Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

feat: Jan supports most llama.cpp params #3508

Open
1 of 3 tasks
Tracked by #3365
imtuyethan opened this issue Aug 30, 2024 · 1 comment
Open
1 of 3 tasks
Tracked by #3365

feat: Jan supports most llama.cpp params #3508

imtuyethan opened this issue Aug 30, 2024 · 1 comment
Assignees
Labels
category: cortex.cpp Related to cortex.cpp category: model settings Inference params, presets, templates category: threads & chat Threads & chat UI UX issues move to Cortex type: feature request A new feature

Comments

@imtuyethan
Copy link
Contributor

imtuyethan commented Aug 30, 2024

Goal

  • Jan supports most llama.cpp params

Tasklist

Cortex

Jan

  • Update Right Sidebar UX for Jan
  • Enable Jan's API server to pass through most llama.cpp params

Related

Original Post

- [X] I have searched the existing issues

### Is your feature request related to a problem? Please describe it

- llama.cpp Settings (e.g. attention) should be consistent across llama.cpp, Cortex and Jan
- From an Eng perspective, we should ensure llama.cpp settings get bubbled up to Cortex and Jan

### Describe the solution

- [ ] Identify all relevant model settings that need to be synced
- [ ] Design a common format for representing these settings across all projects
- [ ] Jan Model Settings should follow common format
- [ ] Cortex should allow user to pass inference-time and runtime parameters
- [ ] Process for llama.cpp updates (who should this be driven by?)

### Teachability, documentation, adoption, migration strategy

-

### What is the motivation / use case for changing the behavior?
@imtuyethan imtuyethan added the type: feature request A new feature label Aug 30, 2024
@imtuyethan imtuyethan moved this to Planning in Jan & Cortex Aug 30, 2024
@dan-homebrew dan-homebrew changed the title feat: Sync Model Settings Across Jan, Cortex, and llama.cpp feat: llama.cpp params should be bubbled up to Cortex and Jan Sep 3, 2024
@dan-homebrew dan-homebrew changed the title feat: llama.cpp params should be bubbled up to Cortex and Jan feat: llama.cpp params should be synced/bubbled up to Cortex and Jan Sep 3, 2024
@0xSage
Copy link
Contributor

0xSage commented Sep 5, 2024

Related #3140

@dan-homebrew dan-homebrew changed the title feat: llama.cpp params should be synced/bubbled up to Cortex and Jan feat: Jan supports most llama.cpp params Sep 10, 2024
@dan-homebrew dan-homebrew moved this from Planning to Scheduled in Jan & Cortex Sep 26, 2024
@0xSage 0xSage added category: threads & chat Threads & chat UI UX issues category: cortex.cpp Related to cortex.cpp category: model settings Inference params, presets, templates and removed category: engines labels Oct 14, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
category: cortex.cpp Related to cortex.cpp category: model settings Inference params, presets, templates category: threads & chat Threads & chat UI UX issues move to Cortex type: feature request A new feature
Projects
Status: Scheduled
Development

No branches or pull requests

5 participants