-
Notifications
You must be signed in to change notification settings - Fork 967
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Investigate model aliasing #39
Labels
Comments
Is this still an issue, or you meant to delete? |
Just tracking an idea I had related to #17 basically allow the user to alias their local models to OpenAI model names as many tools have those hard-coded. |
Ahhhhh, okay, I thought you were somehow having models aliasing as an issue 😆 |
Closed
carmonajca
added a commit
to carmonajca/llama-cpp-python
that referenced
this issue
May 17, 2023
* Bugfix: Ensure logs are printed when streaming * Update llama.cpp * Update llama.cpp * Add missing tfs_z paramter * Bump version * Fix docker command * Revert "llama_cpp server: prompt is a string". Closes abetlen#187 This reverts commit b9098b0. * Only support generating one prompt at a time. * Allow model to tokenize strings longer than context length and set add_bos. Closes abetlen#92 * Update llama.cpp * Bump version * Update llama.cpp * Fix obscure Wndows DLL issue. Closes abetlen#208 * chore: add note for Mac m1 installation * Add winmode arg only on windows if python version supports it * Bump mkdocs-material from 9.1.11 to 9.1.12 Bumps [mkdocs-material](https://github.com/squidfunk/mkdocs-material) from 9.1.11 to 9.1.12. - [Release notes](https://github.com/squidfunk/mkdocs-material/releases) - [Changelog](https://github.com/squidfunk/mkdocs-material/blob/master/CHANGELOG) - [Commits](squidfunk/mkdocs-material@9.1.11...9.1.12) --- updated-dependencies: - dependency-name: mkdocs-material dependency-type: direct:development update-type: version-update:semver-patch ... Signed-off-by: dependabot[bot] <support@github.com> * Update README.md Fix typo. * Fix CMakeLists.txt * Add sampling defaults for generate * Update llama.cpp * Add model_alias option to override model_path in completions. Closes abetlen#39 * Update variable name * Update llama.cpp * Fix top_k value. Closes abetlen#220 * Fix last_n_tokens_size * Implement penalize_nl * Format * Update token checks * Move docs link up * Fixd CUBLAS dll load issue in Windows * Check for CUDA_PATH before adding --------- Signed-off-by: dependabot[bot] <support@github.com> Co-authored-by: Andrei Betlen <abetlen@gmail.com> Co-authored-by: Anchen <anchen.li+alias@pepperstone.com> Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com> Co-authored-by: Xiyou Zhou <xiyou.zhou@gmail.com> Co-authored-by: Aneesh Joy <aneeshjoy@gmail.com>
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Allow the user to alias their local models to OpenAI model names as many tools have those hard-coded.
This may cause unexpected issues with tokenization mismatches.
The text was updated successfully, but these errors were encountered: