Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Fix color getting reset before prompt output done #65

Merged
merged 1 commit into from
Mar 12, 2023

Conversation

blackhole89
Copy link
Contributor

This should fix the issue where sometimes the last few tokens of the initial prompt will not get colored correctly, because the color reset ANSI code was emitted early.

@ggerganov ggerganov merged commit 404fac0 into ggml-org:master Mar 12, 2023
44670 pushed a commit to 44670/llama.cpp that referenced this pull request Aug 2, 2023
* Long-range falcon upgrade (16k context)
Default context is not 2048
The embedding rotation has been adapted to react to context and expected generation
Uses "NTK" fourier aware scaling of the rotation space.

7B and 40B have been tested to work well up to a context of 8k
Tests at > 8k are incoming once performance at these sizes works better

RAM requirements for K/V caches:
Falcon 7B at 8k context : ~2 GB RAM
Falcon 40B at 8k context : ~5.5 GB RAM

In addition falcon_eval() now uses a configuration struct instead of passing many parameters through multiple abstraction layers.
This makes it much easier to pass new features from main into libfalcon

* perplexity bugfix

---------

Co-authored-by: John <cmt-nct@users.noreply.github.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants