-
-
Notifications
You must be signed in to change notification settings - Fork 5.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
codellama 70b don't stop generating #2686
Comments
you need to either: set the eos token to |
I also have this problem with
I don't know if that's the right way to use it. |
I tried setting Sometimes it works and ends properly
But sometimes it includes Source: assistant EOT: true in response
How can I set eos token to |
Changing EOS is hard (or maybe not possible in vLLM), you would have to
clone the model and update the EOS token in tokenizer_config.json, also in
config.json and generation config.
Another approach is to add "Source: assistant" and "Source: user" as stop
sequences too.
…On Thu, Feb 1, 2024 at 2:24 PM victorserbu2709 ***@***.***> wrote:
I tried setting stop=""
Sometimes it works and ends properly
curl http://localhost:8000/v1/completions -H 'Content-Type: application/json' -d '{
"model": "codellama/CodeLlama-70b-Instruct-hf",
"max_tokens": 2048,
"prompt": "<s>Source: system\n\n You are an helpful assistant <step> Source: user\n\n hello<step> Source: assistant\nDestination: user\n\n ",
"stop": "<step>", "stream":"true", "temperature":0.09
}'
data: {"id": "cmpl-d99811e57dc0492fbee4e9e5082333e2", "created": 4162325, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": "", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-d99811e57dc0492fbee4e9e5082333e2", "created": 4162325, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": "", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-d99811e57dc0492fbee4e9e5082333e2", "created": 4162325, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": "", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-d99811e57dc0492fbee4e9e5082333e2", "created": 4162325, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": "😊", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-d99811e57dc0492fbee4e9e5082333e2", "created": 4162325, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " Hi", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-d99811e57dc0492fbee4e9e5082333e2", "created": 4162325, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " there", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-d99811e57dc0492fbee4e9e5082333e2", "created": 4162325, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": "!", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-d99811e57dc0492fbee4e9e5082333e2", "created": 4162325, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " I", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-d99811e57dc0492fbee4e9e5082333e2", "created": 4162325, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": "'", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-d99811e57dc0492fbee4e9e5082333e2", "created": 4162325, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": "m", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-d99811e57dc0492fbee4e9e5082333e2", "created": 4162325, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " an", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-d99811e57dc0492fbee4e9e5082333e2", "created": 4162325, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " A", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-d99811e57dc0492fbee4e9e5082333e2", "created": 4162325, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": "I", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-d99811e57dc0492fbee4e9e5082333e2", "created": 4162325, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " assistant", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-d99811e57dc0492fbee4e9e5082333e2", "created": 4162325, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": ",", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-d99811e57dc0492fbee4e9e5082333e2", "created": 4162325, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " and", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-d99811e57dc0492fbee4e9e5082333e2", "created": 4162325, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " I", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-d99811e57dc0492fbee4e9e5082333e2", "created": 4162325, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": "'", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-d99811e57dc0492fbee4e9e5082333e2", "created": 4162325, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": "m", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-d99811e57dc0492fbee4e9e5082333e2", "created": 4162325, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " here", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-d99811e57dc0492fbee4e9e5082333e2", "created": 4162325, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " to", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-d99811e57dc0492fbee4e9e5082333e2", "created": 4162325, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " help", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-d99811e57dc0492fbee4e9e5082333e2", "created": 4162325, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " you", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-d99811e57dc0492fbee4e9e5082333e2", "created": 4162325, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " with", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-d99811e57dc0492fbee4e9e5082333e2", "created": 4162325, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " any", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-d99811e57dc0492fbee4e9e5082333e2", "created": 4162325, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " questions", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-d99811e57dc0492fbee4e9e5082333e2", "created": 4162325, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " or", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-d99811e57dc0492fbee4e9e5082333e2", "created": 4162325, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " tasks", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-d99811e57dc0492fbee4e9e5082333e2", "created": 4162325, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " you", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-d99811e57dc0492fbee4e9e5082333e2", "created": 4162325, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " may", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-d99811e57dc0492fbee4e9e5082333e2", "created": 4162325, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " have", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-d99811e57dc0492fbee4e9e5082333e2", "created": 4162325, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": ".", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-d99811e57dc0492fbee4e9e5082333e2", "created": 4162325, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " What", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-d99811e57dc0492fbee4e9e5082333e2", "created": 4162325, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " would", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-d99811e57dc0492fbee4e9e5082333e2", "created": 4162325, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " you", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-d99811e57dc0492fbee4e9e5082333e2", "created": 4162325, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " like", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-d99811e57dc0492fbee4e9e5082333e2", "created": 4162325, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " to", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-d99811e57dc0492fbee4e9e5082333e2", "created": 4162325, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " chat", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-d99811e57dc0492fbee4e9e5082333e2", "created": 4162325, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " about", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-d99811e57dc0492fbee4e9e5082333e2", "created": 4162325, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " or", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-d99811e57dc0492fbee4e9e5082333e2", "created": 4162325, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " get", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-d99811e57dc0492fbee4e9e5082333e2", "created": 4162325, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " help", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-d99811e57dc0492fbee4e9e5082333e2", "created": 4162325, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " with", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-d99811e57dc0492fbee4e9e5082333e2", "created": 4162325, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " today", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-d99811e57dc0492fbee4e9e5082333e2", "created": 4162325, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": "?", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-d99811e57dc0492fbee4e9e5082333e2", "created": 4162325, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " ", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-d99811e57dc0492fbee4e9e5082333e2", "created": 4162325, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": "", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-d99811e57dc0492fbee4e9e5082333e2", "created": 4162325, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": "", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-d99811e57dc0492fbee4e9e5082333e2", "created": 4162325, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": "", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-d99811e57dc0492fbee4e9e5082333e2", "created": 4162325, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": "🤔", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-d99811e57dc0492fbee4e9e5082333e2", "created": 4162325, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": "", "logprobs": null, "finish_reason": "stop"}]}
data: {"id": "cmpl-d99811e57dc0492fbee4e9e5082333e2", "created": 4162325, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": "", "logprobs": null, "finish_reason": "stop"}], "usage": {"prompt_tokens": 33, "total_tokens": 84, "completion_tokens": 51}}
But sometimes it includes Source: assistant EOT: true in response
curl http://localhost:8000/v1/completions -H 'Content-Type: application/json' -d '{
"model": "codellama/CodeLlama-70b-Instruct-hf",
"max_tokens": 2048,
"prompt": "<s>Source: system\n\n You are an helpful assistant <step> Source: user\n\n hello<step> Source: assistant\nDestination: user\n\n ",
"stop": "<step>", "stream":"true", "temperature":0.09
}'
data: {"id": "cmpl-f9871eed2ce1424e8b533fb3fe8a910c", "created": 4162329, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": "", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-f9871eed2ce1424e8b533fb3fe8a910c", "created": 4162329, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": "", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-f9871eed2ce1424e8b533fb3fe8a910c", "created": 4162329, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": "", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-f9871eed2ce1424e8b533fb3fe8a910c", "created": 4162329, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": "😊", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-f9871eed2ce1424e8b533fb3fe8a910c", "created": 4162329, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " Hi", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-f9871eed2ce1424e8b533fb3fe8a910c", "created": 4162329, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " there", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-f9871eed2ce1424e8b533fb3fe8a910c", "created": 4162329, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": "!", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-f9871eed2ce1424e8b533fb3fe8a910c", "created": 4162329, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " I", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-f9871eed2ce1424e8b533fb3fe8a910c", "created": 4162329, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": "'", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-f9871eed2ce1424e8b533fb3fe8a910c", "created": 4162329, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": "m", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-f9871eed2ce1424e8b533fb3fe8a910c", "created": 4162329, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " an", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-f9871eed2ce1424e8b533fb3fe8a910c", "created": 4162329, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " A", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-f9871eed2ce1424e8b533fb3fe8a910c", "created": 4162329, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": "I", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-f9871eed2ce1424e8b533fb3fe8a910c", "created": 4162329, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " assistant", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-f9871eed2ce1424e8b533fb3fe8a910c", "created": 4162329, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": ",", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-f9871eed2ce1424e8b533fb3fe8a910c", "created": 4162329, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " and", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-f9871eed2ce1424e8b533fb3fe8a910c", "created": 4162329, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " I", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-f9871eed2ce1424e8b533fb3fe8a910c", "created": 4162329, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": "'", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-f9871eed2ce1424e8b533fb3fe8a910c", "created": 4162329, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": "m", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-f9871eed2ce1424e8b533fb3fe8a910c", "created": 4162329, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " here", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-f9871eed2ce1424e8b533fb3fe8a910c", "created": 4162329, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " to", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-f9871eed2ce1424e8b533fb3fe8a910c", "created": 4162329, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " help", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-f9871eed2ce1424e8b533fb3fe8a910c", "created": 4162329, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": ".", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-f9871eed2ce1424e8b533fb3fe8a910c", "created": 4162329, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " What", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-f9871eed2ce1424e8b533fb3fe8a910c", "created": 4162329, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " can", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-f9871eed2ce1424e8b533fb3fe8a910c", "created": 4162329, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " I", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-f9871eed2ce1424e8b533fb3fe8a910c", "created": 4162329, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " assist", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-f9871eed2ce1424e8b533fb3fe8a910c", "created": 4162329, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " you", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-f9871eed2ce1424e8b533fb3fe8a910c", "created": 4162329, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " with", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-f9871eed2ce1424e8b533fb3fe8a910c", "created": 4162329, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " today", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-f9871eed2ce1424e8b533fb3fe8a910c", "created": 4162329, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": "?", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-f9871eed2ce1424e8b533fb3fe8a910c", "created": 4162329, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " ", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-f9871eed2ce1424e8b533fb3fe8a910c", "created": 4162329, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": "", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-f9871eed2ce1424e8b533fb3fe8a910c", "created": 4162329, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": "", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-f9871eed2ce1424e8b533fb3fe8a910c", "created": 4162329, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": "", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-f9871eed2ce1424e8b533fb3fe8a910c", "created": 4162329, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": "🤔", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-f9871eed2ce1424e8b533fb3fe8a910c", "created": 4162329, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " Source", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-f9871eed2ce1424e8b533fb3fe8a910c", "created": 4162329, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": ":", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-f9871eed2ce1424e8b533fb3fe8a910c", "created": 4162329, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " assistant", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-f9871eed2ce1424e8b533fb3fe8a910c", "created": 4162329, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": "\n", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-f9871eed2ce1424e8b533fb3fe8a910c", "created": 4162329, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": "E", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-f9871eed2ce1424e8b533fb3fe8a910c", "created": 4162329, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": "OT", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-f9871eed2ce1424e8b533fb3fe8a910c", "created": 4162329, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": ":", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-f9871eed2ce1424e8b533fb3fe8a910c", "created": 4162329, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": " true", "logprobs": null, "finish_reason": null}]}
data: {"id": "cmpl-f9871eed2ce1424e8b533fb3fe8a910c", "created": 4162329, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": "", "logprobs": null, "finish_reason": "stop"}]}
data: {"id": "cmpl-f9871eed2ce1424e8b533fb3fe8a910c", "created": 4162329, "model": "codellama/CodeLlama-70b-Instruct-hf", "choices": [{"index": 0, "text": "", "logprobs": null, "finish_reason": "stop"}], "usage": {"prompt_tokens": 33, "total_tokens": 78, "completion_tokens": 45}}
data: [DONE]
How can I set eos token to ?
—
Reply to this email directly, view it on GitHub
<#2686 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/ASVG6CW4DTCEZISL6H7K67LYROQR7AVCNFSM6AAAAABCTCTA32VHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMYTSMRRGQ2TKOJVGY>
.
You are receiving this because you commented.Message ID:
***@***.***>
|
@RonanKMcGovern What are the stop sequences for FIM case? |
What do you mean by FIM?
…On Sat, Mar 9, 2024 at 6:08 AM icnahom ***@***.***> wrote:
@RonanKMcGovern <https://github.com/RonanKMcGovern> What are the stop
sequences for FIM case?
—
Reply to this email directly, view it on GitHub
<#2686 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/ASVG6CW2G6FORBGHYNRJMFDYXKRO5AVCNFSM6AAAAABCTCTA32VHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMYTSOBWG42DKNBRGM>
.
You are receiving this because you were mentioned.Message ID:
***@***.***>
|
Fill-in-middle or Code Infilling
|
ah, sorry, yeah I don't know as I haven't tried that
…On Sat, Mar 9, 2024 at 1:50 PM icnahom ***@***.***> wrote:
Fill-in-middle or Code Infilling
On Sat, Mar 9, 2024, 3:55 PM Ronan McGovern ***@***.***>
wrote:
> What do you mean by FIM?
>
> On Sat, Mar 9, 2024 at 6:08 AM icnahom ***@***.***> wrote:
>
> > @RonanKMcGovern <https://github.com/RonanKMcGovern> What are the stop
> > sequences for FIM case?
> >
> > —
> > Reply to this email directly, view it on GitHub
> > <
> #2686 (comment)>,
>
> > or unsubscribe
> > <
>
https://github.com/notifications/unsubscribe-auth/ASVG6CW2G6FORBGHYNRJMFDYXKRO5AVCNFSM6AAAAABCTCTA32VHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMYTSOBWG42DKNBRGM>
>
> > .
> > You are receiving this because you were mentioned.Message ID:
> > ***@***.***>
> >
>
> —
> Reply to this email directly, view it on GitHub
> <
#2686 (comment)>,
> or unsubscribe
> <
https://github.com/notifications/unsubscribe-auth/AONTRXO5IZU5YGBCWBZWHETYXMBDZAVCNFSM6AAAAABCTCTA32VHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMYTSOBWHA2DSOBUGA>
> .
> You are receiving this because you commented.Message ID:
> ***@***.***>
>
—
Reply to this email directly, view it on GitHub
<#2686 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/ASVG6CTV27NE4XSQTIYLGODYXMHTNAVCNFSM6AAAAABCTCTA32VHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMYTSOBWHA3DENZWHA>
.
You are receiving this because you were mentioned.Message ID:
***@***.***>
|
Should be resolved by #4182 |
Hello
I tried running codellama 70b using docker.io/vllm/vllm-openai:v0.2.7 docker image.
But it seems to stop only when max_tokens are returned
If I set "stop": [" Source: assistant\nEOT: true"] it seems to work:
But the problem with stop field is that it is included when stream:true
The text was updated successfully, but these errors were encountered: