Help regarding understanding Alpaca output without sampling. #342
Unanswered
aniquetahir
asked this question in
Q&A
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Can someone help me understand what is happening if I call Alpaca without the sampling boilerplate. Intuitively, without sampling, the output should just be the most probable next words. However when I call the model, the output does not start with the sentence start token. Additionally, the words seem to be mixed up.
e.g.
Following is my input:
I run the following to get the output:
Following is the output of
tokenizer.decode(torch.argmax(output.logits[0], dim=1).tolist())
:This is not what I would expect.
Beta Was this translation helpful? Give feedback.
All reactions