-
Notifications
You must be signed in to change notification settings - Fork 8
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[WIP] Support logprob calculation for loglikelihood approach #69
Conversation
Why do we need this for |
Hello @masahi! Thank you for your reference. The main idea is support accuracy benchmark of octoml endpoints on tasks (like MMLU, HellaSwag) with loglikelihood approach. Unfortunately I'm not familiar with serve implementation and made mistake when implemented it on "old" part of mlc-llm. I plan to use this PR and do it on serve side. Possibly logprobs calculation has been already done or can be easily done, but it also needs some high-level API for request-response of logprobs. |
ae713cd
to
a54943c
Compare
7dbdbc4
to
8b20bb9
Compare
fd393a1
to
a822016
Compare
86d63fa
to
26692df
Compare
Since we have #82, do we still need this PR? |
Hello @sunggg! Yes, of course. Functionality from #82 allows to get logprobs info for new generated tokens. This PR allows to get logprobs for all tokens from input prompt (prefill step) which used for loglikelihood calculation. Particularly here I modify Relax model due to it cut the last set of logits, but I need all of them. |
6c9c57d
to
b0b10ef
Compare
e3bff68
to
5a4bde6
Compare
63492da
to
aaca70f
Compare
79780ac
to
e34d042
Compare
Close due to transfer to octoml/mlc-serve/pull/56 |
Loglikelihood approach is needed to check accuracy of a model on popular datasets and tasks like MMLU and BigBench. Particularly HuggingFace leaderboard bases on such tasks only.
Notes: