Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

specific model #48

Open
qspang opened this issue Jan 22, 2024 · 13 comments
Open

specific model #48

qspang opened this issue Jan 22, 2024 · 13 comments

Comments

@qspang
Copy link

qspang commented Jan 22, 2024

Can I ask you what models the following three images correspond to?image

@Viol2000
Copy link
Collaborator

  1. codellama 2) codellama-instruct 3) llama-2-chat

@qspang
Copy link
Author

qspang commented Jan 23, 2024

btw, may I ask if AVG COMPRESS RATIO uses Lookahead’s acceleration effect?
image
image

@Viol2000
Copy link
Collaborator

AVG COMPRESS RATIO shows the #generated tokens/#decoding steps with lookahead decoding. It is the upper bound of the speedups.

@qspang
Copy link
Author

qspang commented Jan 23, 2024

I want to know the specific acceleration effect. Which indicator should I look at?

@Viol2000
Copy link
Collaborator

I guess you need to re-run the experiments without lade, like running with USE_LADE=0. Then you can compare the average throughputs shown in the figure above with the throughputs without lade.

@qspang
Copy link
Author

qspang commented Jan 23, 2024

but there are two average throughputs:average throughputs1、average throughputs2,Which one should be used for comparison, or should both be used for comparison and then averaged?

@Viol2000
Copy link
Collaborator

I think both are reasonable. Throughtput1 is the sum(throughput for each questions)/#questions. Throughput2 is the #generated tokens/#decoding steps for the whole dataset.

@qspang
Copy link
Author

qspang commented Jan 23, 2024

Thank you for your patient reply!!!:)

@jivanph
Copy link

jivanph commented Jan 23, 2024

How do you measure "throughput for each question"?

@Viol2000
Copy link
Collaborator

#generated tokens/#time

@xinlong-yang
Copy link

#generated tokens/#time

Hello, can you provide script to evaluate codellama on Human-Eval, like what you do on MT-bench?

@Viol2000
Copy link
Collaborator

Viol2000 commented Jan 23, 2024

#generated tokens/#time

Hello, can you provide script to evaluate codellama on Human-Eval, like what you do on MT-bench?

I will upload them in the following one or two weeks.

@xinlong-yang
Copy link

#generated tokens/#time

Hello, can you provide script to evaluate codellama on Human-Eval, like what you do on MT-bench?

I will upload them in the following one or two weeks.

ok, thanks for you amazing work!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants