Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Llama2-7B-chat-4k on PassageRetrieval-zh gets 10.12 #61

Open
fuqichen1998 opened this issue Mar 18, 2024 · 5 comments
Open

Llama2-7B-chat-4k on PassageRetrieval-zh gets 10.12 #61

fuqichen1998 opened this issue Mar 18, 2024 · 5 comments

Comments

@fuqichen1998
Copy link

As the title, my evaluation of Llama2-7B-chat-4k on PassageRetrieval-zh gets 10.12, which is significantly higher than the README (0.5), could you please share why?

@bys0318
Copy link
Member

bys0318 commented Mar 20, 2024

Hi! Are you using the prompt template as in config/dataset2prompt.json?

@bys0318
Copy link
Member

bys0318 commented Mar 20, 2024

We refer to our code here for the llama2 prompt: https://github.com/THUDM/LongBench/blob/main/pred.py#L33

@fuqichen1998
Copy link
Author

Yes, I was using your pred.py to run the inference and evaluation.

@slatter666
Copy link

Yes, I was using your pred.py to run the inference and evaluation.

Acutally I also get the same result

@condy0919
Copy link

We refer to our code here for the llama2 prompt: https://github.com/THUDM/LongBench/blob/main/pred.py#L33

The INST is necessary for llama2-7b/llama2-13b?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants