Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

generation issue on PMC_LLAMA_7B #16

Open
XZhang97666 opened this issue Oct 5, 2023 · 5 comments
Open

generation issue on PMC_LLAMA_7B #16

XZhang97666 opened this issue Oct 5, 2023 · 5 comments

Comments

@XZhang97666
Copy link

I try to use PMC_LLAMA_7B for text generation and MedQA. However, it may run into some issue, e.g. copy the pervious input without generate anything. One thing I notice there is the special tokens map is {}. Could you recheck the model in huggingface ?

@XZhang97666
Copy link
Author

In addition, I also want to check the QA benchmark setting. I utilized the greedy decoding on chatdoctor without any FT on MedQA training dataset. The performance gap is large from your results. I wonder did you utilize any other strategies e.g. COT, for the generation. Thanks.

@WeixiongLin
Copy link
Collaborator

Thanks for your interest.
May I have your input prompt please. Open source LLMs are often sensitive to prompts, so it might have influence on the performance. Besides, you could try out diffrent decoding strategies (e.g. topk).

@XZhang97666
Copy link
Author

Thanks for your interest. May I have your input prompt please. Open source LLMs are often sensitive to prompts, so it might have influence on the performance. Besides, you could try out diffrent decoding strategies (e.g. topk).

I utilized the prompt you provide. For specific instruction, I used "If you are a doctor, please answer the medical questions based on the patient's description." for text generation and "Answer this multiple choice question and direct output final answer." for multiple choice. However, the 7-b model generate wired answer and is hard to stop.

PROMPT_DICT = {
    "prompt_input": (
        "Below is an instruction that describes a task, paired with an input that provides further context. "
        "Write a response that appropriately completes the request.\n\n"
        "### Instruction:\n{instruction}\n\n### Input:\n{input}\n\n### Response:"
    ),
    "prompt_no_input": (
        "Below is an instruction that describes a task. "
        "Write a response that appropriately completes the request.\n\n"
        "### Instruction:\n{instruction}\n\n### Response:"
    ),
}

@WeixiongLin
Copy link
Collaborator

We are still working on the instruction tuning of 7B model now, it's almost done. You could try it on 13B model for now.

@shamanez
Copy link

@WeixiongLin, can we have the instruction tuned to the 7B PMC llama checkpoint?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants