-
Notifications
You must be signed in to change notification settings - Fork 55
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
generation issue on PMC_LLAMA_7B #16
Comments
In addition, I also want to check the QA benchmark setting. I utilized the greedy decoding on chatdoctor without any FT on MedQA training dataset. The performance gap is large from your results. I wonder did you utilize any other strategies e.g. COT, for the generation. Thanks. |
Thanks for your interest. |
I utilized the prompt you provide. For specific instruction, I used "If you are a doctor, please answer the medical questions based on the patient's description." for text generation and "Answer this multiple choice question and direct output final answer." for multiple choice. However, the 7-b model generate wired answer and is hard to stop.
|
We are still working on the instruction tuning of 7B model now, it's almost done. You could try it on 13B model for now. |
@WeixiongLin, can we have the instruction tuned to the 7B PMC llama checkpoint? |
I try to use PMC_LLAMA_7B for text generation and MedQA. However, it may run into some issue, e.g. copy the pervious input without generate anything. One thing I notice there is the special tokens map is {}. Could you recheck the model in huggingface ?
The text was updated successfully, but these errors were encountered: