Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Regarding evaluation format #19

Open
anandsubramanian123 opened this issue Oct 23, 2023 · 1 comment
Open

Regarding evaluation format #19

anandsubramanian123 opened this issue Oct 23, 2023 · 1 comment

Comments

@anandsubramanian123
Copy link

Hi there! I had a query regarding the evaluation of PMC-LLaMA on the MCQ datasets you had used in the paper. I was trying to evaluate models such as PMC-LLaMA on other datasets, and I was curious about the format in which the question was provided to the model. Would it be possible to obtain a sample of how a MEDQA/MEDMCQA question was formatted before it is provided to the model?

Additionally, could you also confirm if models like MedAlpaca and ChatDoctor were evaluated with questions provided in the same format as PMCLLama?

Thank you

@SuperBruceJia
Copy link

Please check our codes for your reference:
https://github.com/vkola-lab/medpodgpt/blob/main/utils/benchmark_utils.py

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants