Question Answering On Pubchemqa
Metrics
BLEU-2
BLEU-4
MEATOR
ROUGE-1
ROUGE-2
ROUGE-L
Results
Performance results of various models on this benchmark
Model Name | BLEU-2 | BLEU-4 | MEATOR | ROUGE-1 | ROUGE-2 | ROUGE-L | Paper Title | Repository |
---|---|---|---|---|---|---|---|---|
BioMedGPT-10B | 0.234 | 0.141 | 0.308 | 0.386 | 0.206 | 0.332 | BioMedGPT: Open Multimodal Generative Pre-trained Transformer for BioMedicine | |
Llama2-7B-chat | 0.075 | 0.009 | 0.149 | 0.184 | 0.043 | 0.142 | Llama 2: Open Foundation and Fine-Tuned Chat Models |
0 of 2 row(s) selected.