Question Answering On Blurb
Metrics
Accuracy
Results
Performance results of various models on this benchmark
Model Name | Accuracy | Paper Title | Repository |
---|---|---|---|
BioLinkBERT (base) | 80.81 | LinkBERT: Pretraining Language Models with Document Links | |
BioLinkBERT (large) | 83.5 | LinkBERT: Pretraining Language Models with Document Links | |
PubMedBERT (uncased; abstracts) | 71.7 | Domain-Specific Language Model Pretraining for Biomedical Natural Language Processing | |
GPT-4 | 80.56 | Evaluation of large language model performance on the Biomedical Language Understanding and Reasoning Benchmark | - |
0 of 4 row(s) selected.