|
--- |
|
license: apache-2.0 |
|
datasets: |
|
- pietrolesci/pubmed-200k-rct |
|
metrics: |
|
- accuracy |
|
base_model: |
|
- openai-community/gpt2 |
|
tags: |
|
- medical |
|
- biology |
|
- research |
|
- pubmed |
|
model-index: |
|
- name: MedGPT |
|
results: [] |
|
demo: |
|
- name: Try in Space |
|
url: https://huggingface.co/spaces/devmanpreet/Medical-GPT2-Classifier |
|
--- |
|
|
|
# MedGPT — GPT-2 Fine-Tuned on PubMed RCT |
|
|
|
MedGPT is a GPT-2 model fine-tuned on the `pubmed-200k-rct` dataset. It classifies individual sentences from biomedical abstracts into one of five standard sections: |
|
|
|
- Background |
|
- Objective |
|
- Methods |
|
- Results |
|
- Conclusion |
|
|
|
This model is useful for tasks requiring structured understanding or summarization of scientific literature. |
|
|
|
## Training Details |
|
|
|
- Base Model: `gpt2` (124M parameters) |
|
- Dataset: `pietrolesci/pubmed-200k-rct` |
|
- Task: Sentence classification |
|
- Labels: Background, Objective, Methods, Results, Conclusion |
|
- Epochs: 1 (partial training) |
|
- Loss Function: CrossEntropy |
|
- Optimizer: AdamW |
|
|