MedGo: Medical Large Language Model Based on Qwen3-32B
English | ็ฎไฝไธญๆ
๐ Table of Contents
- Introduction
- Key Features
- Performance
- Quick Start
- Training Details
- Use Cases
- Limitations & Risks
- Citation
- License
- Contributing
- Contact
๐ฏ Introduction
MedGo is a general-purpose medical large language model fine-tuned from Qwen3-32B, designed for clinical medicine and research scenarios. The model is trained on large-scale multi-source medical corpora and enhanced with complex case data, supporting various capabilities including medical Q&A, clinical summary, clinical reasoning, multi-turn dialogue, and scientific text generation.
๐ Core Capabilities
- ๐ Medical Knowledge Q&A: Professional responses based on authoritative medical literature and clinical guidelines
- ๐ Clinical Documentation: Automated medical record summaries, diagnostic reports, and medical documentation
- ๐ Clinical Reasoning: Differential diagnosis, examination recommendations, and treatment suggestions
- ๐ฌ Multi-turn Dialogue: Patient-doctor interaction simulation and complex case discussions
- ๐ฌ Research Support: Literature summarization, research idea generation, and quality control review
โจ Key Features
| Feature | Details |
|---|---|
| Base Architecture | Qwen3-32B |
| Parameters | 32B |
| Domain | Clinical Medicine, Research Support, Healthcare System Integration |
| Fine-tuning Method | SFT + Preference Alignment (DPO/KTO) |
| Data Sources | Authoritative medical literature, clinical guidelines, real cases (anonymized) |
| Deployment | Local deployment, HIS/EMR system integration |
| License | Apache 2.0 |
๐ Performance
MedGo demonstrates excellent performance across multiple medical and general evaluation benchmarks, showing competitive results among 32B-parameter models:
Key Benchmark Results
- AIMedQA: Medical question answering comprehension
- CME: Clinical reasoning evaluation
- DiagnosisArena: Diagnostic capability assessment
- MedQA / MedMCQA: Medical multiple-choice questions
- PubMedQA: Biomedical literature Q&A
- MMLU-Pro: Comprehensive capability evaluation
Performance Highlights:
- โ Average Score: ~70 points (excellent performance in the 32B parameter class)
- โ Strong Tasks: Clinical reasoning (DiagnosisArena, CME) and multi-turn medical Q&A
- โ Balanced Capability: Good performance in medical semantic understanding and multi-task generalization
๐ Quick Start
Requirements
- Python >= 3.8
- PyTorch >= 2.0
- Transformers >= 4.35.0
- CUDA >= 11.8 (for GPU inference)
Installation
# Clone the repository
git clone https://github.com/OpenMedZoo/MedGo.git
cd MedGo
# Install dependencies
pip install -r requirements.txt
Model Download
Download model weights from HuggingFace:
# Using huggingface-cli
huggingface-cli download OpenMedZoo/MedGo --local-dir ./models/MedGo
# Or using git-lfs
git lfs install
git clone https://huggingface.co/OpenMedZoo/MedGo
Basic Inference
from transformers import AutoModelForCausalLM, AutoTokenizer
# Load model and tokenizer
model_path = "OpenMedZoo/MedGo"
tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained(
model_path,
device_map="auto",
trust_remote_code=True,
torch_dtype="auto"
)
# Medical Q&A example
messages = [
{"role": "system", "content": "You are a professional medical assistant. Please answer questions based on medical knowledge."},
{"role": "user", "content": "What is hypertension and what are the common treatment methods?"}
]
# Generate response
inputs = tokenizer.apply_chat_template(
messages,
tokenize=True,
add_generation_prompt=True,
return_tensors="pt"
).to(model.device)
outputs = model.generate(
inputs,
max_new_tokens=512,
temperature=0.7,
top_p=0.9,
do_sample=True
)
response = tokenizer.decode(outputs[0][len(inputs[0]):], skip_special_tokens=True)
print(response)
Batch Inference
# Use the provided inference script
python scripts/inference.py \
--model_path OpenMedZoo/MedGo \
--input_file examples/medical_qa.jsonl \
--output_file results/predictions.jsonl \
--batch_size 4
Accelerated Inference with vLLM
from vllm import LLM, SamplingParams
# Initialize vLLM
llm = LLM(model="OpenMedZoo/MedGo", trust_remote_code=True)
sampling_params = SamplingParams(temperature=0.7, top_p=0.9, max_tokens=512)
# Batch inference
prompts = [
"What are the symptoms and treatment methods for diabetes?",
"What dietary precautions should hypertensive patients take?"
]
outputs = llm.generate(prompts, sampling_params)
for output in outputs:
print(output.outputs[0].text)
๐ง Training Details
MedGo employs a two-stage fine-tuning strategy to balance general medical knowledge with clinical task adaptation.
Stage I: General Medical Alignment
Objective: Establish a solid foundation of medical knowledge and improve Q&A standardization
Data Sources:
- Authoritative medical literature (PubMed, medical textbooks)
- Clinical guidelines and diagnostic standards
- Medical encyclopedia entries and terminology databases
Training Methods:
- Supervised Fine-Tuning (SFT)
- Chain-of-Thought (CoT) guided samples
- Medical terminology alignment and safety constraints
Stage II: Clinical Task Enhancement
Objective: Enhance complex case reasoning and multi-task processing capabilities
Data Sources:
- Real medical records (fully anonymized)
- Outpatient and emergency records with complex multi-diagnosis samples
- Research articles and quality control cases
Data Augmentation Techniques:
- Semantic paraphrasing and multi-perspective expansion
- Complex case synthesis
- Doctor-patient interaction simulation
Training Methods:
- Multi-Task Learning (medical record summary, differential diagnosis, examination suggestions, etc.)
- Preference Alignment (DPO/KTO)
- Expert feedback iterative optimization
Training Optimization Focus
- โ Strengthen information extraction and cross-evidence reasoning for complex cases
- โ Improve medical consistency and interpretability of outputs
- โ Optimize expression compliance and safety
- โ Continuous iteration through expert samples and automated evaluation
๐ก Use Cases
โ Suitable Scenarios
| Scenario | Description |
|---|---|
| Clinical Assistance | Preliminary diagnosis suggestions, medical record writing, formatted report generation |
| Research Support | Literature summarization, research idea generation, data analysis assistance |
| Quality Control | Medical document compliance checking, clinical process quality control |
| System Integration | Embedded in HIS/EMR systems to provide intelligent decision support |
| Medical Education | Case discussions, medical knowledge Q&A, clinical reasoning training |
๐ซ Unsuitable Scenarios
- โ Cannot Replace Doctors: Only an auxiliary tool, not a standalone diagnostic basis
- โ High-Risk Operations: Not recommended for surgical decisions or other high-risk medical operations
- โ Rare Disease Limitations: May perform poorly on rare diseases outside training data
- โ Emergency Care: Not suitable for scenarios requiring immediate decisions
โ ๏ธ Limitations & Risks
Model Limitations
- Understanding Bias: Despite covering extensive medical knowledge, may still produce understanding biases or incorrect recommendations
- Complex Cases: Higher risk for cases with complex conditions, severe complications, or missing information
- Knowledge Currency: Medical knowledge continuously updates; training data may lag
- Language Limitation: Primarily designed for Chinese medical scenarios; performance in other languages may vary
Usage Recommendations
- โ ๏ธ Use in controlled environments with clinical expert review of generated results
- โ ๏ธ Treat model outputs as auxiliary references, not final diagnostic conclusions
- โ ๏ธ For sensitive cases or high-risk scenarios, expert consultation is mandatory
- โ ๏ธ Deployment requires internal validation, security review, and clinical testing
Data Privacy & Compliance
- ๐ Training data fully anonymized
- ๐ Attention to patient privacy protection during use
- ๐ Production deployment must comply with healthcare data security regulations (e.g., HIPAA, GDPR)
- ๐ Local deployment recommended to avoid sensitive data transmission
๐ Citation
If MedGo is helpful for your research or project, please cite our work:
@misc{openmedzoo_2025,
author = { OpenMedZoo },
title = { MedGo (Revision 640a2e2) },
year = 2025,
url = { https://huggingface.co/OpenMedZoo/MedGo },
doi = { 10.57967/hf/7024 },
publisher = { Hugging Face }
}
๐ License
This project is licensed under the Apache License 2.0.
Commercial Use Notice:
- โ Commercial use and modification allowed
- โ Original license and copyright notice must be retained
- โ Contact us for technical support when integrating into healthcare systems
๐ค Contributing
We welcome community contributions! Here's how to participate:
Contribution Types
- ๐ Submit bug reports
- ๐ก Propose new features
- ๐ Improve documentation
- ๐ง Submit code fixes or optimizations
- ๐ Share evaluation results and use cases
๐ Acknowledgments
Thanks to all contributors to the MedGo project:
- Model development and fine-tuning algorithm team
- Data annotation and quality control team
- Clinical expert guidance and review team
- Open-source community support and feedback
Special thanks to:
- Qwen Team for providing excellent foundation models
- All healthcare institutions that provided data and feedback
๐ง Contact
- HuggingFace: Model Homepage
- Downloads last month
- 57
Model tree for OpenMedZoo/MedGo
Evaluation results
- accuracy on Medical-Eval-HLEself-reported19.400
- accuracy on SuperGPQAself-reported37.200
- accuracy on Medbulletsself-reported57.800
- accuracy on MMLU-proself-reported64.300
- accuracy on AfrimedQAself-reported74.700
- accuracy on MedMCQAself-reported68.300
- accuracy on MedQA-USMLEself-reported76.800
- accuracy on CMBself-reported92.500
- accuracy on CMExamself-reported87.400
- accuracy on PubMedQAself-reported76.600
