--- base_model: - meta-llama/Llama-3.1-8B datasets: - MegaScience/MegaScience language: - en license: llama3.1 metrics: - accuracy pipeline_tag: text-generation library_name: transformers tags: - science --- # [MegaScience: Pushing the Frontiers of Post-Training Datasets for Science Reasoning](https://arxiv.org/abs/2507.16812) **Llama3.1-8B-MegaScience** is a model fine-tuned on **MegaScience**, a large-scale mixture of high-quality open-source scientific datasets totaling 1.25 million instances, as presented in the paper "MegaScience: Pushing the Frontiers of Post-Training Datasets for Science Reasoning". The MegaScience dataset features truthful reference answers extracted from 12k university-level scientific textbooks, comprising 650k reasoning questions spanning 7 scientific disciplines. This model significantly outperforms corresponding official instruct models in average performance on scientific reasoning tasks and exhibits greater effectiveness for larger and stronger models, suggesting a scaling benefit for scientific tuning. For more details on the project, including the data curation pipeline and evaluation system, visit the [official GitHub repository](https://github.com/GAIR-NLP/lm-open-science-evaluation). ## Llama3.1-8B-MegaScience ### Training Recipe - **LR**: 5e-6 - **LR Schedule**: Cosine - **Batch Size**: 512 - **Max Length**: 4,096 - **Warm Up Ratio**: 0.05 - **Epochs**: 3 ### Evaluation Results