File size: 3,133 Bytes
f62904b 1cab402 f62904b a7b3188 f62904b 1cab402 f62904b 8be5f3c f62904b 1cab402 f62904b 1cab402 fffa5ce f62904b 409c788 f62904b dfb0245 f62904b dfb0245 f62904b dfb0245 f62904b dfb0245 f62904b 374b2b5 f62904b fffa5ce f62904b 714662b 21c1e36 714662b f62904b 714662b f62904b dfb0245 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 |
---
library_name: transformers
license: cc-by-sa-4.0
language:
- en
- ja
base_model:
- EQUES/JPharmatron-7B-base
tags:
- pharmacy
- biology
- chemistry
- medical
---
# JPharmatron-7B
<!-- Provide a quick summary of what the model is/does. -->
JPharmatron-7B is a 7B large language model designed for pharmaceutical applications and researches.
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
The JPharmatron-7B is continually pre-trained using 8.8B tokens from Japanese and English datasets, based on Qwen2.5-7B. Compared to the JPharmatron-7B-base model, JPharmatron-7B has enhanced chat capabilities, obtained from Qwen2.5-7B-Instruct's chat vector.
- **Developed by:** EQUES Inc.
- **Funded by [optional]:** [GENIAC Project](https://www.meti.go.jp/policy/mono_info_service/geniac/index.html)
- **Model type:** Causal decoder-only
- **Language(s) (NLP):** Japanese, English
- **License:** CC-BY-SA-4.0
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** https://github.com/EQUES-Inc/pharma-LLM-eval
- **Paper [optional]:** [A Japanese Language Model and Three New Evaluation Benchmarks for Pharmaceutical NLP](https://arxiv.org/abs/2505.16661) (IJCNLP-AACL 2025)
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
This model is intended for applications in pharmaceutical paperwork and research. It is not validated for medical use or any other risk-sensitive use.
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
We evaluated our model, JPharmatron-7B, with other general / domain-specific models of a similar size.
### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[JPharmaBench](https://huggingface.co/collections/EQUES/jpharmabench-680a34acfe96870e41d050d8) and two existing benchmarks (JMMLU (pharma) and IgakuQA) were used.
### Results
Compared to Meditron3-Qwen2.5-7B and Llama3.1-Swallow-8B-Instruct-v0.3, JPharmatron-7B achieved the highest score on all of the five benchmarks.

## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**This paper has been accepted to IJCNLP-AACL 2025. We will update the bibtex info below soon.**
**BibTeX:**
```
@misc{ono2025japaneselanguagemodelnew,
title={A Japanese Language Model and Three New Evaluation Benchmarks for Pharmaceutical NLP},
author={Shinnosuke Ono and Issey Sukeda and Takuro Fujii and Kosei Buma and Shunsuke Sasaki},
year={2025},
eprint={2505.16661},
archivePrefix={arXiv},
primaryClass={cs.CL},
url={https://arxiv.org/abs/2505.16661},
}
```
## More Information [optional]
See our preprint: [A Japanese Language Model and Three New Evaluation Benchmarks for Pharmaceutical NLP](https://arxiv.org/abs/2505.16661).
## Model Card Authors [optional]
[@shinnosukeono](https://shinnosukeono.github.io/)
|