Half the data was geared towards better reasoning (EvolKit-20k and reasoning-base-20k), the other half will help to de-censor the model (WizardLM data set).
Looking for GGUF?
There is a separate upload for that! Download theprint/ReWiz-Nemo-12B-Instruct-GGUF instead.
Uploaded model
- Developed by: theprint
- License: apache-2.0
- Finetuned from model : unsloth/Mistral-Nemo-Instruct-2407-bnb-4bit
This mistral model was trained 2x faster with Unsloth and Huggingface's TRL library.
Open LLM Leaderboard Evaluation Results
Detailed results can be found here
| Metric | Value |
|---|---|
| Avg. | 15.63 |
| IFEval (0-Shot) | 10.62 |
| BBH (3-Shot) | 29.93 |
| MATH Lvl 5 (4-Shot) | 7.18 |
| GPQA (0-shot) | 9.84 |
| MuSR (0-shot) | 10.23 |
| MMLU-PRO (5-shot) | 25.99 |
- Downloads last month
- 25
Hardware compatibility
Log In
to view the estimation
2-bit
3-bit
4-bit
5-bit
6-bit
8-bit
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
🙋
Ask for provider support
Model tree for mav23/ReWiz-Nemo-12B-Instruct-GGUF
Base model
unsloth/Mistral-Nemo-Instruct-2407-bnb-4bitEvaluation results
- strict accuracy on IFEval (0-Shot)Open LLM Leaderboard10.620
- normalized accuracy on BBH (3-Shot)Open LLM Leaderboard29.930
- exact match on MATH Lvl 5 (4-Shot)Open LLM Leaderboard7.180
- acc_norm on GPQA (0-shot)Open LLM Leaderboard9.840
- acc_norm on MuSR (0-shot)Open LLM Leaderboard10.230
- accuracy on MMLU-PRO (5-shot)test set Open LLM Leaderboard25.990
