English
nvidia
math
igitman commited on
Commit
4fe2334
·
verified ·
1 Parent(s): ef3a477

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -3
README.md CHANGED
@@ -54,13 +54,17 @@ The pipeline we used to produce the data and models is fully open-sourced!
54
  - [Models](https://huggingface.co/collections/nvidia/openmath-2-66fb142317d86400783d2c7b)
55
  - [Dataset](https://huggingface.co/datasets/nvidia/OpenMathInstruct-2)
56
 
 
57
 
58
  # How to use the models?
59
 
60
- Our models are fully compatible with Llama3.1-instruct format, so you should be able to just replace an existing Llama3.1 checkpoint and use it in the same way.
61
- Please note that these models have not been instruction tuned and might not provide good answers outside of math domain.
62
 
63
- If you don't know how to use Llama3.1 models, we provide convenient [instructions in our repo](https://github.com/Kipok/NeMo-Skills/blob/main/docs/inference.md).
 
 
 
64
 
65
  # Reproducing our results
66
 
 
54
  - [Models](https://huggingface.co/collections/nvidia/openmath-2-66fb142317d86400783d2c7b)
55
  - [Dataset](https://huggingface.co/datasets/nvidia/OpenMathInstruct-2)
56
 
57
+ See our [paper](https://arxiv.org/abs/2410.01560) to learn more details!
58
 
59
  # How to use the models?
60
 
61
+ Our models are trained with the same "chat format" as Llama3.1-instruct models (same system/user/assistant tokens).
62
+ Please note that these models have not been instruction tuned on general data and thus might not provide good answers outside of math domain.
63
 
64
+ This is a NeMo checkpoint, so you need to use [NeMo Framework](https://github.com/NVIDIA/NeMo) to run inference or finetune it.
65
+ We also release a [HuggingFace checkpoint](https://huggingface.co/nvidia/OpenMath2-Llama3.1-8B) and provide easy instructions on how to
66
+ [convert between different formats](https://github.com/Kipok/NeMo-Skills/blob/main/docs/checkpoint-conversion.md) or
67
+ [run inference](https://github.com/Kipok/NeMo-Skills/blob/main/docs/inference.md) with these models using our codebase.
68
 
69
  # Reproducing our results
70