abbatea commited on
Commit
ce043cf
·
verified ·
1 Parent(s): 7e96257

corrected grammar

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -35,7 +35,7 @@ Repository: **[Github code for SFT Fine-tuning on MathDial](https://github.com/e
35
 
36
  Training input and output:
37
  The model was fine-tuned on the **[MathDial dataset](https://huggingface.co/datasets/eth-nlped/mathdial-chat/viewer/default/train?views%5B%5D=train&row=0)**.
38
- Each training example consisted of a **Instruction**, **Student's Name**, **Math Word Problem and Solution**, **The students initial approach** as input and the **tutor’s step-by-step solution** as the target output.
39
  To incorporate the whole conversation, a sliding window approach was used. Every input has the same format:
40
  For each step in a conversation, the model input included **all previous turns** in the dialogue (sliding window), followed by the student’s next message. The model’s output was then the **next tutor response** from the dataset.
41
  This approach ensures the model learns to generate responses that are context-aware.
 
35
 
36
  Training input and output:
37
  The model was fine-tuned on the **[MathDial dataset](https://huggingface.co/datasets/eth-nlped/mathdial-chat/viewer/default/train?views%5B%5D=train&row=0)**.
38
+ Each training example consisted of a **Instruction**, **Student's Name**, **Math Word Problem and Solution** and **The students initial approach** as input, followed by the **tutor’s step-by-step solution** as the target output.
39
  To incorporate the whole conversation, a sliding window approach was used. Every input has the same format:
40
  For each step in a conversation, the model input included **all previous turns** in the dialogue (sliding window), followed by the student’s next message. The model’s output was then the **next tutor response** from the dataset.
41
  This approach ensures the model learns to generate responses that are context-aware.