itay1itzhak commited on
Commit
3a2f913
·
verified ·
1 Parent(s): 1379259

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +16 -2
README.md CHANGED
@@ -29,7 +29,7 @@ This is one of 3 identical versions trained with different random seeds.
29
  - **Language(s)**: English
30
  - **License**: Apache 2.0
31
  - **Finetuned from**: `google/t5-v1_1-xxl`
32
- - **Paper**: https://arxiv.org/abs/2507.XXXXX
33
  - **Repository**: https://github.com/itay1itzhak/planted-in-pretraining
34
 
35
  ## Uses
@@ -77,4 +77,18 @@ print(tokenizer.decode(outputs[0]))
77
  ## Technical Specifications
78
 
79
  - Architecture: T5-11B
80
- - Instruction dataset: Tulu-2
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
29
  - **Language(s)**: English
30
  - **License**: Apache 2.0
31
  - **Finetuned from**: `google/t5-v1_1-xxl`
32
+ - **Paper**: https://arxiv.org/abs/2507.07186
33
  - **Repository**: https://github.com/itay1itzhak/planted-in-pretraining
34
 
35
  ## Uses
 
77
  ## Technical Specifications
78
 
79
  - Architecture: T5-11B
80
+ - Instruction dataset: Tulu-2
81
+
82
+ ## Citation
83
+
84
+ ```bibtex
85
+ @misc{itzhak2025plantedpretrainingswayedfinetuning,
86
+ title={Planted in Pretraining, Swayed by Finetuning: A Case Study on the Origins of Cognitive Biases in LLMs},
87
+ author={Itay Itzhak and Yonatan Belinkov and Gabriel Stanovsky},
88
+ year={2025},
89
+ eprint={2507.07186},
90
+ archivePrefix={arXiv},
91
+ primaryClass={cs.CL},
92
+ url={https://arxiv.org/abs/2507.07186},
93
+ }
94
+ ```