Update README.md
Browse files
README.md
CHANGED
@@ -29,7 +29,7 @@ This is one of 3 identical versions trained with different random seeds.
|
|
29 |
- **Language(s)**: English
|
30 |
- **License**: Apache 2.0
|
31 |
- **Finetuned from**: `google/t5-v1_1-xxl`
|
32 |
-
- **Paper**: https://arxiv.org/abs/2507.
|
33 |
- **Repository**: https://github.com/itay1itzhak/planted-in-pretraining
|
34 |
|
35 |
## Uses
|
@@ -77,4 +77,18 @@ print(tokenizer.decode(outputs[0]))
|
|
77 |
## Technical Specifications
|
78 |
|
79 |
- Architecture: T5-11B
|
80 |
-
- Instruction dataset: Tulu-2
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
29 |
- **Language(s)**: English
|
30 |
- **License**: Apache 2.0
|
31 |
- **Finetuned from**: `google/t5-v1_1-xxl`
|
32 |
+
- **Paper**: https://arxiv.org/abs/2507.07186
|
33 |
- **Repository**: https://github.com/itay1itzhak/planted-in-pretraining
|
34 |
|
35 |
## Uses
|
|
|
77 |
## Technical Specifications
|
78 |
|
79 |
- Architecture: T5-11B
|
80 |
+
- Instruction dataset: Tulu-2
|
81 |
+
|
82 |
+
## Citation
|
83 |
+
|
84 |
+
```bibtex
|
85 |
+
@misc{itzhak2025plantedpretrainingswayedfinetuning,
|
86 |
+
title={Planted in Pretraining, Swayed by Finetuning: A Case Study on the Origins of Cognitive Biases in LLMs},
|
87 |
+
author={Itay Itzhak and Yonatan Belinkov and Gabriel Stanovsky},
|
88 |
+
year={2025},
|
89 |
+
eprint={2507.07186},
|
90 |
+
archivePrefix={arXiv},
|
91 |
+
primaryClass={cs.CL},
|
92 |
+
url={https://arxiv.org/abs/2507.07186},
|
93 |
+
}
|
94 |
+
```
|