Update README.md
Browse files
README.md
CHANGED
|
@@ -29,7 +29,7 @@ This is one of 3 identical versions trained with different random seeds.
|
|
| 29 |
- **Language(s)**: English
|
| 30 |
- **License**: Apache 2.0
|
| 31 |
- **Finetuned from**: `google/t5-v1_1-xxl`
|
| 32 |
-
- **Paper**: https://arxiv.org/abs/2507.
|
| 33 |
- **Repository**: https://github.com/itay1itzhak/planted-in-pretraining
|
| 34 |
|
| 35 |
## Uses
|
|
@@ -77,4 +77,18 @@ print(tokenizer.decode(outputs[0]))
|
|
| 77 |
## Technical Specifications
|
| 78 |
|
| 79 |
- Architecture: T5-11B
|
| 80 |
-
- Instruction dataset: Tulu-2
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 29 |
- **Language(s)**: English
|
| 30 |
- **License**: Apache 2.0
|
| 31 |
- **Finetuned from**: `google/t5-v1_1-xxl`
|
| 32 |
+
- **Paper**: https://arxiv.org/abs/2507.07186
|
| 33 |
- **Repository**: https://github.com/itay1itzhak/planted-in-pretraining
|
| 34 |
|
| 35 |
## Uses
|
|
|
|
| 77 |
## Technical Specifications
|
| 78 |
|
| 79 |
- Architecture: T5-11B
|
| 80 |
+
- Instruction dataset: Tulu-2
|
| 81 |
+
|
| 82 |
+
## Citation
|
| 83 |
+
|
| 84 |
+
```bibtex
|
| 85 |
+
@misc{itzhak2025plantedpretrainingswayedfinetuning,
|
| 86 |
+
title={Planted in Pretraining, Swayed by Finetuning: A Case Study on the Origins of Cognitive Biases in LLMs},
|
| 87 |
+
author={Itay Itzhak and Yonatan Belinkov and Gabriel Stanovsky},
|
| 88 |
+
year={2025},
|
| 89 |
+
eprint={2507.07186},
|
| 90 |
+
archivePrefix={arXiv},
|
| 91 |
+
primaryClass={cs.CL},
|
| 92 |
+
url={https://arxiv.org/abs/2507.07186},
|
| 93 |
+
}
|
| 94 |
+
```
|