itay1itzhak nielsr HF Staff commited on
Commit
1801905
·
verified ·
1 Parent(s): 3a2f913

Improve model card: Update pipeline tag, add project page and HF paper link (#1)

Browse files

- Improve model card: Update pipeline tag, add project page and HF paper link (d052a70174f2ad602357e83f856b45e38a979a98)


Co-authored-by: Niels Rogge <[email protected]>

Files changed (1) hide show
  1. README.md +11 -10
README.md CHANGED
@@ -1,18 +1,18 @@
1
  ---
 
 
 
 
 
 
 
2
  license: apache-2.0
 
3
  tags:
4
  - language-modeling
5
  - causal-lm
6
  - bias-analysis
7
  - cognitive-bias
8
- datasets:
9
- - allenai/tulu-v2-sft-mixture
10
- language:
11
- - en
12
- base_model:
13
- - google/t5-v1_1-xxl
14
- pipeline_tag: text2text-generation
15
- library_name: transformers
16
  ---
17
 
18
  # Model Card for T5-Tulu
@@ -20,8 +20,8 @@ library_name: transformers
20
  ## Model Details
21
 
22
  **Model Description**
23
- This 🤗 Transformers model was finetuned using LoRA adapters for the arXiv paper:
24
- **"Planted in Pretraining, Swayed by Finetuning: A Case Study on the Origins of Cognitive Biases in LLMs"**
25
  We study whether cognitive biases in LLMs emerge from pretraining, instruction tuning, or training randomness.
26
  This is one of 3 identical versions trained with different random seeds.
27
 
@@ -30,6 +30,7 @@ This is one of 3 identical versions trained with different random seeds.
30
  - **License**: Apache 2.0
31
  - **Finetuned from**: `google/t5-v1_1-xxl`
32
  - **Paper**: https://arxiv.org/abs/2507.07186
 
33
  - **Repository**: https://github.com/itay1itzhak/planted-in-pretraining
34
 
35
  ## Uses
 
1
  ---
2
+ base_model:
3
+ - google/t5-v1_1-xxl
4
+ datasets:
5
+ - allenai/tulu-v2-sft-mixture
6
+ language:
7
+ - en
8
+ library_name: transformers
9
  license: apache-2.0
10
+ pipeline_tag: text-generation
11
  tags:
12
  - language-modeling
13
  - causal-lm
14
  - bias-analysis
15
  - cognitive-bias
 
 
 
 
 
 
 
 
16
  ---
17
 
18
  # Model Card for T5-Tulu
 
20
  ## Model Details
21
 
22
  **Model Description**
23
+ This 🤗 Transformers model was finetuned using LoRA adapters for the paper:
24
+ **"Planted in Pretraining, Swayed by Finetuning: A Case Study on the Origins of Cognitive Biases in LLMs"** ([Hugging Face Paper](https://huggingface.co/papers/2507.07186), [arXiv](https://arxiv.org/abs/2507.07186))
25
  We study whether cognitive biases in LLMs emerge from pretraining, instruction tuning, or training randomness.
26
  This is one of 3 identical versions trained with different random seeds.
27
 
 
30
  - **License**: Apache 2.0
31
  - **Finetuned from**: `google/t5-v1_1-xxl`
32
  - **Paper**: https://arxiv.org/abs/2507.07186
33
+ - **Project Page**: https://itay1itzhak.github.io/planted-in-pretraining
34
  - **Repository**: https://github.com/itay1itzhak/planted-in-pretraining
35
 
36
  ## Uses