Add pipeline tag, library name, and abstract to model card

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +7 -3
README.md CHANGED
@@ -1,13 +1,17 @@
1
  ---
2
- license: apache-2.0
3
  base_model:
4
  - deepseek-ai/Janus-Pro-7B
 
 
 
5
  ---
 
6
  # 🌟🔥 T2I-R1: Reinforcing Image Generation with Collaborative Semantic-level and Token-level CoT
7
 
8
  Official checkpoint for the paper "[T2I-R1: Reinforcing Image Generation with Collaborative Semantic-level and Token-level CoT](https://arxiv.org/pdf/2505.00703)".
9
 
10
- Paper: https://arxiv.org/pdf/2505.00703
 
11
 
12
  Code: https://github.com/CaraJ7/T2I-R1
13
 
@@ -18,5 +22,5 @@ Please refer to the instructions in GitHub to set up the repository first then r
18
  cd t2i-r1/src/infer
19
  python reason_inference.py \
20
  --model_path YOUR_MODEL_CKPT \
21
- --data_path test_data.txt
22
  ```
 
1
  ---
 
2
  base_model:
3
  - deepseek-ai/Janus-Pro-7B
4
+ license: apache-2.0
5
+ pipeline_tag: text-to-image
6
+ library_name: transformers
7
  ---
8
+
9
  # 🌟🔥 T2I-R1: Reinforcing Image Generation with Collaborative Semantic-level and Token-level CoT
10
 
11
  Official checkpoint for the paper "[T2I-R1: Reinforcing Image Generation with Collaborative Semantic-level and Token-level CoT](https://arxiv.org/pdf/2505.00703)".
12
 
13
+ ## Abstract
14
+ Recent advancements in large language models have demonstrated how chain-of-thought (CoT) and reinforcement learning (RL) can improve performance. However, applying such reasoning strategies to the visual generation domain remains largely unexplored. In this paper, we present T2I-R1, a novel reasoning-enhanced text-to-image generation model, powered by RL with a bi-level CoT reasoning process. Specifically, we identify two levels of CoT that can be utilized to enhance different stages of generation: (1) the semantic-level CoT for high-level planning of the prompt and (2) the token-level CoT for low-level pixel processing during patch-by-patch generation. To better coordinate these two levels of CoT, we introduce BiCoT-GRPO with an ensemble of generation rewards, which seamlessly optimizes both generation CoTs within the same training step. By applying our reasoning strategies to the baseline model, Janus-Pro, we achieve superior performance with 13% improvement on T2I-CompBench and 19% improvement on the WISE benchmark, even surpassing the state-of-the-art model FLUX.1. Code is available at: this https URL
15
 
16
  Code: https://github.com/CaraJ7/T2I-R1
17
 
 
22
  cd t2i-r1/src/infer
23
  python reason_inference.py \
24
  --model_path YOUR_MODEL_CKPT \
25
+ --data_path test_data.txt
26
  ```