Add pipeline tag, library name, and abstract to model card
Browse filesThis PR improves the model card for T2I-R1 by:
- Adding the `pipeline_tag: text-to-image` to the metadata, which significantly enhances the model's discoverability on the Hugging Face Hub for users filtering by this specific pipeline.
- Adding `library_name: transformers` to the metadata. The presence of `config.json` and a `transformers_version` suggests compatibility with the Hugging Face Transformers library, allowing users to more easily integrate the model.
- Including the paper's abstract in the content section. This provides immediate, comprehensive context about the model's objectives, approach, and key findings directly on the model page, making it more informative for visitors.
- The existing arXiv paper link has been retained as per instructions.
@@ -1,13 +1,17 @@
|
|
1 |
---
|
2 |
-
license: apache-2.0
|
3 |
base_model:
|
4 |
- deepseek-ai/Janus-Pro-7B
|
|
|
|
|
|
|
5 |
---
|
|
|
6 |
# 🌟🔥 T2I-R1: Reinforcing Image Generation with Collaborative Semantic-level and Token-level CoT
|
7 |
|
8 |
Official checkpoint for the paper "[T2I-R1: Reinforcing Image Generation with Collaborative Semantic-level and Token-level CoT](https://arxiv.org/pdf/2505.00703)".
|
9 |
|
10 |
-
|
|
|
11 |
|
12 |
Code: https://github.com/CaraJ7/T2I-R1
|
13 |
|
@@ -18,5 +22,5 @@ Please refer to the instructions in GitHub to set up the repository first then r
|
|
18 |
cd t2i-r1/src/infer
|
19 |
python reason_inference.py \
|
20 |
--model_path YOUR_MODEL_CKPT \
|
21 |
-
--data_path test_data.txt
|
22 |
```
|
|
|
1 |
---
|
|
|
2 |
base_model:
|
3 |
- deepseek-ai/Janus-Pro-7B
|
4 |
+
license: apache-2.0
|
5 |
+
pipeline_tag: text-to-image
|
6 |
+
library_name: transformers
|
7 |
---
|
8 |
+
|
9 |
# 🌟🔥 T2I-R1: Reinforcing Image Generation with Collaborative Semantic-level and Token-level CoT
|
10 |
|
11 |
Official checkpoint for the paper "[T2I-R1: Reinforcing Image Generation with Collaborative Semantic-level and Token-level CoT](https://arxiv.org/pdf/2505.00703)".
|
12 |
|
13 |
+
## Abstract
|
14 |
+
Recent advancements in large language models have demonstrated how chain-of-thought (CoT) and reinforcement learning (RL) can improve performance. However, applying such reasoning strategies to the visual generation domain remains largely unexplored. In this paper, we present T2I-R1, a novel reasoning-enhanced text-to-image generation model, powered by RL with a bi-level CoT reasoning process. Specifically, we identify two levels of CoT that can be utilized to enhance different stages of generation: (1) the semantic-level CoT for high-level planning of the prompt and (2) the token-level CoT for low-level pixel processing during patch-by-patch generation. To better coordinate these two levels of CoT, we introduce BiCoT-GRPO with an ensemble of generation rewards, which seamlessly optimizes both generation CoTs within the same training step. By applying our reasoning strategies to the baseline model, Janus-Pro, we achieve superior performance with 13% improvement on T2I-CompBench and 19% improvement on the WISE benchmark, even surpassing the state-of-the-art model FLUX.1. Code is available at: this https URL
|
15 |
|
16 |
Code: https://github.com/CaraJ7/T2I-R1
|
17 |
|
|
|
22 |
cd t2i-r1/src/infer
|
23 |
python reason_inference.py \
|
24 |
--model_path YOUR_MODEL_CKPT \
|
25 |
+
--data_path test_data.txt
|
26 |
```
|