Add missing metadata: library_name, license, and pipeline_tag
Browse filesThis PR adds the missing `library_name`, `license`, and `pipeline_tag` metadata to the model card. The model utilizes the Transformers library, is licensed under Apache 2.0, and performs text generation, which is reflected in the updated metadata. This improves the model card's completeness and discoverability on the Hugging Face Hub.
README.md
CHANGED
|
@@ -1,6 +1,9 @@
|
|
| 1 |
---
|
| 2 |
-
|
|
|
|
|
|
|
| 3 |
---
|
|
|
|
| 4 |
<h1 align="center">
|
| 5 |
<em>AReaL</em>: Ant Reasoning Reinforcement Learning for LLMs
|
| 6 |
</h1>
|
|
@@ -37,7 +40,7 @@ In our AReaL-boba² (A-ReaL-double-boba) release, we highlight the top 3 most im
|
|
| 37 |
|
| 38 |
+ Experimental support for **multi-turn** agentic RL training. Check our [complete example](https://inclusionai.github.io/AReaL/customization/agent.html).
|
| 39 |
|
| 40 |
-
For the complete system design and more training details, please check [our v0.3 blog](/blog/AReaL_v0_3.md) and our [research paper](
|
| 41 |
|
| 42 |
### Overview of Asynchronous RL Training
|
| 43 |
|
|
@@ -100,8 +103,6 @@ AReaL-boba² allows you to independently customize the [dataset](https://inclusi
|
|
| 100 |
|
| 101 |
In particular, we show a simple example to develop a multi-turn math agent for RL training. Please see the learning curve below and reference the [step-by-step guide](https://inclusionai.github.io/AReaL/customization/agent.html) if you want to implement your own agentic RL project.
|
| 102 |
|
| 103 |
-
**Multi-turn Agent Learning Curve**
|
| 104 |
-
|
| 105 |
## Getting Started
|
| 106 |
|
| 107 |
### Quick Start
|
|
@@ -214,4 +215,4 @@ We also appreciate all the pioneering works from the community, particularly the
|
|
| 214 |
primaryClass={cs.LG},
|
| 215 |
url={https://arxiv.org/abs/2505.24298},
|
| 216 |
}
|
| 217 |
-
```
|
|
|
|
| 1 |
---
|
| 2 |
+
license: apache-2.0
|
| 3 |
+
library_name: transformers
|
| 4 |
+
pipeline_tag: text-generation
|
| 5 |
---
|
| 6 |
+
|
| 7 |
<h1 align="center">
|
| 8 |
<em>AReaL</em>: Ant Reasoning Reinforcement Learning for LLMs
|
| 9 |
</h1>
|
|
|
|
| 40 |
|
| 41 |
+ Experimental support for **multi-turn** agentic RL training. Check our [complete example](https://inclusionai.github.io/AReaL/customization/agent.html).
|
| 42 |
|
| 43 |
+
For the complete system design and more training details, please check [our v0.3 blog](/blog/AReaL_v0_3.md) and our [research paper](https://arxiv.org/pdf/2505.24298).
|
| 44 |
|
| 45 |
### Overview of Asynchronous RL Training
|
| 46 |
|
|
|
|
| 103 |
|
| 104 |
In particular, we show a simple example to develop a multi-turn math agent for RL training. Please see the learning curve below and reference the [step-by-step guide](https://inclusionai.github.io/AReaL/customization/agent.html) if you want to implement your own agentic RL project.
|
| 105 |
|
|
|
|
|
|
|
| 106 |
## Getting Started
|
| 107 |
|
| 108 |
### Quick Start
|
|
|
|
| 215 |
primaryClass={cs.LG},
|
| 216 |
url={https://arxiv.org/abs/2505.24298},
|
| 217 |
}
|
| 218 |
+
```
|