update model card
Browse files
README.md
CHANGED
|
@@ -11,10 +11,11 @@ language:
|
|
| 11 |
- en
|
| 12 |
base_model: ibm/granite-7b-base
|
| 13 |
---
|
|
|
|
| 14 |
|
| 15 |
### Overview
|
| 16 |
|
| 17 |
-
, we use the taxonomy to drive the sampling process**: For each knowledge/skill, we only use the local examples within the leaf node as seeds to prompt the teacher model.
|
| 56 |
This makes the teacher model better exploit the task distributions defined by the local examples of each node and the diversity in the taxonomy itself ensures the entire generation covers a wide range of tasks, as illustrated below. In turns, this allows for using Mixtral 8x7B as the teacher model for generation while performing very competitively with models such as ORCA-2, WizardLM, and Zephyr Beta that rely on synthetic data generated by much larger and capable models like GPT-4.
|
| 57 |
|
| 58 |
-
 and prompt the model to generate questions and answers based on the document.
|
| 61 |
Foundational skills such as reasoning and compositional skills such as creative writing are generated through in-context learning using the seed examples from the taxonomy.
|
|
@@ -68,7 +69,7 @@ The second step uses replay a replay buffer with data from the first step.
|
|
| 68 |
Both foundational skills and compositional skills are learned during the skills tuning phases, where a replay buffer of data from the knowledge phase is used.
|
| 69 |
Importantly, we use a set of hyper-parameters for training that are very different from standard small-scale supervised fine-training: larger batch size and carefully optimized learning rate and scheduler.
|
| 70 |
|
| 71 |
-

|
| 15 |
|
| 16 |
### Overview
|
| 17 |
|
| 18 |
+

|
| 19 |
|
| 20 |
### Performance
|
| 21 |
|
|
|
|
| 45 |
2. Large-scale synthetic data generator
|
| 46 |
3. Two-phased-training with replay buffers
|
| 47 |
|
| 48 |
+

|
| 49 |
|
| 50 |
LAB approach allows for adding new knowledge and skills, in an incremental fashion, to an already pre-trained model without suffering from catastrophic forgetting.
|
| 51 |
|
| 52 |
Taxonomy is a tree of seed examples that are used to prompt a teacher model to generate synthetic data. Taxonomy allows the data curator or the model designer to easily specify a diverse set of the knowledge-domains and skills that they would like to include in their LLM. At a high level, these can be categorized into three high-level bins - knowledge, foundational skills, and compositional skills. The leaf nodes of the taxonomy are tasks associated with one or more seed examples.
|
| 53 |
|
| 54 |
+

|
| 55 |
|
| 56 |
During the synthetic data generation, **unlike previous approaches where seed examples are uniformly drawn from the entire pool (i.e. self-instruct), we use the taxonomy to drive the sampling process**: For each knowledge/skill, we only use the local examples within the leaf node as seeds to prompt the teacher model.
|
| 57 |
This makes the teacher model better exploit the task distributions defined by the local examples of each node and the diversity in the taxonomy itself ensures the entire generation covers a wide range of tasks, as illustrated below. In turns, this allows for using Mixtral 8x7B as the teacher model for generation while performing very competitively with models such as ORCA-2, WizardLM, and Zephyr Beta that rely on synthetic data generated by much larger and capable models like GPT-4.
|
| 58 |
|
| 59 |
+

|
| 60 |
|
| 61 |
For adding new domain-specific knowledge, we provide an external knowledge source (document) and prompt the model to generate questions and answers based on the document.
|
| 62 |
Foundational skills such as reasoning and compositional skills such as creative writing are generated through in-context learning using the seed examples from the taxonomy.
|
|
|
|
| 69 |
Both foundational skills and compositional skills are learned during the skills tuning phases, where a replay buffer of data from the knowledge phase is used.
|
| 70 |
Importantly, we use a set of hyper-parameters for training that are very different from standard small-scale supervised fine-training: larger batch size and carefully optimized learning rate and scheduler.
|
| 71 |
|
| 72 |
+

|
| 73 |
|
| 74 |
## Model description
|
| 75 |
- **Model Name**: Granite-7b-lab
|