Update README.md
Browse files
README.md
CHANGED
|
@@ -1,199 +1,108 @@
|
|
| 1 |
---
|
| 2 |
library_name: transformers
|
| 3 |
-
tags:
|
|
|
|
|
|
|
|
|
|
|
|
|
| 4 |
---
|
| 5 |
|
| 6 |
-
# Model Card for
|
| 7 |
-
|
| 8 |
-
<!-- Provide a quick summary of what the model is/does. -->
|
| 9 |
-
|
| 10 |
|
|
|
|
| 11 |
|
| 12 |
## Model Details
|
| 13 |
|
| 14 |
### Model Description
|
| 15 |
|
| 16 |
-
|
| 17 |
-
|
| 18 |
-
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
|
| 19 |
|
| 20 |
-
- **Developed by:**
|
| 21 |
-
- **Funded by [optional]:**
|
| 22 |
-
- **Shared by [optional]:**
|
| 23 |
-
- **Model type:**
|
| 24 |
-
- **Language(s) (NLP):**
|
| 25 |
-
- **License:**
|
| 26 |
-
- **Finetuned from model [optional]:**
|
| 27 |
|
| 28 |
-
### Model Sources
|
| 29 |
|
| 30 |
-
|
| 31 |
-
|
| 32 |
-
- **Repository:** [More Information Needed]
|
| 33 |
-
- **Paper [optional]:** [More Information Needed]
|
| 34 |
-
- **Demo [optional]:** [More Information Needed]
|
| 35 |
|
| 36 |
## Uses
|
| 37 |
|
| 38 |
-
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
|
| 39 |
-
|
| 40 |
### Direct Use
|
| 41 |
|
| 42 |
-
|
| 43 |
-
|
| 44 |
-
[More Information Needed]
|
| 45 |
|
| 46 |
### Downstream Use [optional]
|
| 47 |
|
| 48 |
-
|
| 49 |
-
|
| 50 |
-
[More Information Needed]
|
| 51 |
|
| 52 |
### Out-of-Scope Use
|
| 53 |
|
| 54 |
-
|
| 55 |
-
|
| 56 |
-
[More Information Needed]
|
| 57 |
|
| 58 |
## Bias, Risks, and Limitations
|
| 59 |
|
| 60 |
-
|
| 61 |
-
|
| 62 |
-
[More Information Needed]
|
| 63 |
|
| 64 |
### Recommendations
|
| 65 |
|
| 66 |
-
|
| 67 |
-
|
| 68 |
-
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
|
| 69 |
|
| 70 |
## How to Get Started with the Model
|
| 71 |
|
| 72 |
-
Use the code below to get started with the model
|
| 73 |
-
|
| 74 |
-
|
| 75 |
-
|
| 76 |
-
|
| 77 |
-
|
| 78 |
-
|
| 79 |
-
|
| 80 |
-
|
| 81 |
-
|
| 82 |
-
|
| 83 |
-
|
| 84 |
-
|
| 85 |
-
|
| 86 |
-
|
| 87 |
-
|
| 88 |
-
|
| 89 |
-
|
| 90 |
-
|
| 91 |
-
|
| 92 |
-
|
| 93 |
-
|
| 94 |
-
|
| 95 |
-
|
| 96 |
-
|
| 97 |
-
|
| 98 |
-
|
| 99 |
-
|
| 100 |
-
|
| 101 |
-
|
| 102 |
-
|
| 103 |
-
|
| 104 |
-
|
| 105 |
-
|
| 106 |
-
|
| 107 |
-
|
| 108 |
-
|
| 109 |
-
|
| 110 |
-
|
| 111 |
-
|
| 112 |
-
|
| 113 |
-
|
| 114 |
-
|
| 115 |
-
|
| 116 |
-
|
| 117 |
-
|
| 118 |
-
|
| 119 |
-
|
| 120 |
-
|
| 121 |
-
|
| 122 |
-
|
| 123 |
-
|
| 124 |
-
|
| 125 |
-
[More Information Needed]
|
| 126 |
-
|
| 127 |
-
### Results
|
| 128 |
-
|
| 129 |
-
[More Information Needed]
|
| 130 |
-
|
| 131 |
-
#### Summary
|
| 132 |
-
|
| 133 |
-
|
| 134 |
-
|
| 135 |
-
## Model Examination [optional]
|
| 136 |
-
|
| 137 |
-
<!-- Relevant interpretability work for the model goes here -->
|
| 138 |
-
|
| 139 |
-
[More Information Needed]
|
| 140 |
-
|
| 141 |
-
## Environmental Impact
|
| 142 |
-
|
| 143 |
-
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
|
| 144 |
-
|
| 145 |
-
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
|
| 146 |
-
|
| 147 |
-
- **Hardware Type:** [More Information Needed]
|
| 148 |
-
- **Hours used:** [More Information Needed]
|
| 149 |
-
- **Cloud Provider:** [More Information Needed]
|
| 150 |
-
- **Compute Region:** [More Information Needed]
|
| 151 |
-
- **Carbon Emitted:** [More Information Needed]
|
| 152 |
-
|
| 153 |
-
## Technical Specifications [optional]
|
| 154 |
-
|
| 155 |
-
### Model Architecture and Objective
|
| 156 |
-
|
| 157 |
-
[More Information Needed]
|
| 158 |
-
|
| 159 |
-
### Compute Infrastructure
|
| 160 |
-
|
| 161 |
-
[More Information Needed]
|
| 162 |
-
|
| 163 |
-
#### Hardware
|
| 164 |
-
|
| 165 |
-
[More Information Needed]
|
| 166 |
-
|
| 167 |
-
#### Software
|
| 168 |
-
|
| 169 |
-
[More Information Needed]
|
| 170 |
-
|
| 171 |
-
## Citation [optional]
|
| 172 |
-
|
| 173 |
-
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
|
| 174 |
-
|
| 175 |
-
**BibTeX:**
|
| 176 |
-
|
| 177 |
-
[More Information Needed]
|
| 178 |
-
|
| 179 |
-
**APA:**
|
| 180 |
-
|
| 181 |
-
[More Information Needed]
|
| 182 |
-
|
| 183 |
-
## Glossary [optional]
|
| 184 |
-
|
| 185 |
-
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
|
| 186 |
-
|
| 187 |
-
[More Information Needed]
|
| 188 |
-
|
| 189 |
-
## More Information [optional]
|
| 190 |
-
|
| 191 |
-
[More Information Needed]
|
| 192 |
-
|
| 193 |
-
## Model Card Authors [optional]
|
| 194 |
-
|
| 195 |
-
[More Information Needed]
|
| 196 |
-
|
| 197 |
-
## Model Card Contact
|
| 198 |
-
|
| 199 |
-
[More Information Needed]
|
|
|
|
| 1 |
---
|
| 2 |
library_name: transformers
|
| 3 |
+
tags:
|
| 4 |
+
- dolly-v2
|
| 5 |
+
- instruction-tuning
|
| 6 |
+
- peft
|
| 7 |
+
- lora
|
| 8 |
---
|
| 9 |
|
| 10 |
+
# Model Card for dolly-3b-lora
|
|
|
|
|
|
|
|
|
|
| 11 |
|
| 12 |
+
This model is a fine-tuned version of the Dolly V2 3B language model, enhanced with Parameter-Efficient Fine-Tuning (PEFT) using Low-Rank Adaptation (LoRA). It was fine-tuned on the LaMini-instruction dataset to improve its ability to follow instructions and generate coherent responses for various tasks.
|
| 13 |
|
| 14 |
## Model Details
|
| 15 |
|
| 16 |
### Model Description
|
| 17 |
|
| 18 |
+
This is a fine-tuned version of the `databricks/dolly-v2-3b` model, adapted using LoRA on the LaMini-instruction dataset. The model is designed for instruction-following tasks, leveraging the efficiency of LoRA to fine-tune approximately 2.93% of the total parameters while maintaining performance. It supports text generation tasks and has been optimized for use on GPU hardware with 8-bit quantization, with a fallback to CPU if needed.
|
|
|
|
|
|
|
| 19 |
|
| 20 |
+
- **Developed by:** avinashhm
|
| 21 |
+
- **Funded by [optional]:** Not specified
|
| 22 |
+
- **Shared by [optional]:** avinashhm
|
| 23 |
+
- **Model type:** Causal Language Model
|
| 24 |
+
- **Language(s) (NLP):** English
|
| 25 |
+
- **License:** Apache-2.0
|
| 26 |
+
- **Finetuned from model [optional]:** databricks/dolly-v2-3b
|
| 27 |
|
| 28 |
+
### Model Sources
|
| 29 |
|
| 30 |
+
- **Repository:** https://huggingface.co/avinashhm/dolly-3b-lora
|
|
|
|
|
|
|
|
|
|
|
|
|
| 31 |
|
| 32 |
## Uses
|
| 33 |
|
|
|
|
|
|
|
| 34 |
### Direct Use
|
| 35 |
|
| 36 |
+
The model is intended for direct use in text generation tasks, particularly for instruction-following scenarios such as answering questions, generating lists, or writing short narratives. It can be used by developers, researchers, or hobbyists working on natural language processing applications.
|
|
|
|
|
|
|
| 37 |
|
| 38 |
### Downstream Use [optional]
|
| 39 |
|
| 40 |
+
The model can be further fine-tuned for specific tasks, such as chatbots, virtual assistants, or specialized text generation applications. It can be integrated into larger ecosystems requiring instruction-based text generation.
|
|
|
|
|
|
|
| 41 |
|
| 42 |
### Out-of-Scope Use
|
| 43 |
|
| 44 |
+
The model is not designed for real-time, safety-critical applications or tasks requiring factual accuracy without verification, as it may generate incorrect or biased responses. It should not be used for malicious purposes, such as generating harmful content or misinformation.
|
|
|
|
|
|
|
| 45 |
|
| 46 |
## Bias, Risks, and Limitations
|
| 47 |
|
| 48 |
+
The model inherits biases from the LaMini-instruction dataset and the base Dolly V2 3B model. It may produce biased, incomplete, or factually incorrect responses, particularly for sensitive topics. Performance is limited by the small fine-tuning dataset (200 samples) and LoRA configuration, which may not generalize well to all instruction types. Some responses may lack depth or coherence for complex tasks due to limited training data and epochs.
|
|
|
|
|
|
|
| 49 |
|
| 50 |
### Recommendations
|
| 51 |
|
| 52 |
+
Users should verify outputs for accuracy and appropriateness, especially in sensitive applications. Further fine-tuning with a larger, more diverse dataset could improve performance and reduce biases. Caution is advised when deploying in public-facing applications to avoid unintended consequences from biased or harmful outputs.
|
|
|
|
|
|
|
| 53 |
|
| 54 |
## How to Get Started with the Model
|
| 55 |
|
| 56 |
+
Use the code below to get started with the model:
|
| 57 |
+
|
| 58 |
+
```python
|
| 59 |
+
import torch
|
| 60 |
+
from transformers import AutoTokenizer, AutoModelForCausalLM, pipeline
|
| 61 |
+
from peft import PeftModel
|
| 62 |
+
|
| 63 |
+
# Model names
|
| 64 |
+
base_model_name = "databricks/dolly-v2-3b"
|
| 65 |
+
peft_model_name = "avinashhm/dolly-3b-lora"
|
| 66 |
+
|
| 67 |
+
# Load Tokenizer
|
| 68 |
+
tokenizer = AutoTokenizer.from_pretrained(base_model_name)
|
| 69 |
+
|
| 70 |
+
# Load Base Model
|
| 71 |
+
base_model = AutoModelForCausalLM.from_pretrained(
|
| 72 |
+
base_model_name,
|
| 73 |
+
torch_dtype=torch.float16,
|
| 74 |
+
device_map="auto"
|
| 75 |
+
)
|
| 76 |
+
|
| 77 |
+
# Load PEFT (LoRA) Adapter
|
| 78 |
+
model = PeftModel.from_pretrained(
|
| 79 |
+
base_model,
|
| 80 |
+
peft_model_name,
|
| 81 |
+
torch_dtype=torch.float16
|
| 82 |
+
)
|
| 83 |
+
|
| 84 |
+
# Merge adapter weights into base model (optional, improves speed)
|
| 85 |
+
model = model.merge_and_unload()
|
| 86 |
+
|
| 87 |
+
# Define prompt template
|
| 88 |
+
prompt_template = """Below is an instruction that describes a task. Write a response that appropriately completes the request. Instruction: {instruction}\n Response:"""
|
| 89 |
+
|
| 90 |
+
# Create Text Generation Pipeline
|
| 91 |
+
inf_pipeline = pipeline(
|
| 92 |
+
"text-generation",
|
| 93 |
+
model=model,
|
| 94 |
+
tokenizer=tokenizer,
|
| 95 |
+
max_new_tokens=256,
|
| 96 |
+
pad_token_id=tokenizer.eos_token_id,
|
| 97 |
+
truncation=True,
|
| 98 |
+
do_sample=True,
|
| 99 |
+
temperature=0.7,
|
| 100 |
+
top_p=0.95,
|
| 101 |
+
repetition_penalty=1.1
|
| 102 |
+
)
|
| 103 |
+
|
| 104 |
+
# Example prompt
|
| 105 |
+
prompt = "List 5 reasons why someone should learn to cook."
|
| 106 |
+
formatted_prompt = prompt_template.format(instruction=prompt)
|
| 107 |
+
response = inf_pipeline(formatted_prompt)[0]['generated_text'].split(" Response:")[-1].strip()
|
| 108 |
+
print(response)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|