Chanukya Patnaik
commited on
Commit
·
d58c62e
1
Parent(s):
5feaf84
updated readme
Browse files
README.md
CHANGED
|
@@ -1,6 +1,9 @@
|
|
| 1 |
---
|
| 2 |
library_name: peft
|
| 3 |
---
|
|
|
|
|
|
|
|
|
|
| 4 |
## Training procedure
|
| 5 |
|
| 6 |
|
|
@@ -18,3 +21,4 @@ The following `bitsandbytes` quantization config was used during training:
|
|
| 18 |
|
| 19 |
|
| 20 |
- PEFT 0.5.0.dev0
|
|
|
|
|
|
| 1 |
---
|
| 2 |
library_name: peft
|
| 3 |
---
|
| 4 |
+
effi 7b is a 7 billion parameter model built by AI Planet. Inspired by llama, we've built fine-tuned version of llama7b with qlora. The training procedure and framework versions are provided below along with model weighths.
|
| 5 |
+
|
| 6 |
+
|
| 7 |
## Training procedure
|
| 8 |
|
| 9 |
|
|
|
|
| 21 |
|
| 22 |
|
| 23 |
- PEFT 0.5.0.dev0
|
| 24 |
+
|