Update README.md
Browse files
README.md
CHANGED
|
@@ -6,4 +6,5 @@ license: apache-2.0
|
|
| 6 |
Dataset used is uncencored and filtered of Psmathur's wizard orca 72k instructions, ported into a alpaca chat format, and removed all "Input" strings. Trained on only instruction and output.
|
| 7 |
Preforms remarkably well works with the blokes gptq mini orca v2 gptq and allows the model to output 2048 tokens.
|
| 8 |
This adapter was trained to test how lora training could be used to expand context window with less compute.
|
| 9 |
-
The original model trained by "psmathur" was trained with "8x A100(80G) GPUs for around 13 Hours for cost of $195".
|
|
|
|
|
|
| 6 |
Dataset used is uncencored and filtered of Psmathur's wizard orca 72k instructions, ported into a alpaca chat format, and removed all "Input" strings. Trained on only instruction and output.
|
| 7 |
Preforms remarkably well works with the blokes gptq mini orca v2 gptq and allows the model to output 2048 tokens.
|
| 8 |
This adapter was trained to test how lora training could be used to expand context window with less compute.
|
| 9 |
+
The original model trained by "psmathur" was trained with "8x A100(80G) GPUs for around 13 Hours for cost of $195".
|
| 10 |
+
I'm hoping the bloke can merge this with the orignal model and upload a gptq version for everyone.
|