Vezora commited on
Commit
ff8bd7d
·
1 Parent(s): fd71b6e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -1
README.md CHANGED
@@ -6,4 +6,5 @@ license: apache-2.0
6
  Dataset used is uncencored and filtered of Psmathur's wizard orca 72k instructions, ported into a alpaca chat format, and removed all "Input" strings. Trained on only instruction and output.
7
  Preforms remarkably well works with the blokes gptq mini orca v2 gptq and allows the model to output 2048 tokens.
8
  This adapter was trained to test how lora training could be used to expand context window with less compute.
9
- The original model trained by "psmathur" was trained with "8x A100(80G) GPUs for around 13 Hours for cost of $195".
 
 
6
  Dataset used is uncencored and filtered of Psmathur's wizard orca 72k instructions, ported into a alpaca chat format, and removed all "Input" strings. Trained on only instruction and output.
7
  Preforms remarkably well works with the blokes gptq mini orca v2 gptq and allows the model to output 2048 tokens.
8
  This adapter was trained to test how lora training could be used to expand context window with less compute.
9
+ The original model trained by "psmathur" was trained with "8x A100(80G) GPUs for around 13 Hours for cost of $195".
10
+ I'm hoping the bloke can merge this with the orignal model and upload a gptq version for everyone.