LLMJapan commited on
Commit
9733b94
·
verified ·
1 Parent(s): d425cc4

Update README.md

Browse files

Quantization updated

Files changed (1) hide show
  1. README.md +15 -0
README.md CHANGED
@@ -10,6 +10,21 @@ pipeline_tag: text-generation
10
  library_name: transformers
11
  ---
12
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
13
  # Model Card for OlympicCoder-7B
14
 
15
  OlympicCoder-7B is a code model that achieves strong performance on competitive coding benchmarks such as LiveCodeBench and the 2024 International Olympiad in Informatics.
 
10
  library_name: transformers
11
  ---
12
 
13
+ ## Exllama v2 Quantizations of OlympicCoder-7B
14
+
15
+ Using <a href="https://github.com/turboderp/exllamav2/releases/tag/v0.2.8">turboderp's ExLlamaV2 v0.2.8</a> for quantization.
16
+
17
+ average:8.0bpw
18
+ lm_head:8.0bpw
19
+ ```sh
20
+ python convert.py \
21
+ -i {path}/OlympicCoder-7B \
22
+ -o {path}/OlympicCoder-7B/workingdir/ \
23
+ -cf {path}/OlympicCoder-7B_8.0bpw/ \
24
+ -b 8.0 \
25
+ -hb 8
26
+ ```
27
+
28
  # Model Card for OlympicCoder-7B
29
 
30
  OlympicCoder-7B is a code model that achieves strong performance on competitive coding benchmarks such as LiveCodeBench and the 2024 International Olympiad in Informatics.