lonestar108 commited on
Commit
935cf25
·
1 Parent(s): 9fb36cd

Upload model

Browse files
Files changed (3) hide show
  1. README.md +24 -0
  2. adapter_config.json +1 -1
  3. adapter_model.bin +1 -1
README.md CHANGED
@@ -37,6 +37,28 @@ The following `bitsandbytes` quantization config was used during training:
37
  - bnb_4bit_use_double_quant: False
38
  - bnb_4bit_compute_dtype: float16
39
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
40
  The following `bitsandbytes` quantization config was used during training:
41
  - load_in_8bit: False
42
  - load_in_4bit: True
@@ -49,6 +71,8 @@ The following `bitsandbytes` quantization config was used during training:
49
  - bnb_4bit_compute_dtype: float16
50
  ### Framework versions
51
 
 
 
52
  - PEFT 0.4.0
53
  - PEFT 0.4.0
54
  - PEFT 0.4.0
 
37
  - bnb_4bit_use_double_quant: False
38
  - bnb_4bit_compute_dtype: float16
39
 
40
+ The following `bitsandbytes` quantization config was used during training:
41
+ - load_in_8bit: False
42
+ - load_in_4bit: True
43
+ - llm_int8_threshold: 6.0
44
+ - llm_int8_skip_modules: None
45
+ - llm_int8_enable_fp32_cpu_offload: False
46
+ - llm_int8_has_fp16_weight: False
47
+ - bnb_4bit_quant_type: nf4
48
+ - bnb_4bit_use_double_quant: False
49
+ - bnb_4bit_compute_dtype: float16
50
+
51
+ The following `bitsandbytes` quantization config was used during training:
52
+ - load_in_8bit: False
53
+ - load_in_4bit: True
54
+ - llm_int8_threshold: 6.0
55
+ - llm_int8_skip_modules: None
56
+ - llm_int8_enable_fp32_cpu_offload: False
57
+ - llm_int8_has_fp16_weight: False
58
+ - bnb_4bit_quant_type: nf4
59
+ - bnb_4bit_use_double_quant: False
60
+ - bnb_4bit_compute_dtype: float16
61
+
62
  The following `bitsandbytes` quantization config was used during training:
63
  - load_in_8bit: False
64
  - load_in_4bit: True
 
71
  - bnb_4bit_compute_dtype: float16
72
  ### Framework versions
73
 
74
+ - PEFT 0.4.0
75
+ - PEFT 0.4.0
76
  - PEFT 0.4.0
77
  - PEFT 0.4.0
78
  - PEFT 0.4.0
adapter_config.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "auto_mapping": null,
3
- "base_model_name_or_path": "mrm8488/llama-2-coder-7b",
4
  "bias": "none",
5
  "fan_in_fan_out": false,
6
  "inference_mode": true,
 
1
  {
2
  "auto_mapping": null,
3
+ "base_model_name_or_path": "dwitter",
4
  "bias": "none",
5
  "fan_in_fan_out": false,
6
  "inference_mode": true,
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4582a7c73f19baa2831a553409683339b8e28050ec18bee17b2b76a929b2b187
3
  size 134263757
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c78a6e4fccad57d26c69d09770166d2b6b26bcd31c4adb2886dedc1b86b26635
3
  size 134263757