Trained with Unsloth
Browse files
pytorch_model-00001-of-00002.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9db1d62d20b1fb85148fc5a288e47ca4d707283f3479ab9d5561146aaa03d747
|
| 3 |
+
size 4983825789
|
pytorch_model-00002-of-00002.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:169df8ba9ccb4e93588b382233f8d7daecf5567cbc7f89170f8a307d6142be57
|
| 3 |
+
size 1278363922
|
pytorch_model.bin.index.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
| 1 |
{
|
| 2 |
"metadata": {
|
| 3 |
-
"total_size":
|
| 4 |
},
|
| 5 |
"weight_map": {
|
| 6 |
"model.embed_tokens.weight": "pytorch_model-00001-of-00002.bin",
|
|
@@ -244,11 +244,11 @@
|
|
| 244 |
"model.layers.26.self_attn.q_proj.weight": "pytorch_model-00001-of-00002.bin",
|
| 245 |
"model.layers.26.self_attn.v_proj.bias": "pytorch_model-00001-of-00002.bin",
|
| 246 |
"model.layers.26.self_attn.v_proj.weight": "pytorch_model-00001-of-00002.bin",
|
| 247 |
-
"model.layers.27.input_layernorm.weight": "pytorch_model-
|
| 248 |
-
"model.layers.27.mlp.down_proj.weight": "pytorch_model-
|
| 249 |
"model.layers.27.mlp.gate_proj.weight": "pytorch_model-00001-of-00002.bin",
|
| 250 |
"model.layers.27.mlp.up_proj.weight": "pytorch_model-00001-of-00002.bin",
|
| 251 |
-
"model.layers.27.post_attention_layernorm.weight": "pytorch_model-
|
| 252 |
"model.layers.27.self_attn.k_proj.bias": "pytorch_model-00001-of-00002.bin",
|
| 253 |
"model.layers.27.self_attn.k_proj.weight": "pytorch_model-00001-of-00002.bin",
|
| 254 |
"model.layers.27.self_attn.o_proj.weight": "pytorch_model-00001-of-00002.bin",
|
|
@@ -261,13 +261,13 @@
|
|
| 261 |
"model.layers.28.mlp.gate_proj.weight": "pytorch_model-00002-of-00002.bin",
|
| 262 |
"model.layers.28.mlp.up_proj.weight": "pytorch_model-00002-of-00002.bin",
|
| 263 |
"model.layers.28.post_attention_layernorm.weight": "pytorch_model-00002-of-00002.bin",
|
| 264 |
-
"model.layers.28.self_attn.k_proj.bias": "pytorch_model-
|
| 265 |
-
"model.layers.28.self_attn.k_proj.weight": "pytorch_model-
|
| 266 |
-
"model.layers.28.self_attn.o_proj.weight": "pytorch_model-
|
| 267 |
-
"model.layers.28.self_attn.q_proj.bias": "pytorch_model-
|
| 268 |
-
"model.layers.28.self_attn.q_proj.weight": "pytorch_model-
|
| 269 |
-
"model.layers.28.self_attn.v_proj.bias": "pytorch_model-
|
| 270 |
-
"model.layers.28.self_attn.v_proj.weight": "pytorch_model-
|
| 271 |
"model.layers.29.input_layernorm.weight": "pytorch_model-00002-of-00002.bin",
|
| 272 |
"model.layers.29.mlp.down_proj.weight": "pytorch_model-00002-of-00002.bin",
|
| 273 |
"model.layers.29.mlp.gate_proj.weight": "pytorch_model-00002-of-00002.bin",
|
|
|
|
| 1 |
{
|
| 2 |
"metadata": {
|
| 3 |
+
"total_size": 6262054912
|
| 4 |
},
|
| 5 |
"weight_map": {
|
| 6 |
"model.embed_tokens.weight": "pytorch_model-00001-of-00002.bin",
|
|
|
|
| 244 |
"model.layers.26.self_attn.q_proj.weight": "pytorch_model-00001-of-00002.bin",
|
| 245 |
"model.layers.26.self_attn.v_proj.bias": "pytorch_model-00001-of-00002.bin",
|
| 246 |
"model.layers.26.self_attn.v_proj.weight": "pytorch_model-00001-of-00002.bin",
|
| 247 |
+
"model.layers.27.input_layernorm.weight": "pytorch_model-00002-of-00002.bin",
|
| 248 |
+
"model.layers.27.mlp.down_proj.weight": "pytorch_model-00002-of-00002.bin",
|
| 249 |
"model.layers.27.mlp.gate_proj.weight": "pytorch_model-00001-of-00002.bin",
|
| 250 |
"model.layers.27.mlp.up_proj.weight": "pytorch_model-00001-of-00002.bin",
|
| 251 |
+
"model.layers.27.post_attention_layernorm.weight": "pytorch_model-00002-of-00002.bin",
|
| 252 |
"model.layers.27.self_attn.k_proj.bias": "pytorch_model-00001-of-00002.bin",
|
| 253 |
"model.layers.27.self_attn.k_proj.weight": "pytorch_model-00001-of-00002.bin",
|
| 254 |
"model.layers.27.self_attn.o_proj.weight": "pytorch_model-00001-of-00002.bin",
|
|
|
|
| 261 |
"model.layers.28.mlp.gate_proj.weight": "pytorch_model-00002-of-00002.bin",
|
| 262 |
"model.layers.28.mlp.up_proj.weight": "pytorch_model-00002-of-00002.bin",
|
| 263 |
"model.layers.28.post_attention_layernorm.weight": "pytorch_model-00002-of-00002.bin",
|
| 264 |
+
"model.layers.28.self_attn.k_proj.bias": "pytorch_model-00002-of-00002.bin",
|
| 265 |
+
"model.layers.28.self_attn.k_proj.weight": "pytorch_model-00002-of-00002.bin",
|
| 266 |
+
"model.layers.28.self_attn.o_proj.weight": "pytorch_model-00002-of-00002.bin",
|
| 267 |
+
"model.layers.28.self_attn.q_proj.bias": "pytorch_model-00002-of-00002.bin",
|
| 268 |
+
"model.layers.28.self_attn.q_proj.weight": "pytorch_model-00002-of-00002.bin",
|
| 269 |
+
"model.layers.28.self_attn.v_proj.bias": "pytorch_model-00002-of-00002.bin",
|
| 270 |
+
"model.layers.28.self_attn.v_proj.weight": "pytorch_model-00002-of-00002.bin",
|
| 271 |
"model.layers.29.input_layernorm.weight": "pytorch_model-00002-of-00002.bin",
|
| 272 |
"model.layers.29.mlp.down_proj.weight": "pytorch_model-00002-of-00002.bin",
|
| 273 |
"model.layers.29.mlp.gate_proj.weight": "pytorch_model-00002-of-00002.bin",
|