End of training
Browse files- README.md +31 -31
 - config.json +38 -18
 - final_checkpoint/config.json +38 -18
 - final_checkpoint/generation_config.json +3 -3
 - final_checkpoint/model-00001-of-00003.safetensors +2 -2
 - final_checkpoint/model-00002-of-00003.safetensors +2 -2
 - generation_config.json +3 -3
 - model-00001-of-00003.safetensors +2 -2
 - model-00002-of-00003.safetensors +2 -2
 - model-00003-of-00003.safetensors +2 -2
 - model.safetensors.index.json +195 -292
 - special_tokens_map.json +4 -4
 - tokenizer.json +0 -0
 - tokenizer_config.json +190 -21
 - training_args.bin +1 -1
 
    	
        README.md
    CHANGED
    
    | 
         @@ -1,6 +1,6 @@ 
     | 
|
| 1 | 
         
             
            ---
         
     | 
| 2 | 
         
             
            license: apache-2.0
         
     | 
| 3 | 
         
            -
            base_model:  
     | 
| 4 | 
         
             
            tags:
         
     | 
| 5 | 
         
             
            - trl
         
     | 
| 6 | 
         
             
            - dpo
         
     | 
| 
         @@ -15,17 +15,17 @@ should probably proofread and complete it, then remove this comment. --> 
     | 
|
| 15 | 
         | 
| 16 | 
         
             
            # v1_1000_STEPS_1e6_rate_03_beta_DPO
         
     | 
| 17 | 
         | 
| 18 | 
         
            -
            This model is a fine-tuned version of [ 
     | 
| 19 | 
         
             
            It achieves the following results on the evaluation set:
         
     | 
| 20 | 
         
            -
            - Loss: 0. 
     | 
| 21 | 
         
            -
            - Rewards/chosen: - 
     | 
| 22 | 
         
            -
            - Rewards/rejected: - 
     | 
| 23 | 
         
            -
            - Rewards/accuracies: 0. 
     | 
| 24 | 
         
            -
            - Rewards/margins: 0. 
     | 
| 25 | 
         
            -
            - Logps/rejected: - 
     | 
| 26 | 
         
            -
            - Logps/chosen: - 
     | 
| 27 | 
         
            -
            - Logits/rejected:  
     | 
| 28 | 
         
            -
            - Logits/chosen:  
     | 
| 29 | 
         | 
| 30 | 
         
             
            ## Model description
         
     | 
| 31 | 
         | 
| 
         @@ -59,26 +59,26 @@ The following hyperparameters were used during training: 
     | 
|
| 59 | 
         | 
| 60 | 
         
             
            | Training Loss | Epoch | Step | Validation Loss | Rewards/chosen | Rewards/rejected | Rewards/accuracies | Rewards/margins | Logps/rejected | Logps/chosen | Logits/rejected | Logits/chosen |
         
     | 
| 61 | 
         
             
            |:-------------:|:-----:|:----:|:---------------:|:--------------:|:----------------:|:------------------:|:---------------:|:--------------:|:------------:|:---------------:|:-------------:|
         
     | 
| 62 | 
         
            -
            | 0. 
     | 
| 63 | 
         
            -
            | 0. 
     | 
| 64 | 
         
            -
            |  
     | 
| 65 | 
         
            -
            | 0. 
     | 
| 66 | 
         
            -
            | 0. 
     | 
| 67 | 
         
            -
            | 0. 
     | 
| 68 | 
         
            -
            |  
     | 
| 69 | 
         
            -
            |  
     | 
| 70 | 
         
            -
            | 0. 
     | 
| 71 | 
         
            -
            | 0. 
     | 
| 72 | 
         
            -
            | 0. 
     | 
| 73 | 
         
            -
            | 0. 
     | 
| 74 | 
         
            -
            | 0. 
     | 
| 75 | 
         
            -
            | 0. 
     | 
| 76 | 
         
            -
            | 0. 
     | 
| 77 | 
         
            -
            | 0. 
     | 
| 78 | 
         
            -
            |  
     | 
| 79 | 
         
            -
            |  
     | 
| 80 | 
         
            -
            | 0. 
     | 
| 81 | 
         
            -
            | 0. 
     | 
| 82 | 
         | 
| 83 | 
         | 
| 84 | 
         
             
            ### Framework versions
         
     | 
| 
         | 
|
| 1 | 
         
             
            ---
         
     | 
| 2 | 
         
             
            license: apache-2.0
         
     | 
| 3 | 
         
            +
            base_model: mosaicml/mpt-7b-instruct
         
     | 
| 4 | 
         
             
            tags:
         
     | 
| 5 | 
         
             
            - trl
         
     | 
| 6 | 
         
             
            - dpo
         
     | 
| 
         | 
|
| 15 | 
         | 
| 16 | 
         
             
            # v1_1000_STEPS_1e6_rate_03_beta_DPO
         
     | 
| 17 | 
         | 
| 18 | 
         
            +
            This model is a fine-tuned version of [mosaicml/mpt-7b-instruct](https://huggingface.co/mosaicml/mpt-7b-instruct) on an unknown dataset.
         
     | 
| 19 | 
         
             
            It achieves the following results on the evaluation set:
         
     | 
| 20 | 
         
            +
            - Loss: 0.6641
         
     | 
| 21 | 
         
            +
            - Rewards/chosen: -1.4066
         
     | 
| 22 | 
         
            +
            - Rewards/rejected: -1.6576
         
     | 
| 23 | 
         
            +
            - Rewards/accuracies: 0.6198
         
     | 
| 24 | 
         
            +
            - Rewards/margins: 0.2510
         
     | 
| 25 | 
         
            +
            - Logps/rejected: -27.0829
         
     | 
| 26 | 
         
            +
            - Logps/chosen: -25.4808
         
     | 
| 27 | 
         
            +
            - Logits/rejected: 13.3887
         
     | 
| 28 | 
         
            +
            - Logits/chosen: 13.3921
         
     | 
| 29 | 
         | 
| 30 | 
         
             
            ## Model description
         
     | 
| 31 | 
         | 
| 
         | 
|
| 59 | 
         | 
| 60 | 
         
             
            | Training Loss | Epoch | Step | Validation Loss | Rewards/chosen | Rewards/rejected | Rewards/accuracies | Rewards/margins | Logps/rejected | Logps/chosen | Logits/rejected | Logits/chosen |
         
     | 
| 61 | 
         
             
            |:-------------:|:-----:|:----:|:---------------:|:--------------:|:----------------:|:------------------:|:---------------:|:--------------:|:------------:|:---------------:|:-------------:|
         
     | 
| 62 | 
         
            +
            | 0.6901        | 0.05  | 50   | 0.6931          | 0.0510         | 0.0490           | 0.5253             | 0.0019          | -21.3940       | -20.6223     | 14.3181         | 14.3207       |
         
     | 
| 63 | 
         
            +
            | 0.7257        | 0.1   | 100  | 0.6841          | 0.0934         | 0.0501           | 0.5692             | 0.0433          | -21.3906       | -20.4809     | 14.1613         | 14.1641       |
         
     | 
| 64 | 
         
            +
            | 0.7259        | 0.15  | 150  | 0.6925          | -0.0147        | -0.0834          | 0.5451             | 0.0688          | -21.8355       | -20.8411     | 13.9200         | 13.9229       |
         
     | 
| 65 | 
         
            +
            | 0.6593        | 0.2   | 200  | 0.7118          | 0.4903         | 0.3962           | 0.5802             | 0.0941          | -20.2368       | -19.1579     | 13.7791         | 13.7821       |
         
     | 
| 66 | 
         
            +
            | 0.7282        | 0.24  | 250  | 0.7093          | -1.2326        | -1.3686          | 0.5648             | 0.1360          | -26.1195       | -24.9010     | 13.8037         | 13.8067       |
         
     | 
| 67 | 
         
            +
            | 0.6924        | 0.29  | 300  | 0.6944          | -0.7898        | -0.9655          | 0.5626             | 0.1757          | -24.7758       | -23.4250     | 14.0496         | 14.0528       |
         
     | 
| 68 | 
         
            +
            | 0.7523        | 0.34  | 350  | 0.6909          | -0.9371        | -1.1226          | 0.5626             | 0.1855          | -25.2994       | -23.9158     | 14.0003         | 14.0037       |
         
     | 
| 69 | 
         
            +
            | 0.7276        | 0.39  | 400  | 0.6918          | -1.8471        | -2.0415          | 0.5868             | 0.1944          | -28.3625       | -26.9492     | 13.3382         | 13.3414       |
         
     | 
| 70 | 
         
            +
            | 0.6255        | 0.44  | 450  | 0.6860          | -1.5470        | -1.7599          | 0.5934             | 0.2129          | -27.4236       | -25.9489     | 13.2551         | 13.2584       |
         
     | 
| 71 | 
         
            +
            | 0.7342        | 0.49  | 500  | 0.6801          | -1.5841        | -1.7888          | 0.5758             | 0.2046          | -27.5199       | -26.0726     | 13.4186         | 13.4219       |
         
     | 
| 72 | 
         
            +
            | 0.568         | 0.54  | 550  | 0.6694          | -1.5101        | -1.7458          | 0.6022             | 0.2356          | -27.3766       | -25.8260     | 13.5776         | 13.5810       |
         
     | 
| 73 | 
         
            +
            | 0.6217        | 0.59  | 600  | 0.6645          | -1.4050        | -1.6543          | 0.6110             | 0.2492          | -27.0716       | -25.4756     | 13.6337         | 13.6371       |
         
     | 
| 74 | 
         
            +
            | 0.6186        | 0.64  | 650  | 0.6682          | -1.3826        | -1.6291          | 0.5978             | 0.2465          | -26.9876       | -25.4007     | 13.4204         | 13.4237       |
         
     | 
| 75 | 
         
            +
            | 0.6637        | 0.68  | 700  | 0.6633          | -1.3994        | -1.6501          | 0.6220             | 0.2507          | -27.0576       | -25.4569     | 13.4574         | 13.4608       |
         
     | 
| 76 | 
         
            +
            | 0.7482        | 0.73  | 750  | 0.6632          | -1.3772        | -1.6269          | 0.6198             | 0.2497          | -26.9804       | -25.3829     | 13.4047         | 13.4081       |
         
     | 
| 77 | 
         
            +
            | 0.6597        | 0.78  | 800  | 0.6627          | -1.3970        | -1.6527          | 0.6198             | 0.2557          | -27.0664       | -25.4489     | 13.3914         | 13.3948       |
         
     | 
| 78 | 
         
            +
            | 0.7206        | 0.83  | 850  | 0.6613          | -1.4018        | -1.6593          | 0.6220             | 0.2575          | -27.0885       | -25.4648     | 13.3862         | 13.3896       |
         
     | 
| 79 | 
         
            +
            | 0.6715        | 0.88  | 900  | 0.6633          | -1.4047        | -1.6584          | 0.6220             | 0.2537          | -27.0856       | -25.4746     | 13.3969         | 13.4003       |
         
     | 
| 80 | 
         
            +
            | 0.6108        | 0.93  | 950  | 0.6633          | -1.4042        | -1.6585          | 0.6242             | 0.2543          | -27.0857       | -25.4727     | 13.3883         | 13.3917       |
         
     | 
| 81 | 
         
            +
            | 0.5964        | 0.98  | 1000 | 0.6641          | -1.4066        | -1.6576          | 0.6198             | 0.2510          | -27.0829       | -25.4808     | 13.3887         | 13.3921       |
         
     | 
| 82 | 
         | 
| 83 | 
         | 
| 84 | 
         
             
            ### Framework versions
         
     | 
    	
        config.json
    CHANGED
    
    | 
         @@ -1,26 +1,46 @@ 
     | 
|
| 1 | 
         
             
            {
         
     | 
| 2 | 
         
            -
              "_name_or_path": " 
     | 
| 3 | 
         
             
              "architectures": [
         
     | 
| 4 | 
         
            -
                " 
     | 
| 5 | 
         
             
              ],
         
     | 
| 6 | 
         
            -
              " 
     | 
| 7 | 
         
            -
             
     | 
| 8 | 
         
            -
               
     | 
| 9 | 
         
            -
              " 
     | 
| 10 | 
         
            -
             
     | 
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 11 | 
         
             
              "initializer_range": 0.02,
         
     | 
| 12 | 
         
            -
              " 
     | 
| 13 | 
         
            -
              " 
     | 
| 14 | 
         
            -
              " 
     | 
| 15 | 
         
            -
              " 
     | 
| 16 | 
         
            -
              " 
     | 
| 17 | 
         
            -
              " 
     | 
| 18 | 
         
            -
              " 
     | 
| 19 | 
         
            -
              " 
     | 
| 20 | 
         
            -
              " 
     | 
| 21 | 
         
            -
              " 
     | 
| 
         | 
|
| 22 | 
         
             
              "torch_dtype": "float16",
         
     | 
| 23 | 
         
             
              "transformers_version": "4.39.1",
         
     | 
| 24 | 
         
             
              "use_cache": false,
         
     | 
| 25 | 
         
            -
              " 
     | 
| 
         | 
|
| 26 | 
         
             
            }
         
     | 
| 
         | 
|
| 1 | 
         
             
            {
         
     | 
| 2 | 
         
            +
              "_name_or_path": "mosaicml/mpt-7b-instruct",
         
     | 
| 3 | 
         
             
              "architectures": [
         
     | 
| 4 | 
         
            +
                "MptForCausalLM"
         
     | 
| 5 | 
         
             
              ],
         
     | 
| 6 | 
         
            +
              "attn_config": {
         
     | 
| 7 | 
         
            +
                "model_type": ""
         
     | 
| 8 | 
         
            +
              },
         
     | 
| 9 | 
         
            +
              "auto_map": {
         
     | 
| 10 | 
         
            +
                "AutoConfig": "mosaicml/mpt-7b-instruct--configuration_mpt.MPTConfig",
         
     | 
| 11 | 
         
            +
                "AutoModelForCausalLM": "mosaicml/mpt-7b-instruct--modeling_mpt.MPTForCausalLM"
         
     | 
| 12 | 
         
            +
              },
         
     | 
| 13 | 
         
            +
              "d_model": 4096,
         
     | 
| 14 | 
         
            +
              "emb_pdrop": 0,
         
     | 
| 15 | 
         
            +
              "embedding_fraction": 1.0,
         
     | 
| 16 | 
         
            +
              "expansion_ratio": 4,
         
     | 
| 17 | 
         
            +
              "init_config": {
         
     | 
| 18 | 
         
            +
                "emb_init_std": null,
         
     | 
| 19 | 
         
            +
                "emb_init_uniform_lim": null,
         
     | 
| 20 | 
         
            +
                "fan_mode": "fan_in",
         
     | 
| 21 | 
         
            +
                "init_div_is_residual": true,
         
     | 
| 22 | 
         
            +
                "init_gain": 0,
         
     | 
| 23 | 
         
            +
                "init_nonlinearity": "relu",
         
     | 
| 24 | 
         
            +
                "init_std": 0.02,
         
     | 
| 25 | 
         
            +
                "name": "kaiming_normal_",
         
     | 
| 26 | 
         
            +
                "verbose": 0
         
     | 
| 27 | 
         
            +
              },
         
     | 
| 28 | 
         
            +
              "init_device": "cpu",
         
     | 
| 29 | 
         
             
              "initializer_range": 0.02,
         
     | 
| 30 | 
         
            +
              "layer_norm_epsilon": 1e-05,
         
     | 
| 31 | 
         
            +
              "learned_pos_emb": true,
         
     | 
| 32 | 
         
            +
              "logit_scale": null,
         
     | 
| 33 | 
         
            +
              "max_seq_len": 2048,
         
     | 
| 34 | 
         
            +
              "model_type": "mpt",
         
     | 
| 35 | 
         
            +
              "n_heads": 32,
         
     | 
| 36 | 
         
            +
              "n_layers": 32,
         
     | 
| 37 | 
         
            +
              "no_bias": true,
         
     | 
| 38 | 
         
            +
              "norm_type": "low_precision_layernorm",
         
     | 
| 39 | 
         
            +
              "resid_pdrop": 0,
         
     | 
| 40 | 
         
            +
              "tokenizer_name": "EleutherAI/gpt-neox-20b",
         
     | 
| 41 | 
         
             
              "torch_dtype": "float16",
         
     | 
| 42 | 
         
             
              "transformers_version": "4.39.1",
         
     | 
| 43 | 
         
             
              "use_cache": false,
         
     | 
| 44 | 
         
            +
              "verbose": 0,
         
     | 
| 45 | 
         
            +
              "vocab_size": 50432
         
     | 
| 46 | 
         
             
            }
         
     | 
    	
        final_checkpoint/config.json
    CHANGED
    
    | 
         @@ -1,26 +1,46 @@ 
     | 
|
| 1 | 
         
             
            {
         
     | 
| 2 | 
         
            -
              "_name_or_path": " 
     | 
| 3 | 
         
             
              "architectures": [
         
     | 
| 4 | 
         
            -
                " 
     | 
| 5 | 
         
             
              ],
         
     | 
| 6 | 
         
            -
              " 
     | 
| 7 | 
         
            -
             
     | 
| 8 | 
         
            -
               
     | 
| 9 | 
         
            -
              " 
     | 
| 10 | 
         
            -
             
     | 
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 11 | 
         
             
              "initializer_range": 0.02,
         
     | 
| 12 | 
         
            -
              " 
     | 
| 13 | 
         
            -
              " 
     | 
| 14 | 
         
            -
              " 
     | 
| 15 | 
         
            -
              " 
     | 
| 16 | 
         
            -
              " 
     | 
| 17 | 
         
            -
              " 
     | 
| 18 | 
         
            -
              " 
     | 
| 19 | 
         
            -
              " 
     | 
| 20 | 
         
            -
              " 
     | 
| 21 | 
         
            -
              " 
     | 
| 
         | 
|
| 22 | 
         
             
              "torch_dtype": "float16",
         
     | 
| 23 | 
         
             
              "transformers_version": "4.39.1",
         
     | 
| 24 | 
         
             
              "use_cache": false,
         
     | 
| 25 | 
         
            -
              " 
     | 
| 
         | 
|
| 26 | 
         
             
            }
         
     | 
| 
         | 
|
| 1 | 
         
             
            {
         
     | 
| 2 | 
         
            +
              "_name_or_path": "mosaicml/mpt-7b-instruct",
         
     | 
| 3 | 
         
             
              "architectures": [
         
     | 
| 4 | 
         
            +
                "MptForCausalLM"
         
     | 
| 5 | 
         
             
              ],
         
     | 
| 6 | 
         
            +
              "attn_config": {
         
     | 
| 7 | 
         
            +
                "model_type": ""
         
     | 
| 8 | 
         
            +
              },
         
     | 
| 9 | 
         
            +
              "auto_map": {
         
     | 
| 10 | 
         
            +
                "AutoConfig": "mosaicml/mpt-7b-instruct--configuration_mpt.MPTConfig",
         
     | 
| 11 | 
         
            +
                "AutoModelForCausalLM": "mosaicml/mpt-7b-instruct--modeling_mpt.MPTForCausalLM"
         
     | 
| 12 | 
         
            +
              },
         
     | 
| 13 | 
         
            +
              "d_model": 4096,
         
     | 
| 14 | 
         
            +
              "emb_pdrop": 0,
         
     | 
| 15 | 
         
            +
              "embedding_fraction": 1.0,
         
     | 
| 16 | 
         
            +
              "expansion_ratio": 4,
         
     | 
| 17 | 
         
            +
              "init_config": {
         
     | 
| 18 | 
         
            +
                "emb_init_std": null,
         
     | 
| 19 | 
         
            +
                "emb_init_uniform_lim": null,
         
     | 
| 20 | 
         
            +
                "fan_mode": "fan_in",
         
     | 
| 21 | 
         
            +
                "init_div_is_residual": true,
         
     | 
| 22 | 
         
            +
                "init_gain": 0,
         
     | 
| 23 | 
         
            +
                "init_nonlinearity": "relu",
         
     | 
| 24 | 
         
            +
                "init_std": 0.02,
         
     | 
| 25 | 
         
            +
                "name": "kaiming_normal_",
         
     | 
| 26 | 
         
            +
                "verbose": 0
         
     | 
| 27 | 
         
            +
              },
         
     | 
| 28 | 
         
            +
              "init_device": "cpu",
         
     | 
| 29 | 
         
             
              "initializer_range": 0.02,
         
     | 
| 30 | 
         
            +
              "layer_norm_epsilon": 1e-05,
         
     | 
| 31 | 
         
            +
              "learned_pos_emb": true,
         
     | 
| 32 | 
         
            +
              "logit_scale": null,
         
     | 
| 33 | 
         
            +
              "max_seq_len": 2048,
         
     | 
| 34 | 
         
            +
              "model_type": "mpt",
         
     | 
| 35 | 
         
            +
              "n_heads": 32,
         
     | 
| 36 | 
         
            +
              "n_layers": 32,
         
     | 
| 37 | 
         
            +
              "no_bias": true,
         
     | 
| 38 | 
         
            +
              "norm_type": "low_precision_layernorm",
         
     | 
| 39 | 
         
            +
              "resid_pdrop": 0,
         
     | 
| 40 | 
         
            +
              "tokenizer_name": "EleutherAI/gpt-neox-20b",
         
     | 
| 41 | 
         
             
              "torch_dtype": "float16",
         
     | 
| 42 | 
         
             
              "transformers_version": "4.39.1",
         
     | 
| 43 | 
         
             
              "use_cache": false,
         
     | 
| 44 | 
         
            +
              "verbose": 0,
         
     | 
| 45 | 
         
            +
              "vocab_size": 50432
         
     | 
| 46 | 
         
             
            }
         
     | 
    	
        final_checkpoint/generation_config.json
    CHANGED
    
    | 
         @@ -1,6 +1,6 @@ 
     | 
|
| 1 | 
         
             
            {
         
     | 
| 2 | 
         
             
              "_from_model_config": true,
         
     | 
| 3 | 
         
            -
              " 
     | 
| 4 | 
         
            -
              " 
     | 
| 5 | 
         
            -
              " 
     | 
| 6 | 
         
             
            }
         
     | 
| 
         | 
|
| 1 | 
         
             
            {
         
     | 
| 2 | 
         
             
              "_from_model_config": true,
         
     | 
| 3 | 
         
            +
              "eos_token_id": 0,
         
     | 
| 4 | 
         
            +
              "transformers_version": "4.39.1",
         
     | 
| 5 | 
         
            +
              "use_cache": false
         
     | 
| 6 | 
         
             
            }
         
     | 
    	
        final_checkpoint/model-00001-of-00003.safetensors
    CHANGED
    
    | 
         @@ -1,3 +1,3 @@ 
     | 
|
| 1 | 
         
             
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            -
            oid sha256: 
     | 
| 3 | 
         
            -
            size  
     | 
| 
         | 
|
| 1 | 
         
             
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:1aef2d0bb5b9f576cd1873ec9de034fac83e25af00d50919492ddf51431a2114
         
     | 
| 3 | 
         
            +
            size 4976746424
         
     | 
    	
        final_checkpoint/model-00002-of-00003.safetensors
    CHANGED
    
    | 
         @@ -1,3 +1,3 @@ 
     | 
|
| 1 | 
         
             
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            -
            oid sha256: 
     | 
| 3 | 
         
            -
            size  
     | 
| 
         | 
|
| 1 | 
         
             
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:0494df8dc581f6d568140d4a3d0961d84abfc239aa2b155e682064f04e5b9ab7
         
     | 
| 3 | 
         
            +
            size 2386755584
         
     | 
    	
        generation_config.json
    CHANGED
    
    | 
         @@ -1,6 +1,6 @@ 
     | 
|
| 1 | 
         
             
            {
         
     | 
| 2 | 
         
             
              "_from_model_config": true,
         
     | 
| 3 | 
         
            -
              " 
     | 
| 4 | 
         
            -
              " 
     | 
| 5 | 
         
            -
              " 
     | 
| 6 | 
         
             
            }
         
     | 
| 
         | 
|
| 1 | 
         
             
            {
         
     | 
| 2 | 
         
             
              "_from_model_config": true,
         
     | 
| 3 | 
         
            +
              "eos_token_id": 0,
         
     | 
| 4 | 
         
            +
              "transformers_version": "4.39.1",
         
     | 
| 5 | 
         
            +
              "use_cache": false
         
     | 
| 6 | 
         
             
            }
         
     | 
    	
        model-00001-of-00003.safetensors
    CHANGED
    
    | 
         @@ -1,3 +1,3 @@ 
     | 
|
| 1 | 
         
             
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            -
            oid sha256: 
     | 
| 3 | 
         
            -
            size  
     | 
| 
         | 
|
| 1 | 
         
             
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:1aef2d0bb5b9f576cd1873ec9de034fac83e25af00d50919492ddf51431a2114
         
     | 
| 3 | 
         
            +
            size 4976746424
         
     | 
    	
        model-00002-of-00003.safetensors
    CHANGED
    
    | 
         @@ -1,3 +1,3 @@ 
     | 
|
| 1 | 
         
             
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            -
            oid sha256: 
     | 
| 3 | 
         
            -
            size  
     | 
| 
         | 
|
| 1 | 
         
             
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:14b36e78d8fb271835ecc916c45a4c949f6f210664b73a5a8118bdd8690949e0
         
     | 
| 3 | 
         
            +
            size 4966260992
         
     | 
    	
        model-00003-of-00003.safetensors
    CHANGED
    
    | 
         @@ -1,3 +1,3 @@ 
     | 
|
| 1 | 
         
             
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            -
            oid sha256: 
     | 
| 3 | 
         
            -
            size  
     | 
| 
         | 
|
| 1 | 
         
             
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:b8dd232e235e5190c9b5d9bb3080d51e9c2304e42dde3b47aea255f0e44f0b0b
         
     | 
| 3 | 
         
            +
            size 3355588232
         
     | 
    	
        model.safetensors.index.json
    CHANGED
    
    | 
         @@ -1,298 +1,201 @@ 
     | 
|
| 1 | 
         
             
            {
         
     | 
| 2 | 
         
             
              "metadata": {
         
     | 
| 3 | 
         
            -
                "total_size":  
     | 
| 4 | 
         
             
              },
         
     | 
| 5 | 
         
             
              "weight_map": {
         
     | 
| 6 | 
         
            -
                " 
     | 
| 7 | 
         
            -
                " 
     | 
| 8 | 
         
            -
                " 
     | 
| 9 | 
         
            -
                " 
     | 
| 10 | 
         
            -
                " 
     | 
| 11 | 
         
            -
                " 
     | 
| 12 | 
         
            -
                " 
     | 
| 13 | 
         
            -
                " 
     | 
| 14 | 
         
            -
                " 
     | 
| 15 | 
         
            -
                " 
     | 
| 16 | 
         
            -
                " 
     | 
| 17 | 
         
            -
                " 
     | 
| 18 | 
         
            -
                " 
     | 
| 19 | 
         
            -
                " 
     | 
| 20 | 
         
            -
                " 
     | 
| 21 | 
         
            -
                " 
     | 
| 22 | 
         
            -
                " 
     | 
| 23 | 
         
            -
                " 
     | 
| 24 | 
         
            -
                " 
     | 
| 25 | 
         
            -
                " 
     | 
| 26 | 
         
            -
                " 
     | 
| 27 | 
         
            -
                " 
     | 
| 28 | 
         
            -
                " 
     | 
| 29 | 
         
            -
                " 
     | 
| 30 | 
         
            -
                " 
     | 
| 31 | 
         
            -
                " 
     | 
| 32 | 
         
            -
                " 
     | 
| 33 | 
         
            -
                " 
     | 
| 34 | 
         
            -
                " 
     | 
| 35 | 
         
            -
                " 
     | 
| 36 | 
         
            -
                " 
     | 
| 37 | 
         
            -
                " 
     | 
| 38 | 
         
            -
                " 
     | 
| 39 | 
         
            -
                " 
     | 
| 40 | 
         
            -
                " 
     | 
| 41 | 
         
            -
                " 
     | 
| 42 | 
         
            -
                " 
     | 
| 43 | 
         
            -
                " 
     | 
| 44 | 
         
            -
                " 
     | 
| 45 | 
         
            -
                " 
     | 
| 46 | 
         
            -
                " 
     | 
| 47 | 
         
            -
                " 
     | 
| 48 | 
         
            -
                " 
     | 
| 49 | 
         
            -
                " 
     | 
| 50 | 
         
            -
                " 
     | 
| 51 | 
         
            -
                " 
     | 
| 52 | 
         
            -
                " 
     | 
| 53 | 
         
            -
                " 
     | 
| 54 | 
         
            -
                " 
     | 
| 55 | 
         
            -
                " 
     | 
| 56 | 
         
            -
                " 
     | 
| 57 | 
         
            -
                " 
     | 
| 58 | 
         
            -
                " 
     | 
| 59 | 
         
            -
                " 
     | 
| 60 | 
         
            -
                " 
     | 
| 61 | 
         
            -
                " 
     | 
| 62 | 
         
            -
                " 
     | 
| 63 | 
         
            -
                " 
     | 
| 64 | 
         
            -
                " 
     | 
| 65 | 
         
            -
                " 
     | 
| 66 | 
         
            -
                " 
     | 
| 67 | 
         
            -
                " 
     | 
| 68 | 
         
            -
                " 
     | 
| 69 | 
         
            -
                " 
     | 
| 70 | 
         
            -
                " 
     | 
| 71 | 
         
            -
                " 
     | 
| 72 | 
         
            -
                " 
     | 
| 73 | 
         
            -
                " 
     | 
| 74 | 
         
            -
                " 
     | 
| 75 | 
         
            -
                " 
     | 
| 76 | 
         
            -
                " 
     | 
| 77 | 
         
            -
                " 
     | 
| 78 | 
         
            -
                " 
     | 
| 79 | 
         
            -
                " 
     | 
| 80 | 
         
            -
                " 
     | 
| 81 | 
         
            -
                " 
     | 
| 82 | 
         
            -
                " 
     | 
| 83 | 
         
            -
                " 
     | 
| 84 | 
         
            -
                " 
     | 
| 85 | 
         
            -
                " 
     | 
| 86 | 
         
            -
                " 
     | 
| 87 | 
         
            -
                " 
     | 
| 88 | 
         
            -
                " 
     | 
| 89 | 
         
            -
                " 
     | 
| 90 | 
         
            -
                " 
     | 
| 91 | 
         
            -
                " 
     | 
| 92 | 
         
            -
                " 
     | 
| 93 | 
         
            -
                " 
     | 
| 94 | 
         
            -
                " 
     | 
| 95 | 
         
            -
                " 
     | 
| 96 | 
         
            -
                " 
     | 
| 97 | 
         
            -
                " 
     | 
| 98 | 
         
            -
                " 
     | 
| 99 | 
         
            -
                " 
     | 
| 100 | 
         
            -
                " 
     | 
| 101 | 
         
            -
                " 
     | 
| 102 | 
         
            -
                " 
     | 
| 103 | 
         
            -
                " 
     | 
| 104 | 
         
            -
                " 
     | 
| 105 | 
         
            -
                " 
     | 
| 106 | 
         
            -
                " 
     | 
| 107 | 
         
            -
                " 
     | 
| 108 | 
         
            -
                " 
     | 
| 109 | 
         
            -
                " 
     | 
| 110 | 
         
            -
                " 
     | 
| 111 | 
         
            -
                " 
     | 
| 112 | 
         
            -
                " 
     | 
| 113 | 
         
            -
                " 
     | 
| 114 | 
         
            -
                " 
     | 
| 115 | 
         
            -
                " 
     | 
| 116 | 
         
            -
                " 
     | 
| 117 | 
         
            -
                " 
     | 
| 118 | 
         
            -
                " 
     | 
| 119 | 
         
            -
                " 
     | 
| 120 | 
         
            -
                " 
     | 
| 121 | 
         
            -
                " 
     | 
| 122 | 
         
            -
                " 
     | 
| 123 | 
         
            -
                " 
     | 
| 124 | 
         
            -
                " 
     | 
| 125 | 
         
            -
                " 
     | 
| 126 | 
         
            -
                " 
     | 
| 127 | 
         
            -
                " 
     | 
| 128 | 
         
            -
                " 
     | 
| 129 | 
         
            -
                " 
     | 
| 130 | 
         
            -
                " 
     | 
| 131 | 
         
            -
                " 
     | 
| 132 | 
         
            -
                " 
     | 
| 133 | 
         
            -
                " 
     | 
| 134 | 
         
            -
                " 
     | 
| 135 | 
         
            -
                " 
     | 
| 136 | 
         
            -
                " 
     | 
| 137 | 
         
            -
                " 
     | 
| 138 | 
         
            -
                " 
     | 
| 139 | 
         
            -
                " 
     | 
| 140 | 
         
            -
                " 
     | 
| 141 | 
         
            -
                " 
     | 
| 142 | 
         
            -
                " 
     | 
| 143 | 
         
            -
                " 
     | 
| 144 | 
         
            -
                " 
     | 
| 145 | 
         
            -
                " 
     | 
| 146 | 
         
            -
                " 
     | 
| 147 | 
         
            -
                " 
     | 
| 148 | 
         
            -
                " 
     | 
| 149 | 
         
            -
                " 
     | 
| 150 | 
         
            -
                " 
     | 
| 151 | 
         
            -
                " 
     | 
| 152 | 
         
            -
                " 
     | 
| 153 | 
         
            -
                " 
     | 
| 154 | 
         
            -
                " 
     | 
| 155 | 
         
            -
                " 
     | 
| 156 | 
         
            -
                " 
     | 
| 157 | 
         
            -
                " 
     | 
| 158 | 
         
            -
                " 
     | 
| 159 | 
         
            -
                " 
     | 
| 160 | 
         
            -
                " 
     | 
| 161 | 
         
            -
                " 
     | 
| 162 | 
         
            -
                " 
     | 
| 163 | 
         
            -
                " 
     | 
| 164 | 
         
            -
                " 
     | 
| 165 | 
         
            -
                " 
     | 
| 166 | 
         
            -
                " 
     | 
| 167 | 
         
            -
                " 
     | 
| 168 | 
         
            -
                " 
     | 
| 169 | 
         
            -
                " 
     | 
| 170 | 
         
            -
                " 
     | 
| 171 | 
         
            -
                " 
     | 
| 172 | 
         
            -
                " 
     | 
| 173 | 
         
            -
                " 
     | 
| 174 | 
         
            -
                " 
     | 
| 175 | 
         
            -
                " 
     | 
| 176 | 
         
            -
                " 
     | 
| 177 | 
         
            -
                " 
     | 
| 178 | 
         
            -
                " 
     | 
| 179 | 
         
            -
                " 
     | 
| 180 | 
         
            -
                " 
     | 
| 181 | 
         
            -
                " 
     | 
| 182 | 
         
            -
                " 
     | 
| 183 | 
         
            -
                " 
     | 
| 184 | 
         
            -
                " 
     | 
| 185 | 
         
            -
                " 
     | 
| 186 | 
         
            -
                " 
     | 
| 187 | 
         
            -
                " 
     | 
| 188 | 
         
            -
                " 
     | 
| 189 | 
         
            -
                " 
     | 
| 190 | 
         
            -
                " 
     | 
| 191 | 
         
            -
                " 
     | 
| 192 | 
         
            -
                " 
     | 
| 193 | 
         
            -
                " 
     | 
| 194 | 
         
            -
                " 
     | 
| 195 | 
         
            -
                " 
     | 
| 196 | 
         
            -
                " 
     | 
| 197 | 
         
            -
                " 
     | 
| 198 | 
         
            -
                " 
     | 
| 199 | 
         
            -
                " 
     | 
| 200 | 
         
            -
                "model.layers.28.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
         
     | 
| 201 | 
         
            -
                "model.layers.28.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
         
     | 
| 202 | 
         
            -
                "model.layers.28.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
         
     | 
| 203 | 
         
            -
                "model.layers.28.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
         
     | 
| 204 | 
         
            -
                "model.layers.28.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
         
     | 
| 205 | 
         
            -
                "model.layers.28.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
         
     | 
| 206 | 
         
            -
                "model.layers.29.input_layernorm.weight": "model-00003-of-00003.safetensors",
         
     | 
| 207 | 
         
            -
                "model.layers.29.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
         
     | 
| 208 | 
         
            -
                "model.layers.29.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
         
     | 
| 209 | 
         
            -
                "model.layers.29.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
         
     | 
| 210 | 
         
            -
                "model.layers.29.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
         
     | 
| 211 | 
         
            -
                "model.layers.29.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
         
     | 
| 212 | 
         
            -
                "model.layers.29.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
         
     | 
| 213 | 
         
            -
                "model.layers.29.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
         
     | 
| 214 | 
         
            -
                "model.layers.29.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
         
     | 
| 215 | 
         
            -
                "model.layers.3.input_layernorm.weight": "model-00001-of-00003.safetensors",
         
     | 
| 216 | 
         
            -
                "model.layers.3.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 217 | 
         
            -
                "model.layers.3.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 218 | 
         
            -
                "model.layers.3.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 219 | 
         
            -
                "model.layers.3.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
         
     | 
| 220 | 
         
            -
                "model.layers.3.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 221 | 
         
            -
                "model.layers.3.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 222 | 
         
            -
                "model.layers.3.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 223 | 
         
            -
                "model.layers.3.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 224 | 
         
            -
                "model.layers.30.input_layernorm.weight": "model-00003-of-00003.safetensors",
         
     | 
| 225 | 
         
            -
                "model.layers.30.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
         
     | 
| 226 | 
         
            -
                "model.layers.30.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
         
     | 
| 227 | 
         
            -
                "model.layers.30.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
         
     | 
| 228 | 
         
            -
                "model.layers.30.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
         
     | 
| 229 | 
         
            -
                "model.layers.30.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
         
     | 
| 230 | 
         
            -
                "model.layers.30.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
         
     | 
| 231 | 
         
            -
                "model.layers.30.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
         
     | 
| 232 | 
         
            -
                "model.layers.30.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
         
     | 
| 233 | 
         
            -
                "model.layers.31.input_layernorm.weight": "model-00003-of-00003.safetensors",
         
     | 
| 234 | 
         
            -
                "model.layers.31.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
         
     | 
| 235 | 
         
            -
                "model.layers.31.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
         
     | 
| 236 | 
         
            -
                "model.layers.31.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
         
     | 
| 237 | 
         
            -
                "model.layers.31.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
         
     | 
| 238 | 
         
            -
                "model.layers.31.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
         
     | 
| 239 | 
         
            -
                "model.layers.31.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
         
     | 
| 240 | 
         
            -
                "model.layers.31.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
         
     | 
| 241 | 
         
            -
                "model.layers.31.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
         
     | 
| 242 | 
         
            -
                "model.layers.4.input_layernorm.weight": "model-00001-of-00003.safetensors",
         
     | 
| 243 | 
         
            -
                "model.layers.4.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 244 | 
         
            -
                "model.layers.4.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 245 | 
         
            -
                "model.layers.4.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 246 | 
         
            -
                "model.layers.4.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
         
     | 
| 247 | 
         
            -
                "model.layers.4.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 248 | 
         
            -
                "model.layers.4.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 249 | 
         
            -
                "model.layers.4.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 250 | 
         
            -
                "model.layers.4.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 251 | 
         
            -
                "model.layers.5.input_layernorm.weight": "model-00001-of-00003.safetensors",
         
     | 
| 252 | 
         
            -
                "model.layers.5.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 253 | 
         
            -
                "model.layers.5.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 254 | 
         
            -
                "model.layers.5.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 255 | 
         
            -
                "model.layers.5.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
         
     | 
| 256 | 
         
            -
                "model.layers.5.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 257 | 
         
            -
                "model.layers.5.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 258 | 
         
            -
                "model.layers.5.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 259 | 
         
            -
                "model.layers.5.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 260 | 
         
            -
                "model.layers.6.input_layernorm.weight": "model-00001-of-00003.safetensors",
         
     | 
| 261 | 
         
            -
                "model.layers.6.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 262 | 
         
            -
                "model.layers.6.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 263 | 
         
            -
                "model.layers.6.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 264 | 
         
            -
                "model.layers.6.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
         
     | 
| 265 | 
         
            -
                "model.layers.6.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 266 | 
         
            -
                "model.layers.6.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 267 | 
         
            -
                "model.layers.6.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 268 | 
         
            -
                "model.layers.6.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 269 | 
         
            -
                "model.layers.7.input_layernorm.weight": "model-00001-of-00003.safetensors",
         
     | 
| 270 | 
         
            -
                "model.layers.7.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 271 | 
         
            -
                "model.layers.7.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 272 | 
         
            -
                "model.layers.7.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 273 | 
         
            -
                "model.layers.7.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
         
     | 
| 274 | 
         
            -
                "model.layers.7.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 275 | 
         
            -
                "model.layers.7.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 276 | 
         
            -
                "model.layers.7.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 277 | 
         
            -
                "model.layers.7.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 278 | 
         
            -
                "model.layers.8.input_layernorm.weight": "model-00001-of-00003.safetensors",
         
     | 
| 279 | 
         
            -
                "model.layers.8.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 280 | 
         
            -
                "model.layers.8.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 281 | 
         
            -
                "model.layers.8.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 282 | 
         
            -
                "model.layers.8.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
         
     | 
| 283 | 
         
            -
                "model.layers.8.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 284 | 
         
            -
                "model.layers.8.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 285 | 
         
            -
                "model.layers.8.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 286 | 
         
            -
                "model.layers.8.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 287 | 
         
            -
                "model.layers.9.input_layernorm.weight": "model-00001-of-00003.safetensors",
         
     | 
| 288 | 
         
            -
                "model.layers.9.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 289 | 
         
            -
                "model.layers.9.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 290 | 
         
            -
                "model.layers.9.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 291 | 
         
            -
                "model.layers.9.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
         
     | 
| 292 | 
         
            -
                "model.layers.9.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 293 | 
         
            -
                "model.layers.9.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 294 | 
         
            -
                "model.layers.9.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 295 | 
         
            -
                "model.layers.9.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 296 | 
         
            -
                "model.norm.weight": "model-00003-of-00003.safetensors"
         
     | 
| 297 | 
         
             
              }
         
     | 
| 298 | 
         
             
            }
         
     | 
| 
         | 
|
| 1 | 
         
             
            {
         
     | 
| 2 | 
         
             
              "metadata": {
         
     | 
| 3 | 
         
            +
                "total_size": 13298573312
         
     | 
| 4 | 
         
             
              },
         
     | 
| 5 | 
         
             
              "weight_map": {
         
     | 
| 6 | 
         
            +
                "transformer.blocks.0.attn.Wqkv.weight": "model-00001-of-00003.safetensors",
         
     | 
| 7 | 
         
            +
                "transformer.blocks.0.attn.out_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 8 | 
         
            +
                "transformer.blocks.0.ffn.down_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 9 | 
         
            +
                "transformer.blocks.0.ffn.up_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 10 | 
         
            +
                "transformer.blocks.0.norm_1.weight": "model-00001-of-00003.safetensors",
         
     | 
| 11 | 
         
            +
                "transformer.blocks.0.norm_2.weight": "model-00001-of-00003.safetensors",
         
     | 
| 12 | 
         
            +
                "transformer.blocks.1.attn.Wqkv.weight": "model-00001-of-00003.safetensors",
         
     | 
| 13 | 
         
            +
                "transformer.blocks.1.attn.out_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 14 | 
         
            +
                "transformer.blocks.1.ffn.down_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 15 | 
         
            +
                "transformer.blocks.1.ffn.up_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 16 | 
         
            +
                "transformer.blocks.1.norm_1.weight": "model-00001-of-00003.safetensors",
         
     | 
| 17 | 
         
            +
                "transformer.blocks.1.norm_2.weight": "model-00001-of-00003.safetensors",
         
     | 
| 18 | 
         
            +
                "transformer.blocks.10.attn.Wqkv.weight": "model-00001-of-00003.safetensors",
         
     | 
| 19 | 
         
            +
                "transformer.blocks.10.attn.out_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 20 | 
         
            +
                "transformer.blocks.10.ffn.down_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 21 | 
         
            +
                "transformer.blocks.10.ffn.up_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 22 | 
         
            +
                "transformer.blocks.10.norm_1.weight": "model-00001-of-00003.safetensors",
         
     | 
| 23 | 
         
            +
                "transformer.blocks.10.norm_2.weight": "model-00001-of-00003.safetensors",
         
     | 
| 24 | 
         
            +
                "transformer.blocks.11.attn.Wqkv.weight": "model-00001-of-00003.safetensors",
         
     | 
| 25 | 
         
            +
                "transformer.blocks.11.attn.out_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 26 | 
         
            +
                "transformer.blocks.11.ffn.down_proj.weight": "model-00002-of-00003.safetensors",
         
     | 
| 27 | 
         
            +
                "transformer.blocks.11.ffn.up_proj.weight": "model-00002-of-00003.safetensors",
         
     | 
| 28 | 
         
            +
                "transformer.blocks.11.norm_1.weight": "model-00001-of-00003.safetensors",
         
     | 
| 29 | 
         
            +
                "transformer.blocks.11.norm_2.weight": "model-00001-of-00003.safetensors",
         
     | 
| 30 | 
         
            +
                "transformer.blocks.12.attn.Wqkv.weight": "model-00002-of-00003.safetensors",
         
     | 
| 31 | 
         
            +
                "transformer.blocks.12.attn.out_proj.weight": "model-00002-of-00003.safetensors",
         
     | 
| 32 | 
         
            +
                "transformer.blocks.12.ffn.down_proj.weight": "model-00002-of-00003.safetensors",
         
     | 
| 33 | 
         
            +
                "transformer.blocks.12.ffn.up_proj.weight": "model-00002-of-00003.safetensors",
         
     | 
| 34 | 
         
            +
                "transformer.blocks.12.norm_1.weight": "model-00002-of-00003.safetensors",
         
     | 
| 35 | 
         
            +
                "transformer.blocks.12.norm_2.weight": "model-00002-of-00003.safetensors",
         
     | 
| 36 | 
         
            +
                "transformer.blocks.13.attn.Wqkv.weight": "model-00002-of-00003.safetensors",
         
     | 
| 37 | 
         
            +
                "transformer.blocks.13.attn.out_proj.weight": "model-00002-of-00003.safetensors",
         
     | 
| 38 | 
         
            +
                "transformer.blocks.13.ffn.down_proj.weight": "model-00002-of-00003.safetensors",
         
     | 
| 39 | 
         
            +
                "transformer.blocks.13.ffn.up_proj.weight": "model-00002-of-00003.safetensors",
         
     | 
| 40 | 
         
            +
                "transformer.blocks.13.norm_1.weight": "model-00002-of-00003.safetensors",
         
     | 
| 41 | 
         
            +
                "transformer.blocks.13.norm_2.weight": "model-00002-of-00003.safetensors",
         
     | 
| 42 | 
         
            +
                "transformer.blocks.14.attn.Wqkv.weight": "model-00002-of-00003.safetensors",
         
     | 
| 43 | 
         
            +
                "transformer.blocks.14.attn.out_proj.weight": "model-00002-of-00003.safetensors",
         
     | 
| 44 | 
         
            +
                "transformer.blocks.14.ffn.down_proj.weight": "model-00002-of-00003.safetensors",
         
     | 
| 45 | 
         
            +
                "transformer.blocks.14.ffn.up_proj.weight": "model-00002-of-00003.safetensors",
         
     | 
| 46 | 
         
            +
                "transformer.blocks.14.norm_1.weight": "model-00002-of-00003.safetensors",
         
     | 
| 47 | 
         
            +
                "transformer.blocks.14.norm_2.weight": "model-00002-of-00003.safetensors",
         
     | 
| 48 | 
         
            +
                "transformer.blocks.15.attn.Wqkv.weight": "model-00002-of-00003.safetensors",
         
     | 
| 49 | 
         
            +
                "transformer.blocks.15.attn.out_proj.weight": "model-00002-of-00003.safetensors",
         
     | 
| 50 | 
         
            +
                "transformer.blocks.15.ffn.down_proj.weight": "model-00002-of-00003.safetensors",
         
     | 
| 51 | 
         
            +
                "transformer.blocks.15.ffn.up_proj.weight": "model-00002-of-00003.safetensors",
         
     | 
| 52 | 
         
            +
                "transformer.blocks.15.norm_1.weight": "model-00002-of-00003.safetensors",
         
     | 
| 53 | 
         
            +
                "transformer.blocks.15.norm_2.weight": "model-00002-of-00003.safetensors",
         
     | 
| 54 | 
         
            +
                "transformer.blocks.16.attn.Wqkv.weight": "model-00002-of-00003.safetensors",
         
     | 
| 55 | 
         
            +
                "transformer.blocks.16.attn.out_proj.weight": "model-00002-of-00003.safetensors",
         
     | 
| 56 | 
         
            +
                "transformer.blocks.16.ffn.down_proj.weight": "model-00002-of-00003.safetensors",
         
     | 
| 57 | 
         
            +
                "transformer.blocks.16.ffn.up_proj.weight": "model-00002-of-00003.safetensors",
         
     | 
| 58 | 
         
            +
                "transformer.blocks.16.norm_1.weight": "model-00002-of-00003.safetensors",
         
     | 
| 59 | 
         
            +
                "transformer.blocks.16.norm_2.weight": "model-00002-of-00003.safetensors",
         
     | 
| 60 | 
         
            +
                "transformer.blocks.17.attn.Wqkv.weight": "model-00002-of-00003.safetensors",
         
     | 
| 61 | 
         
            +
                "transformer.blocks.17.attn.out_proj.weight": "model-00002-of-00003.safetensors",
         
     | 
| 62 | 
         
            +
                "transformer.blocks.17.ffn.down_proj.weight": "model-00002-of-00003.safetensors",
         
     | 
| 63 | 
         
            +
                "transformer.blocks.17.ffn.up_proj.weight": "model-00002-of-00003.safetensors",
         
     | 
| 64 | 
         
            +
                "transformer.blocks.17.norm_1.weight": "model-00002-of-00003.safetensors",
         
     | 
| 65 | 
         
            +
                "transformer.blocks.17.norm_2.weight": "model-00002-of-00003.safetensors",
         
     | 
| 66 | 
         
            +
                "transformer.blocks.18.attn.Wqkv.weight": "model-00002-of-00003.safetensors",
         
     | 
| 67 | 
         
            +
                "transformer.blocks.18.attn.out_proj.weight": "model-00002-of-00003.safetensors",
         
     | 
| 68 | 
         
            +
                "transformer.blocks.18.ffn.down_proj.weight": "model-00002-of-00003.safetensors",
         
     | 
| 69 | 
         
            +
                "transformer.blocks.18.ffn.up_proj.weight": "model-00002-of-00003.safetensors",
         
     | 
| 70 | 
         
            +
                "transformer.blocks.18.norm_1.weight": "model-00002-of-00003.safetensors",
         
     | 
| 71 | 
         
            +
                "transformer.blocks.18.norm_2.weight": "model-00002-of-00003.safetensors",
         
     | 
| 72 | 
         
            +
                "transformer.blocks.19.attn.Wqkv.weight": "model-00002-of-00003.safetensors",
         
     | 
| 73 | 
         
            +
                "transformer.blocks.19.attn.out_proj.weight": "model-00002-of-00003.safetensors",
         
     | 
| 74 | 
         
            +
                "transformer.blocks.19.ffn.down_proj.weight": "model-00002-of-00003.safetensors",
         
     | 
| 75 | 
         
            +
                "transformer.blocks.19.ffn.up_proj.weight": "model-00002-of-00003.safetensors",
         
     | 
| 76 | 
         
            +
                "transformer.blocks.19.norm_1.weight": "model-00002-of-00003.safetensors",
         
     | 
| 77 | 
         
            +
                "transformer.blocks.19.norm_2.weight": "model-00002-of-00003.safetensors",
         
     | 
| 78 | 
         
            +
                "transformer.blocks.2.attn.Wqkv.weight": "model-00001-of-00003.safetensors",
         
     | 
| 79 | 
         
            +
                "transformer.blocks.2.attn.out_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 80 | 
         
            +
                "transformer.blocks.2.ffn.down_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 81 | 
         
            +
                "transformer.blocks.2.ffn.up_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 82 | 
         
            +
                "transformer.blocks.2.norm_1.weight": "model-00001-of-00003.safetensors",
         
     | 
| 83 | 
         
            +
                "transformer.blocks.2.norm_2.weight": "model-00001-of-00003.safetensors",
         
     | 
| 84 | 
         
            +
                "transformer.blocks.20.attn.Wqkv.weight": "model-00002-of-00003.safetensors",
         
     | 
| 85 | 
         
            +
                "transformer.blocks.20.attn.out_proj.weight": "model-00002-of-00003.safetensors",
         
     | 
| 86 | 
         
            +
                "transformer.blocks.20.ffn.down_proj.weight": "model-00002-of-00003.safetensors",
         
     | 
| 87 | 
         
            +
                "transformer.blocks.20.ffn.up_proj.weight": "model-00002-of-00003.safetensors",
         
     | 
| 88 | 
         
            +
                "transformer.blocks.20.norm_1.weight": "model-00002-of-00003.safetensors",
         
     | 
| 89 | 
         
            +
                "transformer.blocks.20.norm_2.weight": "model-00002-of-00003.safetensors",
         
     | 
| 90 | 
         
            +
                "transformer.blocks.21.attn.Wqkv.weight": "model-00002-of-00003.safetensors",
         
     | 
| 91 | 
         
            +
                "transformer.blocks.21.attn.out_proj.weight": "model-00002-of-00003.safetensors",
         
     | 
| 92 | 
         
            +
                "transformer.blocks.21.ffn.down_proj.weight": "model-00002-of-00003.safetensors",
         
     | 
| 93 | 
         
            +
                "transformer.blocks.21.ffn.up_proj.weight": "model-00002-of-00003.safetensors",
         
     | 
| 94 | 
         
            +
                "transformer.blocks.21.norm_1.weight": "model-00002-of-00003.safetensors",
         
     | 
| 95 | 
         
            +
                "transformer.blocks.21.norm_2.weight": "model-00002-of-00003.safetensors",
         
     | 
| 96 | 
         
            +
                "transformer.blocks.22.attn.Wqkv.weight": "model-00002-of-00003.safetensors",
         
     | 
| 97 | 
         
            +
                "transformer.blocks.22.attn.out_proj.weight": "model-00002-of-00003.safetensors",
         
     | 
| 98 | 
         
            +
                "transformer.blocks.22.ffn.down_proj.weight": "model-00002-of-00003.safetensors",
         
     | 
| 99 | 
         
            +
                "transformer.blocks.22.ffn.up_proj.weight": "model-00002-of-00003.safetensors",
         
     | 
| 100 | 
         
            +
                "transformer.blocks.22.norm_1.weight": "model-00002-of-00003.safetensors",
         
     | 
| 101 | 
         
            +
                "transformer.blocks.22.norm_2.weight": "model-00002-of-00003.safetensors",
         
     | 
| 102 | 
         
            +
                "transformer.blocks.23.attn.Wqkv.weight": "model-00002-of-00003.safetensors",
         
     | 
| 103 | 
         
            +
                "transformer.blocks.23.attn.out_proj.weight": "model-00002-of-00003.safetensors",
         
     | 
| 104 | 
         
            +
                "transformer.blocks.23.ffn.down_proj.weight": "model-00003-of-00003.safetensors",
         
     | 
| 105 | 
         
            +
                "transformer.blocks.23.ffn.up_proj.weight": "model-00002-of-00003.safetensors",
         
     | 
| 106 | 
         
            +
                "transformer.blocks.23.norm_1.weight": "model-00002-of-00003.safetensors",
         
     | 
| 107 | 
         
            +
                "transformer.blocks.23.norm_2.weight": "model-00002-of-00003.safetensors",
         
     | 
| 108 | 
         
            +
                "transformer.blocks.24.attn.Wqkv.weight": "model-00003-of-00003.safetensors",
         
     | 
| 109 | 
         
            +
                "transformer.blocks.24.attn.out_proj.weight": "model-00003-of-00003.safetensors",
         
     | 
| 110 | 
         
            +
                "transformer.blocks.24.ffn.down_proj.weight": "model-00003-of-00003.safetensors",
         
     | 
| 111 | 
         
            +
                "transformer.blocks.24.ffn.up_proj.weight": "model-00003-of-00003.safetensors",
         
     | 
| 112 | 
         
            +
                "transformer.blocks.24.norm_1.weight": "model-00003-of-00003.safetensors",
         
     | 
| 113 | 
         
            +
                "transformer.blocks.24.norm_2.weight": "model-00003-of-00003.safetensors",
         
     | 
| 114 | 
         
            +
                "transformer.blocks.25.attn.Wqkv.weight": "model-00003-of-00003.safetensors",
         
     | 
| 115 | 
         
            +
                "transformer.blocks.25.attn.out_proj.weight": "model-00003-of-00003.safetensors",
         
     | 
| 116 | 
         
            +
                "transformer.blocks.25.ffn.down_proj.weight": "model-00003-of-00003.safetensors",
         
     | 
| 117 | 
         
            +
                "transformer.blocks.25.ffn.up_proj.weight": "model-00003-of-00003.safetensors",
         
     | 
| 118 | 
         
            +
                "transformer.blocks.25.norm_1.weight": "model-00003-of-00003.safetensors",
         
     | 
| 119 | 
         
            +
                "transformer.blocks.25.norm_2.weight": "model-00003-of-00003.safetensors",
         
     | 
| 120 | 
         
            +
                "transformer.blocks.26.attn.Wqkv.weight": "model-00003-of-00003.safetensors",
         
     | 
| 121 | 
         
            +
                "transformer.blocks.26.attn.out_proj.weight": "model-00003-of-00003.safetensors",
         
     | 
| 122 | 
         
            +
                "transformer.blocks.26.ffn.down_proj.weight": "model-00003-of-00003.safetensors",
         
     | 
| 123 | 
         
            +
                "transformer.blocks.26.ffn.up_proj.weight": "model-00003-of-00003.safetensors",
         
     | 
| 124 | 
         
            +
                "transformer.blocks.26.norm_1.weight": "model-00003-of-00003.safetensors",
         
     | 
| 125 | 
         
            +
                "transformer.blocks.26.norm_2.weight": "model-00003-of-00003.safetensors",
         
     | 
| 126 | 
         
            +
                "transformer.blocks.27.attn.Wqkv.weight": "model-00003-of-00003.safetensors",
         
     | 
| 127 | 
         
            +
                "transformer.blocks.27.attn.out_proj.weight": "model-00003-of-00003.safetensors",
         
     | 
| 128 | 
         
            +
                "transformer.blocks.27.ffn.down_proj.weight": "model-00003-of-00003.safetensors",
         
     | 
| 129 | 
         
            +
                "transformer.blocks.27.ffn.up_proj.weight": "model-00003-of-00003.safetensors",
         
     | 
| 130 | 
         
            +
                "transformer.blocks.27.norm_1.weight": "model-00003-of-00003.safetensors",
         
     | 
| 131 | 
         
            +
                "transformer.blocks.27.norm_2.weight": "model-00003-of-00003.safetensors",
         
     | 
| 132 | 
         
            +
                "transformer.blocks.28.attn.Wqkv.weight": "model-00003-of-00003.safetensors",
         
     | 
| 133 | 
         
            +
                "transformer.blocks.28.attn.out_proj.weight": "model-00003-of-00003.safetensors",
         
     | 
| 134 | 
         
            +
                "transformer.blocks.28.ffn.down_proj.weight": "model-00003-of-00003.safetensors",
         
     | 
| 135 | 
         
            +
                "transformer.blocks.28.ffn.up_proj.weight": "model-00003-of-00003.safetensors",
         
     | 
| 136 | 
         
            +
                "transformer.blocks.28.norm_1.weight": "model-00003-of-00003.safetensors",
         
     | 
| 137 | 
         
            +
                "transformer.blocks.28.norm_2.weight": "model-00003-of-00003.safetensors",
         
     | 
| 138 | 
         
            +
                "transformer.blocks.29.attn.Wqkv.weight": "model-00003-of-00003.safetensors",
         
     | 
| 139 | 
         
            +
                "transformer.blocks.29.attn.out_proj.weight": "model-00003-of-00003.safetensors",
         
     | 
| 140 | 
         
            +
                "transformer.blocks.29.ffn.down_proj.weight": "model-00003-of-00003.safetensors",
         
     | 
| 141 | 
         
            +
                "transformer.blocks.29.ffn.up_proj.weight": "model-00003-of-00003.safetensors",
         
     | 
| 142 | 
         
            +
                "transformer.blocks.29.norm_1.weight": "model-00003-of-00003.safetensors",
         
     | 
| 143 | 
         
            +
                "transformer.blocks.29.norm_2.weight": "model-00003-of-00003.safetensors",
         
     | 
| 144 | 
         
            +
                "transformer.blocks.3.attn.Wqkv.weight": "model-00001-of-00003.safetensors",
         
     | 
| 145 | 
         
            +
                "transformer.blocks.3.attn.out_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 146 | 
         
            +
                "transformer.blocks.3.ffn.down_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 147 | 
         
            +
                "transformer.blocks.3.ffn.up_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 148 | 
         
            +
                "transformer.blocks.3.norm_1.weight": "model-00001-of-00003.safetensors",
         
     | 
| 149 | 
         
            +
                "transformer.blocks.3.norm_2.weight": "model-00001-of-00003.safetensors",
         
     | 
| 150 | 
         
            +
                "transformer.blocks.30.attn.Wqkv.weight": "model-00003-of-00003.safetensors",
         
     | 
| 151 | 
         
            +
                "transformer.blocks.30.attn.out_proj.weight": "model-00003-of-00003.safetensors",
         
     | 
| 152 | 
         
            +
                "transformer.blocks.30.ffn.down_proj.weight": "model-00003-of-00003.safetensors",
         
     | 
| 153 | 
         
            +
                "transformer.blocks.30.ffn.up_proj.weight": "model-00003-of-00003.safetensors",
         
     | 
| 154 | 
         
            +
                "transformer.blocks.30.norm_1.weight": "model-00003-of-00003.safetensors",
         
     | 
| 155 | 
         
            +
                "transformer.blocks.30.norm_2.weight": "model-00003-of-00003.safetensors",
         
     | 
| 156 | 
         
            +
                "transformer.blocks.31.attn.Wqkv.weight": "model-00003-of-00003.safetensors",
         
     | 
| 157 | 
         
            +
                "transformer.blocks.31.attn.out_proj.weight": "model-00003-of-00003.safetensors",
         
     | 
| 158 | 
         
            +
                "transformer.blocks.31.ffn.down_proj.weight": "model-00003-of-00003.safetensors",
         
     | 
| 159 | 
         
            +
                "transformer.blocks.31.ffn.up_proj.weight": "model-00003-of-00003.safetensors",
         
     | 
| 160 | 
         
            +
                "transformer.blocks.31.norm_1.weight": "model-00003-of-00003.safetensors",
         
     | 
| 161 | 
         
            +
                "transformer.blocks.31.norm_2.weight": "model-00003-of-00003.safetensors",
         
     | 
| 162 | 
         
            +
                "transformer.blocks.4.attn.Wqkv.weight": "model-00001-of-00003.safetensors",
         
     | 
| 163 | 
         
            +
                "transformer.blocks.4.attn.out_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 164 | 
         
            +
                "transformer.blocks.4.ffn.down_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 165 | 
         
            +
                "transformer.blocks.4.ffn.up_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 166 | 
         
            +
                "transformer.blocks.4.norm_1.weight": "model-00001-of-00003.safetensors",
         
     | 
| 167 | 
         
            +
                "transformer.blocks.4.norm_2.weight": "model-00001-of-00003.safetensors",
         
     | 
| 168 | 
         
            +
                "transformer.blocks.5.attn.Wqkv.weight": "model-00001-of-00003.safetensors",
         
     | 
| 169 | 
         
            +
                "transformer.blocks.5.attn.out_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 170 | 
         
            +
                "transformer.blocks.5.ffn.down_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 171 | 
         
            +
                "transformer.blocks.5.ffn.up_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 172 | 
         
            +
                "transformer.blocks.5.norm_1.weight": "model-00001-of-00003.safetensors",
         
     | 
| 173 | 
         
            +
                "transformer.blocks.5.norm_2.weight": "model-00001-of-00003.safetensors",
         
     | 
| 174 | 
         
            +
                "transformer.blocks.6.attn.Wqkv.weight": "model-00001-of-00003.safetensors",
         
     | 
| 175 | 
         
            +
                "transformer.blocks.6.attn.out_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 176 | 
         
            +
                "transformer.blocks.6.ffn.down_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 177 | 
         
            +
                "transformer.blocks.6.ffn.up_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 178 | 
         
            +
                "transformer.blocks.6.norm_1.weight": "model-00001-of-00003.safetensors",
         
     | 
| 179 | 
         
            +
                "transformer.blocks.6.norm_2.weight": "model-00001-of-00003.safetensors",
         
     | 
| 180 | 
         
            +
                "transformer.blocks.7.attn.Wqkv.weight": "model-00001-of-00003.safetensors",
         
     | 
| 181 | 
         
            +
                "transformer.blocks.7.attn.out_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 182 | 
         
            +
                "transformer.blocks.7.ffn.down_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 183 | 
         
            +
                "transformer.blocks.7.ffn.up_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 184 | 
         
            +
                "transformer.blocks.7.norm_1.weight": "model-00001-of-00003.safetensors",
         
     | 
| 185 | 
         
            +
                "transformer.blocks.7.norm_2.weight": "model-00001-of-00003.safetensors",
         
     | 
| 186 | 
         
            +
                "transformer.blocks.8.attn.Wqkv.weight": "model-00001-of-00003.safetensors",
         
     | 
| 187 | 
         
            +
                "transformer.blocks.8.attn.out_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 188 | 
         
            +
                "transformer.blocks.8.ffn.down_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 189 | 
         
            +
                "transformer.blocks.8.ffn.up_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 190 | 
         
            +
                "transformer.blocks.8.norm_1.weight": "model-00001-of-00003.safetensors",
         
     | 
| 191 | 
         
            +
                "transformer.blocks.8.norm_2.weight": "model-00001-of-00003.safetensors",
         
     | 
| 192 | 
         
            +
                "transformer.blocks.9.attn.Wqkv.weight": "model-00001-of-00003.safetensors",
         
     | 
| 193 | 
         
            +
                "transformer.blocks.9.attn.out_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 194 | 
         
            +
                "transformer.blocks.9.ffn.down_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 195 | 
         
            +
                "transformer.blocks.9.ffn.up_proj.weight": "model-00001-of-00003.safetensors",
         
     | 
| 196 | 
         
            +
                "transformer.blocks.9.norm_1.weight": "model-00001-of-00003.safetensors",
         
     | 
| 197 | 
         
            +
                "transformer.blocks.9.norm_2.weight": "model-00001-of-00003.safetensors",
         
     | 
| 198 | 
         
            +
                "transformer.norm_f.weight": "model-00003-of-00003.safetensors",
         
     | 
| 199 | 
         
            +
                "transformer.wte.weight": "model-00001-of-00003.safetensors"
         
     | 
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 200 | 
         
             
              }
         
     | 
| 201 | 
         
             
            }
         
     | 
    	
        special_tokens_map.json
    CHANGED
    
    | 
         @@ -1,21 +1,21 @@ 
     | 
|
| 1 | 
         
             
            {
         
     | 
| 2 | 
         
             
              "bos_token": {
         
     | 
| 3 | 
         
            -
                "content": " 
     | 
| 4 | 
         
             
                "lstrip": false,
         
     | 
| 5 | 
         
             
                "normalized": false,
         
     | 
| 6 | 
         
             
                "rstrip": false,
         
     | 
| 7 | 
         
             
                "single_word": false
         
     | 
| 8 | 
         
             
              },
         
     | 
| 9 | 
         
             
              "eos_token": {
         
     | 
| 10 | 
         
            -
                "content": " 
     | 
| 11 | 
         
             
                "lstrip": false,
         
     | 
| 12 | 
         
             
                "normalized": false,
         
     | 
| 13 | 
         
             
                "rstrip": false,
         
     | 
| 14 | 
         
             
                "single_word": false
         
     | 
| 15 | 
         
             
              },
         
     | 
| 16 | 
         
            -
              "pad_token": " 
     | 
| 17 | 
         
             
              "unk_token": {
         
     | 
| 18 | 
         
            -
                "content": " 
     | 
| 19 | 
         
             
                "lstrip": false,
         
     | 
| 20 | 
         
             
                "normalized": false,
         
     | 
| 21 | 
         
             
                "rstrip": false,
         
     | 
| 
         | 
|
| 1 | 
         
             
            {
         
     | 
| 2 | 
         
             
              "bos_token": {
         
     | 
| 3 | 
         
            +
                "content": "<|endoftext|>",
         
     | 
| 4 | 
         
             
                "lstrip": false,
         
     | 
| 5 | 
         
             
                "normalized": false,
         
     | 
| 6 | 
         
             
                "rstrip": false,
         
     | 
| 7 | 
         
             
                "single_word": false
         
     | 
| 8 | 
         
             
              },
         
     | 
| 9 | 
         
             
              "eos_token": {
         
     | 
| 10 | 
         
            +
                "content": "<|endoftext|>",
         
     | 
| 11 | 
         
             
                "lstrip": false,
         
     | 
| 12 | 
         
             
                "normalized": false,
         
     | 
| 13 | 
         
             
                "rstrip": false,
         
     | 
| 14 | 
         
             
                "single_word": false
         
     | 
| 15 | 
         
             
              },
         
     | 
| 16 | 
         
            +
              "pad_token": "<|endoftext|>",
         
     | 
| 17 | 
         
             
              "unk_token": {
         
     | 
| 18 | 
         
            +
                "content": "<|endoftext|>",
         
     | 
| 19 | 
         
             
                "lstrip": false,
         
     | 
| 20 | 
         
             
                "normalized": false,
         
     | 
| 21 | 
         
             
                "rstrip": false,
         
     | 
    	
        tokenizer.json
    CHANGED
    
    | 
         The diff for this file is too large to render. 
		See raw diff 
     | 
| 
         | 
    	
        tokenizer_config.json
    CHANGED
    
    | 
         @@ -1,9 +1,8 @@ 
     | 
|
| 1 | 
         
             
            {
         
     | 
| 2 | 
         
            -
              " 
     | 
| 3 | 
         
            -
              "add_eos_token": false,
         
     | 
| 4 | 
         
             
              "added_tokens_decoder": {
         
     | 
| 5 | 
         
             
                "0": {
         
     | 
| 6 | 
         
            -
                  "content": " 
     | 
| 7 | 
         
             
                  "lstrip": false,
         
     | 
| 8 | 
         
             
                  "normalized": false,
         
     | 
| 9 | 
         
             
                  "rstrip": false,
         
     | 
| 
         @@ -11,33 +10,203 @@ 
     | 
|
| 11 | 
         
             
                  "special": true
         
     | 
| 12 | 
         
             
                },
         
     | 
| 13 | 
         
             
                "1": {
         
     | 
| 14 | 
         
            -
                  "content": " 
     | 
| 15 | 
         
             
                  "lstrip": false,
         
     | 
| 16 | 
         
             
                  "normalized": false,
         
     | 
| 17 | 
         
             
                  "rstrip": false,
         
     | 
| 18 | 
         
             
                  "single_word": false,
         
     | 
| 19 | 
         
             
                  "special": true
         
     | 
| 20 | 
         
             
                },
         
     | 
| 21 | 
         
            -
                " 
     | 
| 22 | 
         
            -
                  "content": " 
     | 
| 23 | 
         
             
                  "lstrip": false,
         
     | 
| 24 | 
         
            -
                  "normalized":  
     | 
| 25 | 
         
             
                  "rstrip": false,
         
     | 
| 26 | 
         
             
                  "single_word": false,
         
     | 
| 27 | 
         
            -
                  "special":  
     | 
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 28 | 
         
             
                }
         
     | 
| 29 | 
         
             
              },
         
     | 
| 30 | 
         
            -
              " 
     | 
| 31 | 
         
            -
              " 
     | 
| 32 | 
         
            -
              " 
     | 
| 33 | 
         
            -
              " 
     | 
| 34 | 
         
            -
              " 
     | 
| 35 | 
         
            -
              " 
     | 
| 36 | 
         
            -
              " 
     | 
| 37 | 
         
            -
              "pad_token": "</s>",
         
     | 
| 38 | 
         
            -
              "sp_model_kwargs": {},
         
     | 
| 39 | 
         
            -
              "spaces_between_special_tokens": false,
         
     | 
| 40 | 
         
            -
              "tokenizer_class": "LlamaTokenizer",
         
     | 
| 41 | 
         
            -
              "unk_token": "<unk>",
         
     | 
| 42 | 
         
            -
              "use_default_system_prompt": false
         
     | 
| 43 | 
         
             
            }
         
     | 
| 
         | 
|
| 1 | 
         
             
            {
         
     | 
| 2 | 
         
            +
              "add_prefix_space": false,
         
     | 
| 
         | 
|
| 3 | 
         
             
              "added_tokens_decoder": {
         
     | 
| 4 | 
         
             
                "0": {
         
     | 
| 5 | 
         
            +
                  "content": "<|endoftext|>",
         
     | 
| 6 | 
         
             
                  "lstrip": false,
         
     | 
| 7 | 
         
             
                  "normalized": false,
         
     | 
| 8 | 
         
             
                  "rstrip": false,
         
     | 
| 
         | 
|
| 10 | 
         
             
                  "special": true
         
     | 
| 11 | 
         
             
                },
         
     | 
| 12 | 
         
             
                "1": {
         
     | 
| 13 | 
         
            +
                  "content": "<|padding|>",
         
     | 
| 14 | 
         
             
                  "lstrip": false,
         
     | 
| 15 | 
         
             
                  "normalized": false,
         
     | 
| 16 | 
         
             
                  "rstrip": false,
         
     | 
| 17 | 
         
             
                  "single_word": false,
         
     | 
| 18 | 
         
             
                  "special": true
         
     | 
| 19 | 
         
             
                },
         
     | 
| 20 | 
         
            +
                "50254": {
         
     | 
| 21 | 
         
            +
                  "content": "                        ",
         
     | 
| 22 | 
         
             
                  "lstrip": false,
         
     | 
| 23 | 
         
            +
                  "normalized": true,
         
     | 
| 24 | 
         
             
                  "rstrip": false,
         
     | 
| 25 | 
         
             
                  "single_word": false,
         
     | 
| 26 | 
         
            +
                  "special": false
         
     | 
| 27 | 
         
            +
                },
         
     | 
| 28 | 
         
            +
                "50255": {
         
     | 
| 29 | 
         
            +
                  "content": "                       ",
         
     | 
| 30 | 
         
            +
                  "lstrip": false,
         
     | 
| 31 | 
         
            +
                  "normalized": true,
         
     | 
| 32 | 
         
            +
                  "rstrip": false,
         
     | 
| 33 | 
         
            +
                  "single_word": false,
         
     | 
| 34 | 
         
            +
                  "special": false
         
     | 
| 35 | 
         
            +
                },
         
     | 
| 36 | 
         
            +
                "50256": {
         
     | 
| 37 | 
         
            +
                  "content": "                      ",
         
     | 
| 38 | 
         
            +
                  "lstrip": false,
         
     | 
| 39 | 
         
            +
                  "normalized": true,
         
     | 
| 40 | 
         
            +
                  "rstrip": false,
         
     | 
| 41 | 
         
            +
                  "single_word": false,
         
     | 
| 42 | 
         
            +
                  "special": false
         
     | 
| 43 | 
         
            +
                },
         
     | 
| 44 | 
         
            +
                "50257": {
         
     | 
| 45 | 
         
            +
                  "content": "                     ",
         
     | 
| 46 | 
         
            +
                  "lstrip": false,
         
     | 
| 47 | 
         
            +
                  "normalized": true,
         
     | 
| 48 | 
         
            +
                  "rstrip": false,
         
     | 
| 49 | 
         
            +
                  "single_word": false,
         
     | 
| 50 | 
         
            +
                  "special": false
         
     | 
| 51 | 
         
            +
                },
         
     | 
| 52 | 
         
            +
                "50258": {
         
     | 
| 53 | 
         
            +
                  "content": "                    ",
         
     | 
| 54 | 
         
            +
                  "lstrip": false,
         
     | 
| 55 | 
         
            +
                  "normalized": true,
         
     | 
| 56 | 
         
            +
                  "rstrip": false,
         
     | 
| 57 | 
         
            +
                  "single_word": false,
         
     | 
| 58 | 
         
            +
                  "special": false
         
     | 
| 59 | 
         
            +
                },
         
     | 
| 60 | 
         
            +
                "50259": {
         
     | 
| 61 | 
         
            +
                  "content": "                   ",
         
     | 
| 62 | 
         
            +
                  "lstrip": false,
         
     | 
| 63 | 
         
            +
                  "normalized": true,
         
     | 
| 64 | 
         
            +
                  "rstrip": false,
         
     | 
| 65 | 
         
            +
                  "single_word": false,
         
     | 
| 66 | 
         
            +
                  "special": false
         
     | 
| 67 | 
         
            +
                },
         
     | 
| 68 | 
         
            +
                "50260": {
         
     | 
| 69 | 
         
            +
                  "content": "                  ",
         
     | 
| 70 | 
         
            +
                  "lstrip": false,
         
     | 
| 71 | 
         
            +
                  "normalized": true,
         
     | 
| 72 | 
         
            +
                  "rstrip": false,
         
     | 
| 73 | 
         
            +
                  "single_word": false,
         
     | 
| 74 | 
         
            +
                  "special": false
         
     | 
| 75 | 
         
            +
                },
         
     | 
| 76 | 
         
            +
                "50261": {
         
     | 
| 77 | 
         
            +
                  "content": "                 ",
         
     | 
| 78 | 
         
            +
                  "lstrip": false,
         
     | 
| 79 | 
         
            +
                  "normalized": true,
         
     | 
| 80 | 
         
            +
                  "rstrip": false,
         
     | 
| 81 | 
         
            +
                  "single_word": false,
         
     | 
| 82 | 
         
            +
                  "special": false
         
     | 
| 83 | 
         
            +
                },
         
     | 
| 84 | 
         
            +
                "50262": {
         
     | 
| 85 | 
         
            +
                  "content": "                ",
         
     | 
| 86 | 
         
            +
                  "lstrip": false,
         
     | 
| 87 | 
         
            +
                  "normalized": true,
         
     | 
| 88 | 
         
            +
                  "rstrip": false,
         
     | 
| 89 | 
         
            +
                  "single_word": false,
         
     | 
| 90 | 
         
            +
                  "special": false
         
     | 
| 91 | 
         
            +
                },
         
     | 
| 92 | 
         
            +
                "50263": {
         
     | 
| 93 | 
         
            +
                  "content": "               ",
         
     | 
| 94 | 
         
            +
                  "lstrip": false,
         
     | 
| 95 | 
         
            +
                  "normalized": true,
         
     | 
| 96 | 
         
            +
                  "rstrip": false,
         
     | 
| 97 | 
         
            +
                  "single_word": false,
         
     | 
| 98 | 
         
            +
                  "special": false
         
     | 
| 99 | 
         
            +
                },
         
     | 
| 100 | 
         
            +
                "50264": {
         
     | 
| 101 | 
         
            +
                  "content": "              ",
         
     | 
| 102 | 
         
            +
                  "lstrip": false,
         
     | 
| 103 | 
         
            +
                  "normalized": true,
         
     | 
| 104 | 
         
            +
                  "rstrip": false,
         
     | 
| 105 | 
         
            +
                  "single_word": false,
         
     | 
| 106 | 
         
            +
                  "special": false
         
     | 
| 107 | 
         
            +
                },
         
     | 
| 108 | 
         
            +
                "50265": {
         
     | 
| 109 | 
         
            +
                  "content": "             ",
         
     | 
| 110 | 
         
            +
                  "lstrip": false,
         
     | 
| 111 | 
         
            +
                  "normalized": true,
         
     | 
| 112 | 
         
            +
                  "rstrip": false,
         
     | 
| 113 | 
         
            +
                  "single_word": false,
         
     | 
| 114 | 
         
            +
                  "special": false
         
     | 
| 115 | 
         
            +
                },
         
     | 
| 116 | 
         
            +
                "50266": {
         
     | 
| 117 | 
         
            +
                  "content": "            ",
         
     | 
| 118 | 
         
            +
                  "lstrip": false,
         
     | 
| 119 | 
         
            +
                  "normalized": true,
         
     | 
| 120 | 
         
            +
                  "rstrip": false,
         
     | 
| 121 | 
         
            +
                  "single_word": false,
         
     | 
| 122 | 
         
            +
                  "special": false
         
     | 
| 123 | 
         
            +
                },
         
     | 
| 124 | 
         
            +
                "50267": {
         
     | 
| 125 | 
         
            +
                  "content": "           ",
         
     | 
| 126 | 
         
            +
                  "lstrip": false,
         
     | 
| 127 | 
         
            +
                  "normalized": true,
         
     | 
| 128 | 
         
            +
                  "rstrip": false,
         
     | 
| 129 | 
         
            +
                  "single_word": false,
         
     | 
| 130 | 
         
            +
                  "special": false
         
     | 
| 131 | 
         
            +
                },
         
     | 
| 132 | 
         
            +
                "50268": {
         
     | 
| 133 | 
         
            +
                  "content": "          ",
         
     | 
| 134 | 
         
            +
                  "lstrip": false,
         
     | 
| 135 | 
         
            +
                  "normalized": true,
         
     | 
| 136 | 
         
            +
                  "rstrip": false,
         
     | 
| 137 | 
         
            +
                  "single_word": false,
         
     | 
| 138 | 
         
            +
                  "special": false
         
     | 
| 139 | 
         
            +
                },
         
     | 
| 140 | 
         
            +
                "50269": {
         
     | 
| 141 | 
         
            +
                  "content": "         ",
         
     | 
| 142 | 
         
            +
                  "lstrip": false,
         
     | 
| 143 | 
         
            +
                  "normalized": true,
         
     | 
| 144 | 
         
            +
                  "rstrip": false,
         
     | 
| 145 | 
         
            +
                  "single_word": false,
         
     | 
| 146 | 
         
            +
                  "special": false
         
     | 
| 147 | 
         
            +
                },
         
     | 
| 148 | 
         
            +
                "50270": {
         
     | 
| 149 | 
         
            +
                  "content": "        ",
         
     | 
| 150 | 
         
            +
                  "lstrip": false,
         
     | 
| 151 | 
         
            +
                  "normalized": true,
         
     | 
| 152 | 
         
            +
                  "rstrip": false,
         
     | 
| 153 | 
         
            +
                  "single_word": false,
         
     | 
| 154 | 
         
            +
                  "special": false
         
     | 
| 155 | 
         
            +
                },
         
     | 
| 156 | 
         
            +
                "50271": {
         
     | 
| 157 | 
         
            +
                  "content": "       ",
         
     | 
| 158 | 
         
            +
                  "lstrip": false,
         
     | 
| 159 | 
         
            +
                  "normalized": true,
         
     | 
| 160 | 
         
            +
                  "rstrip": false,
         
     | 
| 161 | 
         
            +
                  "single_word": false,
         
     | 
| 162 | 
         
            +
                  "special": false
         
     | 
| 163 | 
         
            +
                },
         
     | 
| 164 | 
         
            +
                "50272": {
         
     | 
| 165 | 
         
            +
                  "content": "      ",
         
     | 
| 166 | 
         
            +
                  "lstrip": false,
         
     | 
| 167 | 
         
            +
                  "normalized": true,
         
     | 
| 168 | 
         
            +
                  "rstrip": false,
         
     | 
| 169 | 
         
            +
                  "single_word": false,
         
     | 
| 170 | 
         
            +
                  "special": false
         
     | 
| 171 | 
         
            +
                },
         
     | 
| 172 | 
         
            +
                "50273": {
         
     | 
| 173 | 
         
            +
                  "content": "     ",
         
     | 
| 174 | 
         
            +
                  "lstrip": false,
         
     | 
| 175 | 
         
            +
                  "normalized": true,
         
     | 
| 176 | 
         
            +
                  "rstrip": false,
         
     | 
| 177 | 
         
            +
                  "single_word": false,
         
     | 
| 178 | 
         
            +
                  "special": false
         
     | 
| 179 | 
         
            +
                },
         
     | 
| 180 | 
         
            +
                "50274": {
         
     | 
| 181 | 
         
            +
                  "content": "    ",
         
     | 
| 182 | 
         
            +
                  "lstrip": false,
         
     | 
| 183 | 
         
            +
                  "normalized": true,
         
     | 
| 184 | 
         
            +
                  "rstrip": false,
         
     | 
| 185 | 
         
            +
                  "single_word": false,
         
     | 
| 186 | 
         
            +
                  "special": false
         
     | 
| 187 | 
         
            +
                },
         
     | 
| 188 | 
         
            +
                "50275": {
         
     | 
| 189 | 
         
            +
                  "content": "   ",
         
     | 
| 190 | 
         
            +
                  "lstrip": false,
         
     | 
| 191 | 
         
            +
                  "normalized": true,
         
     | 
| 192 | 
         
            +
                  "rstrip": false,
         
     | 
| 193 | 
         
            +
                  "single_word": false,
         
     | 
| 194 | 
         
            +
                  "special": false
         
     | 
| 195 | 
         
            +
                },
         
     | 
| 196 | 
         
            +
                "50276": {
         
     | 
| 197 | 
         
            +
                  "content": "  ",
         
     | 
| 198 | 
         
            +
                  "lstrip": false,
         
     | 
| 199 | 
         
            +
                  "normalized": true,
         
     | 
| 200 | 
         
            +
                  "rstrip": false,
         
     | 
| 201 | 
         
            +
                  "single_word": false,
         
     | 
| 202 | 
         
            +
                  "special": false
         
     | 
| 203 | 
         
             
                }
         
     | 
| 204 | 
         
             
              },
         
     | 
| 205 | 
         
            +
              "bos_token": "<|endoftext|>",
         
     | 
| 206 | 
         
            +
              "clean_up_tokenization_spaces": true,
         
     | 
| 207 | 
         
            +
              "eos_token": "<|endoftext|>",
         
     | 
| 208 | 
         
            +
              "model_max_length": 2048,
         
     | 
| 209 | 
         
            +
              "pad_token": "<|endoftext|>",
         
     | 
| 210 | 
         
            +
              "tokenizer_class": "GPTNeoXTokenizer",
         
     | 
| 211 | 
         
            +
              "unk_token": "<|endoftext|>"
         
     | 
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 212 | 
         
             
            }
         
     | 
    	
        training_args.bin
    CHANGED
    
    | 
         @@ -1,3 +1,3 @@ 
     | 
|
| 1 | 
         
             
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            -
            oid sha256: 
     | 
| 3 | 
         
             
            size 4475
         
     | 
| 
         | 
|
| 1 | 
         
             
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:96a468d1102d1d9c46447bce3b5a0a66d05ed25a247c1bee457c1a3971d444fa
         
     | 
| 3 | 
         
             
            size 4475
         
     |