tejas1206 commited on
Commit
ed9aeb0
·
verified ·
1 Parent(s): 7a8f152

End of training

Browse files
README.md CHANGED
@@ -34,7 +34,7 @@ This model was trained with SFT.
34
 
35
  ### Framework versions
36
 
37
- - TRL: 0.15.1
38
  - Transformers: 4.48.3
39
  - Pytorch: 2.5.1+cu124
40
  - Datasets: 3.3.2
 
34
 
35
  ### Framework versions
36
 
37
+ - TRL: 0.15.2
38
  - Transformers: 4.48.3
39
  - Pytorch: 2.5.1+cu124
40
  - Datasets: 3.3.2
adapter_config.json CHANGED
@@ -5,7 +5,7 @@
5
  "bias": "lora_only",
6
  "eva_config": null,
7
  "exclude_modules": null,
8
- "fan_in_fan_out": true,
9
  "inference_mode": true,
10
  "init_lora_weights": true,
11
  "layer_replication": null,
@@ -23,9 +23,9 @@
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
 
26
  "c_proj",
27
- "c_attn",
28
- "c_fc"
29
  ],
30
  "task_type": "CAUSAL_LM",
31
  "use_dora": false,
 
5
  "bias": "lora_only",
6
  "eva_config": null,
7
  "exclude_modules": null,
8
+ "fan_in_fan_out": false,
9
  "inference_mode": true,
10
  "init_lora_weights": true,
11
  "layer_replication": null,
 
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
+ "c_fc",
27
  "c_proj",
28
+ "c_attn"
 
29
  ],
30
  "task_type": "CAUSAL_LM",
31
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5a985c949388a45893d301d366e100fb9422b24a62cd6d5cd8084759f27012ed
3
- size 397959576
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8e69dd0819ff9b5022a5c890d9fb44c2b10ee605bc39bcff1533232a58547332
3
+ size 4730632
special_tokens_map.json CHANGED
@@ -1,22 +1,6 @@
1
  {
2
- "additional_special_tokens": [
3
- {
4
- "content": "<|im_start|>",
5
- "lstrip": false,
6
- "normalized": false,
7
- "rstrip": false,
8
- "single_word": false
9
- },
10
- {
11
- "content": "<|im_end|>",
12
- "lstrip": false,
13
- "normalized": false,
14
- "rstrip": false,
15
- "single_word": false
16
- }
17
- ],
18
- "bos_token": "<|im_start|>",
19
- "eos_token": "<|im_end|>",
20
- "pad_token": "<|im_end|>",
21
  "unk_token": "<|endoftext|>"
22
  }
 
1
  {
2
+ "bos_token": "<|endoftext|>",
3
+ "eos_token": "<|endoftext|>",
4
+ "pad_token": "<|endoftext|>",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
  "unk_token": "<|endoftext|>"
6
  }
tokenizer.json CHANGED
@@ -1,6 +1,11 @@
1
  {
2
  "version": "1.0",
3
- "truncation": null,
 
 
 
 
 
4
  "padding": null,
5
  "added_tokens": [
6
  {
@@ -38,24 +43,6 @@
38
  "rstrip": false,
39
  "normalized": false,
40
  "special": true
41
- },
42
- {
43
- "id": 64000,
44
- "content": "<|im_start|>",
45
- "single_word": false,
46
- "lstrip": false,
47
- "rstrip": false,
48
- "normalized": false,
49
- "special": true
50
- },
51
- {
52
- "id": 64001,
53
- "content": "<|im_end|>",
54
- "single_word": false,
55
- "lstrip": false,
56
- "rstrip": false,
57
- "normalized": false,
58
- "special": true
59
  }
60
  ],
61
  "normalizer": null,
 
1
  {
2
  "version": "1.0",
3
+ "truncation": {
4
+ "direction": "Right",
5
+ "max_length": 256,
6
+ "strategy": "LongestFirst",
7
+ "stride": 0
8
+ },
9
  "padding": null,
10
  "added_tokens": [
11
  {
 
43
  "rstrip": false,
44
  "normalized": false,
45
  "special": true
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
46
  }
47
  ],
48
  "normalizer": null,
tokenizer_config.json CHANGED
@@ -32,36 +32,14 @@
32
  "rstrip": false,
33
  "single_word": false,
34
  "special": true
35
- },
36
- "64000": {
37
- "content": "<|im_start|>",
38
- "lstrip": false,
39
- "normalized": false,
40
- "rstrip": false,
41
- "single_word": false,
42
- "special": true
43
- },
44
- "64001": {
45
- "content": "<|im_end|>",
46
- "lstrip": false,
47
- "normalized": false,
48
- "rstrip": false,
49
- "single_word": false,
50
- "special": true
51
  }
52
  },
53
- "additional_special_tokens": [
54
- "<|im_start|>",
55
- "<|im_end|>"
56
- ],
57
- "bos_token": "<|im_start|>",
58
- "chat_template": "{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
59
  "clean_up_tokenization_spaces": false,
60
- "eos_token": "<|im_end|>",
61
  "extra_special_tokens": {},
62
  "model_max_length": 1000000000000000019884624838656,
63
- "pad_token": "<|im_end|>",
64
  "tokenizer_class": "GPT2Tokenizer",
65
- "trust_remote_model": true,
66
  "unk_token": "<|endoftext|>"
67
  }
 
32
  "rstrip": false,
33
  "single_word": false,
34
  "special": true
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
35
  }
36
  },
37
+ "bos_token": "<|endoftext|>",
 
 
 
 
 
38
  "clean_up_tokenization_spaces": false,
39
+ "eos_token": "<|endoftext|>",
40
  "extra_special_tokens": {},
41
  "model_max_length": 1000000000000000019884624838656,
42
+ "pad_token": "<|endoftext|>",
43
  "tokenizer_class": "GPT2Tokenizer",
 
44
  "unk_token": "<|endoftext|>"
45
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:44329eba918abad06f64871058cbe7feec8541eb6a0cd805c7aca3a98a599272
3
  size 5560
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9907bdf5c44b343a9fb589f61e65b0b7b8baff0d5edff6054e6ed0f0ae46acfc
3
  size 5560