Upload folder using huggingface_hub
Browse files- .gitattributes +2 -0
- checkpoint-146000/config.json +79 -0
- checkpoint-146000/generation_config.json +6 -0
- checkpoint-146000/model.safetensors +3 -0
- checkpoint-146000/optimizer.pt +3 -0
- checkpoint-146000/rng_state_0.pth +3 -0
- checkpoint-146000/rng_state_1.pth +3 -0
- checkpoint-146000/rng_state_10.pth +3 -0
- checkpoint-146000/rng_state_11.pth +3 -0
- checkpoint-146000/rng_state_12.pth +3 -0
- checkpoint-146000/rng_state_13.pth +3 -0
- checkpoint-146000/rng_state_14.pth +3 -0
- checkpoint-146000/rng_state_15.pth +3 -0
- checkpoint-146000/rng_state_2.pth +3 -0
- checkpoint-146000/rng_state_3.pth +3 -0
- checkpoint-146000/rng_state_4.pth +3 -0
- checkpoint-146000/rng_state_5.pth +3 -0
- checkpoint-146000/rng_state_6.pth +3 -0
- checkpoint-146000/rng_state_7.pth +3 -0
- checkpoint-146000/rng_state_8.pth +3 -0
- checkpoint-146000/rng_state_9.pth +3 -0
- checkpoint-146000/scheduler.pt +3 -0
- checkpoint-146000/trainer_state.json +3 -0
- checkpoint-146000/training_args.bin +3 -0
- checkpoint-148000/config.json +79 -0
- checkpoint-148000/generation_config.json +6 -0
- checkpoint-148000/model.safetensors +3 -0
- checkpoint-148000/optimizer.pt +3 -0
- checkpoint-148000/rng_state_0.pth +3 -0
- checkpoint-148000/rng_state_1.pth +3 -0
- checkpoint-148000/rng_state_10.pth +3 -0
- checkpoint-148000/rng_state_11.pth +3 -0
- checkpoint-148000/rng_state_12.pth +3 -0
- checkpoint-148000/rng_state_13.pth +3 -0
- checkpoint-148000/rng_state_14.pth +3 -0
- checkpoint-148000/rng_state_15.pth +3 -0
- checkpoint-148000/rng_state_2.pth +3 -0
- checkpoint-148000/rng_state_3.pth +3 -0
- checkpoint-148000/rng_state_4.pth +3 -0
- checkpoint-148000/rng_state_5.pth +3 -0
- checkpoint-148000/rng_state_6.pth +3 -0
- checkpoint-148000/rng_state_7.pth +3 -0
- checkpoint-148000/rng_state_8.pth +3 -0
- checkpoint-148000/rng_state_9.pth +3 -0
- checkpoint-148000/scheduler.pt +3 -0
- checkpoint-148000/trainer_state.json +3 -0
- checkpoint-148000/training_args.bin +3 -0
.gitattributes
CHANGED
|
@@ -58,3 +58,5 @@ checkpoint-130000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
|
|
| 58 |
checkpoint-136000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
|
| 59 |
checkpoint-138000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
|
| 60 |
checkpoint-140000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
| 58 |
checkpoint-136000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
|
| 59 |
checkpoint-138000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
|
| 60 |
checkpoint-140000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
|
| 61 |
+
checkpoint-146000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
|
| 62 |
+
checkpoint-148000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
|
checkpoint-146000/config.json
ADDED
|
@@ -0,0 +1,79 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"ar_steps": 1,
|
| 3 |
+
"architectures": [
|
| 4 |
+
"DiffVLMDiffusion"
|
| 5 |
+
],
|
| 6 |
+
"attention_dropout": 0.0,
|
| 7 |
+
"bos_token_id": 151643,
|
| 8 |
+
"condition_layer": -1,
|
| 9 |
+
"eos_token_id": 151645,
|
| 10 |
+
"hidden_act": "silu",
|
| 11 |
+
"hidden_size": 1536,
|
| 12 |
+
"image_token_id": 151655,
|
| 13 |
+
"img_cross_attention_dim": 2048,
|
| 14 |
+
"img_diffuser_depth": 2,
|
| 15 |
+
"img_ffn_dim_multiplier": null,
|
| 16 |
+
"img_hidden_size": 1536,
|
| 17 |
+
"img_multiple_of": 256,
|
| 18 |
+
"img_norm_eps": 1e-05,
|
| 19 |
+
"img_num_attention_heads": 12,
|
| 20 |
+
"img_num_kv_heads": 12,
|
| 21 |
+
"img_qk_norm": true,
|
| 22 |
+
"in_channels": 32,
|
| 23 |
+
"initializer_range": 0.02,
|
| 24 |
+
"inject_img_diffuser": false,
|
| 25 |
+
"input_size": 32,
|
| 26 |
+
"intermediate_size": 8960,
|
| 27 |
+
"layer_group_size": 7,
|
| 28 |
+
"layerwise_start_idx": 0,
|
| 29 |
+
"lora_alpha": 128,
|
| 30 |
+
"lora_bias": "none",
|
| 31 |
+
"lora_dropout": 0.05,
|
| 32 |
+
"lora_enable": false,
|
| 33 |
+
"lora_r": 64,
|
| 34 |
+
"max_position_embeddings": 32768,
|
| 35 |
+
"max_window_layers": 28,
|
| 36 |
+
"model_type": "qwen2_vl",
|
| 37 |
+
"non_linearity": 1,
|
| 38 |
+
"norm_elementwise_affine": true,
|
| 39 |
+
"num_attention_heads": 12,
|
| 40 |
+
"num_hidden_layers": 28,
|
| 41 |
+
"num_key_value_heads": 2,
|
| 42 |
+
"patch_size": 2,
|
| 43 |
+
"repa_coeff": 0.1,
|
| 44 |
+
"repa_layers": "2",
|
| 45 |
+
"repa_shared": false,
|
| 46 |
+
"rms_norm_eps": 1e-06,
|
| 47 |
+
"rope_scaling": {
|
| 48 |
+
"mrope_section": [
|
| 49 |
+
16,
|
| 50 |
+
24,
|
| 51 |
+
24
|
| 52 |
+
],
|
| 53 |
+
"rope_type": "default",
|
| 54 |
+
"type": "default"
|
| 55 |
+
},
|
| 56 |
+
"rope_theta": 1000000.0,
|
| 57 |
+
"sample_size": 128,
|
| 58 |
+
"sampling_steps": 28,
|
| 59 |
+
"sliding_window": null,
|
| 60 |
+
"tie_word_embeddings": true,
|
| 61 |
+
"torch_dtype": "bfloat16",
|
| 62 |
+
"transformers_version": "4.47.0",
|
| 63 |
+
"use_cache": true,
|
| 64 |
+
"use_repa": false,
|
| 65 |
+
"use_residual_attn": true,
|
| 66 |
+
"use_sliding_window": false,
|
| 67 |
+
"vae_path": "mit-han-lab/dc-ae-f32c32-sana-1.1-diffusers",
|
| 68 |
+
"video_token_id": 151656,
|
| 69 |
+
"vision_config": {
|
| 70 |
+
"hidden_size": 1536,
|
| 71 |
+
"in_chans": 3,
|
| 72 |
+
"model_type": "qwen2_vl",
|
| 73 |
+
"spatial_patch_size": 14
|
| 74 |
+
},
|
| 75 |
+
"vision_end_token_id": 151653,
|
| 76 |
+
"vision_start_token_id": 151652,
|
| 77 |
+
"vision_token_id": 151654,
|
| 78 |
+
"vocab_size": 151936
|
| 79 |
+
}
|
checkpoint-146000/generation_config.json
ADDED
|
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_from_model_config": true,
|
| 3 |
+
"bos_token_id": 151643,
|
| 4 |
+
"eos_token_id": 151645,
|
| 5 |
+
"transformers_version": "4.47.0"
|
| 6 |
+
}
|
checkpoint-146000/model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f6e2cf59adea94dec0bb8e21abbb5ad537f831caab40171c37ad6cd143679db2
|
| 3 |
+
size 4539487822
|
checkpoint-146000/optimizer.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5f9b605320e5bcc1b9302b25664d845365c3a1ba2e65824e41a60e148bee826d
|
| 3 |
+
size 6632345482
|
checkpoint-146000/rng_state_0.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:757a6d5e9900ded973ca0aba59c46cbc1c7a1414d229ba54e1c9df38cafdfe82
|
| 3 |
+
size 15984
|
checkpoint-146000/rng_state_1.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ba32a2523c9ff34c222919f87e65ad14bce39f7c5cbfb1d38319b108481f6375
|
| 3 |
+
size 15984
|
checkpoint-146000/rng_state_10.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9ab390f32bfec962d414f0950ad2a046f59adfa51270aaf213ccf18dd4b3a65a
|
| 3 |
+
size 15997
|
checkpoint-146000/rng_state_11.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d2bb7b189d3ab873b14d5fa8fe427919464aad7817c8778cf007e635eade899d
|
| 3 |
+
size 15997
|
checkpoint-146000/rng_state_12.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:57e3b314b348c34f7dbc06121dcfd78259fc6356c31a03683e8dc8fdee1c3ab1
|
| 3 |
+
size 15997
|
checkpoint-146000/rng_state_13.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9d551ab6d7edb6ecaf2aa22f955a157409ca5b0aca76b22c39b5504734e24c77
|
| 3 |
+
size 15997
|
checkpoint-146000/rng_state_14.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1eca0d867da1cba0d5ad983d0efab42a85eb3875fa0ffd30cf31385fad298627
|
| 3 |
+
size 15997
|
checkpoint-146000/rng_state_15.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:59998c7ce1ca5fdde96b7d8a74938d8b2705c8b74c39243d04adffaacdb670b7
|
| 3 |
+
size 15997
|
checkpoint-146000/rng_state_2.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f222102e88ae7687f471d708b6e176d9f8198fe0527053c4367da94c9c762ce2
|
| 3 |
+
size 15984
|
checkpoint-146000/rng_state_3.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6fba274c279fb6b3622d2b3eed3a5c4227ec5577a43719bb877d94687e236915
|
| 3 |
+
size 15984
|
checkpoint-146000/rng_state_4.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3cc8133fde118ecc84965f60636c744c58ae15db3357dc9684e823847d28dd6c
|
| 3 |
+
size 15984
|
checkpoint-146000/rng_state_5.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:50e3586e3fa6a23d9a1e88080bc8fd8d607b593bb1c9195aaaea67e53bb835bd
|
| 3 |
+
size 15984
|
checkpoint-146000/rng_state_6.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ab108e4c3d560d015e5d01f4292bc49c8ea2d207a3681b8816feb1ae58072fb4
|
| 3 |
+
size 15984
|
checkpoint-146000/rng_state_7.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4a4180dc4af97a5cbb961fa990993e6ab09579458b4f98b858eb1b79717ecac2
|
| 3 |
+
size 15984
|
checkpoint-146000/rng_state_8.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8fddf3d21ab75beadb8f7a29a3a7fe4639531177bb51dfe46ac3c1c7496b9420
|
| 3 |
+
size 15984
|
checkpoint-146000/rng_state_9.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9d59e6f663814199ab493c44246b9813808f15a0a8ac7cc76e9bb7153b84f999
|
| 3 |
+
size 15984
|
checkpoint-146000/scheduler.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:07a16afba48b165ea61af0ce817c3459f3ecaad53ba968fbbb072410ee66b8d1
|
| 3 |
+
size 1064
|
checkpoint-146000/trainer_state.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3006ac2b48d495348d86dc2af2336543079ee01c835fefc0d5f0c1a48613e87b
|
| 3 |
+
size 22562942
|
checkpoint-146000/training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9154b5e1923b09fb0f9d0965b6e2a2f29aa4552e1522e2ce13d538188c5558cc
|
| 3 |
+
size 6008
|
checkpoint-148000/config.json
ADDED
|
@@ -0,0 +1,79 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"ar_steps": 1,
|
| 3 |
+
"architectures": [
|
| 4 |
+
"DiffVLMDiffusion"
|
| 5 |
+
],
|
| 6 |
+
"attention_dropout": 0.0,
|
| 7 |
+
"bos_token_id": 151643,
|
| 8 |
+
"condition_layer": -1,
|
| 9 |
+
"eos_token_id": 151645,
|
| 10 |
+
"hidden_act": "silu",
|
| 11 |
+
"hidden_size": 1536,
|
| 12 |
+
"image_token_id": 151655,
|
| 13 |
+
"img_cross_attention_dim": 2048,
|
| 14 |
+
"img_diffuser_depth": 2,
|
| 15 |
+
"img_ffn_dim_multiplier": null,
|
| 16 |
+
"img_hidden_size": 1536,
|
| 17 |
+
"img_multiple_of": 256,
|
| 18 |
+
"img_norm_eps": 1e-05,
|
| 19 |
+
"img_num_attention_heads": 12,
|
| 20 |
+
"img_num_kv_heads": 12,
|
| 21 |
+
"img_qk_norm": true,
|
| 22 |
+
"in_channels": 32,
|
| 23 |
+
"initializer_range": 0.02,
|
| 24 |
+
"inject_img_diffuser": false,
|
| 25 |
+
"input_size": 32,
|
| 26 |
+
"intermediate_size": 8960,
|
| 27 |
+
"layer_group_size": 7,
|
| 28 |
+
"layerwise_start_idx": 0,
|
| 29 |
+
"lora_alpha": 128,
|
| 30 |
+
"lora_bias": "none",
|
| 31 |
+
"lora_dropout": 0.05,
|
| 32 |
+
"lora_enable": false,
|
| 33 |
+
"lora_r": 64,
|
| 34 |
+
"max_position_embeddings": 32768,
|
| 35 |
+
"max_window_layers": 28,
|
| 36 |
+
"model_type": "qwen2_vl",
|
| 37 |
+
"non_linearity": 1,
|
| 38 |
+
"norm_elementwise_affine": true,
|
| 39 |
+
"num_attention_heads": 12,
|
| 40 |
+
"num_hidden_layers": 28,
|
| 41 |
+
"num_key_value_heads": 2,
|
| 42 |
+
"patch_size": 2,
|
| 43 |
+
"repa_coeff": 0.1,
|
| 44 |
+
"repa_layers": "2",
|
| 45 |
+
"repa_shared": false,
|
| 46 |
+
"rms_norm_eps": 1e-06,
|
| 47 |
+
"rope_scaling": {
|
| 48 |
+
"mrope_section": [
|
| 49 |
+
16,
|
| 50 |
+
24,
|
| 51 |
+
24
|
| 52 |
+
],
|
| 53 |
+
"rope_type": "default",
|
| 54 |
+
"type": "default"
|
| 55 |
+
},
|
| 56 |
+
"rope_theta": 1000000.0,
|
| 57 |
+
"sample_size": 128,
|
| 58 |
+
"sampling_steps": 28,
|
| 59 |
+
"sliding_window": null,
|
| 60 |
+
"tie_word_embeddings": true,
|
| 61 |
+
"torch_dtype": "bfloat16",
|
| 62 |
+
"transformers_version": "4.47.0",
|
| 63 |
+
"use_cache": true,
|
| 64 |
+
"use_repa": false,
|
| 65 |
+
"use_residual_attn": true,
|
| 66 |
+
"use_sliding_window": false,
|
| 67 |
+
"vae_path": "mit-han-lab/dc-ae-f32c32-sana-1.1-diffusers",
|
| 68 |
+
"video_token_id": 151656,
|
| 69 |
+
"vision_config": {
|
| 70 |
+
"hidden_size": 1536,
|
| 71 |
+
"in_chans": 3,
|
| 72 |
+
"model_type": "qwen2_vl",
|
| 73 |
+
"spatial_patch_size": 14
|
| 74 |
+
},
|
| 75 |
+
"vision_end_token_id": 151653,
|
| 76 |
+
"vision_start_token_id": 151652,
|
| 77 |
+
"vision_token_id": 151654,
|
| 78 |
+
"vocab_size": 151936
|
| 79 |
+
}
|
checkpoint-148000/generation_config.json
ADDED
|
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_from_model_config": true,
|
| 3 |
+
"bos_token_id": 151643,
|
| 4 |
+
"eos_token_id": 151645,
|
| 5 |
+
"transformers_version": "4.47.0"
|
| 6 |
+
}
|
checkpoint-148000/model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0115d0f4b4891c2dbdb0ce224e7e7e83bc2ab8f6b3fe6883a637f4a4309b3b48
|
| 3 |
+
size 4539487822
|
checkpoint-148000/optimizer.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:243fa906bf849f6f64d1e160eedc41408401796cd0950f58490afdd128dd9e70
|
| 3 |
+
size 6632345482
|
checkpoint-148000/rng_state_0.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ffe9ff8d8879b3ff0756fe5687daeea5c91d3868da1dac3f835a07fbafd16076
|
| 3 |
+
size 15984
|
checkpoint-148000/rng_state_1.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9d45063af2fa57577617eba8d15d139cfdff3b82820dee3fedf32fdb40631cd8
|
| 3 |
+
size 15984
|
checkpoint-148000/rng_state_10.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:607eb54d19f8889ae330aa98fe7bd0fea7af9f6f55481ee3936ee589ce05894a
|
| 3 |
+
size 15997
|
checkpoint-148000/rng_state_11.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b45869cf53d3d507518b8648f05bbd2ad8d4a500cc5f7d7ff2b189a73e480b5f
|
| 3 |
+
size 15997
|
checkpoint-148000/rng_state_12.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4aa9daa47de0578e7e9894731e8ecfec5b63daf1d0048af9e746c6deff261c5e
|
| 3 |
+
size 15997
|
checkpoint-148000/rng_state_13.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:91ad40fcb3d9c344f24230df83489a6e4204127438e8faaea33cc11330ce8f07
|
| 3 |
+
size 15997
|
checkpoint-148000/rng_state_14.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:98355cbd55c2f24b16a8bd9e9301792a234d4f41cf96b9212ee16af876c68981
|
| 3 |
+
size 15997
|
checkpoint-148000/rng_state_15.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6004bad76aa9791af3c16c650ad9cd82b6d20d7801b84cb0c4e159210a31c7e4
|
| 3 |
+
size 15997
|
checkpoint-148000/rng_state_2.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:cbff1a05d878b69c4d7f8ada60c56b9353e610a5ce1e2f748a3d5c47e114d5fe
|
| 3 |
+
size 15984
|
checkpoint-148000/rng_state_3.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ad654ad4782304f212bc2f7cfd2f79b30dec79bfc9a30edeca78fe862cb0607d
|
| 3 |
+
size 15984
|
checkpoint-148000/rng_state_4.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0333f1cc7caf6da072ca0f1ec2786dbfdd04ce60a550b04c6640ef3c23ffb4b7
|
| 3 |
+
size 15984
|
checkpoint-148000/rng_state_5.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a1a18e334316fa84342cd10fa70179eb58f7b00a8a6232fc953cd254d9e01568
|
| 3 |
+
size 15984
|
checkpoint-148000/rng_state_6.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b8b9c3fdbf0489a15abe53601f75af7964f610be1ada16bf57b8437318c524b0
|
| 3 |
+
size 15984
|
checkpoint-148000/rng_state_7.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9bae277416c1c31e83a87d9f2b86c8ca35214510a2b22a6efd084a1717ebb040
|
| 3 |
+
size 15984
|
checkpoint-148000/rng_state_8.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6d983d94555fc238f0d265430b96cfcda7783e77853aeed9a4c301106d767381
|
| 3 |
+
size 15984
|
checkpoint-148000/rng_state_9.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d331e1a299254e3470161a9d3f16313ec357405b8ac655d671ef7aa86ac0f2a6
|
| 3 |
+
size 15984
|
checkpoint-148000/scheduler.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8ede12b49b56707a56eb6cff6946ba010bfc21e4d6c9d40bd05227680119aabc
|
| 3 |
+
size 1064
|
checkpoint-148000/trainer_state.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e66f86217b85d3a2bc25312149c988009704621e5ac5554b9bc5c73c777e5260
|
| 3 |
+
size 22873515
|
checkpoint-148000/training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:55ebe8280cc9de8749c2676fb8565d0bd9655cfb03330c8e587160de7f3a553e
|
| 3 |
+
size 6008
|