iaravagni commited on
Commit
d04a889
·
verified ·
1 Parent(s): 93f3cce

Upload folder using huggingface_hub

Browse files
finetuned_ttm_model/config.json ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "adaptive_patching_levels": 0,
3
+ "architectures": [
4
+ "TinyTimeMixerForPrediction"
5
+ ],
6
+ "categorical_vocab_size_list": null,
7
+ "context_length": 52,
8
+ "d_model": 20,
9
+ "d_model_scale": 5,
10
+ "decoder_adaptive_patching_levels": 0,
11
+ "decoder_d_model": 16,
12
+ "decoder_d_model_scale": 4,
13
+ "decoder_mode": "mix_channel",
14
+ "decoder_num_layers": 4,
15
+ "decoder_raw_residual": false,
16
+ "distribution_output": "student_t",
17
+ "dropout": 0.3,
18
+ "enable_forecast_channel_mixing": false,
19
+ "exogenous_channel_indices": null,
20
+ "expansion_factor": 3,
21
+ "fcm_context_length": 1,
22
+ "fcm_gated_attn": true,
23
+ "fcm_mix_layers": 3,
24
+ "fcm_prepend_past": true,
25
+ "fcm_prepend_past_offset": null,
26
+ "fcm_use_mixer": true,
27
+ "frequency_token_vocab_size": 10,
28
+ "gated_attn": true,
29
+ "head_dropout": 0.3,
30
+ "huber_delta": 1,
31
+ "init_embed": "pytorch",
32
+ "init_linear": "pytorch",
33
+ "init_processing": true,
34
+ "init_std": 0.02,
35
+ "loss": "mse",
36
+ "mask_value": 0,
37
+ "masked_context_length": null,
38
+ "mode": "common_channel",
39
+ "model_type": "tinytimemixer",
40
+ "norm_eps": 1e-05,
41
+ "norm_mlp": "LayerNorm",
42
+ "num_input_channels": 8,
43
+ "num_layers": 12,
44
+ "num_parallel_samples": 100,
45
+ "num_patches": 14,
46
+ "patch_last": true,
47
+ "patch_length": 4,
48
+ "patch_stride": 4,
49
+ "positional_encoding_type": "sincos",
50
+ "post_init": false,
51
+ "prediction_channel_indices": [
52
+ 0
53
+ ],
54
+ "prediction_filter_length": 6,
55
+ "prediction_length": 16,
56
+ "quantile": 0.5,
57
+ "resolution_prefix_tuning": true,
58
+ "scaling": "std",
59
+ "self_attn": false,
60
+ "self_attn_heads": 1,
61
+ "stride_ratio": 1,
62
+ "torch_dtype": "float32",
63
+ "transformers_version": "4.51.3",
64
+ "use_decoder": true,
65
+ "use_positional_encoding": false
66
+ }
finetuned_ttm_model/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a697a662117fd92d5a872b07456e7d507f2d7cf51db5030baa015bd4d18ce501
3
+ size 332048
logs/events.out.tfevents.1745270714.dbfd7c3c91c7.1623.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a21c50543671d323491ff23170c9b9a258b56102d0b72350d1f4a9af71008b62
3
+ size 12124
output/checkpoint-2786/config.json ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "adaptive_patching_levels": 0,
3
+ "architectures": [
4
+ "TinyTimeMixerForPrediction"
5
+ ],
6
+ "categorical_vocab_size_list": null,
7
+ "context_length": 52,
8
+ "d_model": 20,
9
+ "d_model_scale": 5,
10
+ "decoder_adaptive_patching_levels": 0,
11
+ "decoder_d_model": 16,
12
+ "decoder_d_model_scale": 4,
13
+ "decoder_mode": "mix_channel",
14
+ "decoder_num_layers": 4,
15
+ "decoder_raw_residual": false,
16
+ "distribution_output": "student_t",
17
+ "dropout": 0.3,
18
+ "enable_forecast_channel_mixing": false,
19
+ "exogenous_channel_indices": null,
20
+ "expansion_factor": 3,
21
+ "fcm_context_length": 1,
22
+ "fcm_gated_attn": true,
23
+ "fcm_mix_layers": 3,
24
+ "fcm_prepend_past": true,
25
+ "fcm_prepend_past_offset": null,
26
+ "fcm_use_mixer": true,
27
+ "frequency_token_vocab_size": 10,
28
+ "gated_attn": true,
29
+ "head_dropout": 0.3,
30
+ "huber_delta": 1,
31
+ "init_embed": "pytorch",
32
+ "init_linear": "pytorch",
33
+ "init_processing": true,
34
+ "init_std": 0.02,
35
+ "loss": "mse",
36
+ "mask_value": 0,
37
+ "masked_context_length": null,
38
+ "mode": "common_channel",
39
+ "model_type": "tinytimemixer",
40
+ "norm_eps": 1e-05,
41
+ "norm_mlp": "LayerNorm",
42
+ "num_input_channels": 8,
43
+ "num_layers": 12,
44
+ "num_parallel_samples": 100,
45
+ "num_patches": 14,
46
+ "patch_last": true,
47
+ "patch_length": 4,
48
+ "patch_stride": 4,
49
+ "positional_encoding_type": "sincos",
50
+ "post_init": false,
51
+ "prediction_channel_indices": [
52
+ 0
53
+ ],
54
+ "prediction_filter_length": 6,
55
+ "prediction_length": 16,
56
+ "quantile": 0.5,
57
+ "resolution_prefix_tuning": true,
58
+ "scaling": "std",
59
+ "self_attn": false,
60
+ "self_attn_heads": 1,
61
+ "stride_ratio": 1,
62
+ "torch_dtype": "float32",
63
+ "transformers_version": "4.51.3",
64
+ "use_decoder": true,
65
+ "use_positional_encoding": false
66
+ }
output/checkpoint-2786/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a697a662117fd92d5a872b07456e7d507f2d7cf51db5030baa015bd4d18ce501
3
+ size 332048
output/checkpoint-2786/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5de32065dbd7d9e934710939301d37c86ed9519d38bbe7684d87e3c798f467be
3
+ size 811194
output/checkpoint-2786/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3dd30c2f71a549d40d88501427dd6c934beddaec3a87220f6f567c9139ce997e
3
+ size 13990
output/checkpoint-2786/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:28758e9bf3a0e44359800d4da07f861242d93e71b6cd40bf30e1f8d873dba782
3
+ size 1320
output/checkpoint-2786/trainer_state.json ADDED
@@ -0,0 +1,148 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 2786,
3
+ "best_metric": 72.45570373535156,
4
+ "best_model_checkpoint": "model/output/checkpoint-2786",
5
+ "epoch": 7.0,
6
+ "eval_steps": 500,
7
+ "global_step": 2786,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 1.0,
14
+ "grad_norm": 1758.0869140625,
15
+ "learning_rate": 0.00033533537238222464,
16
+ "loss": 165.6237,
17
+ "step": 398
18
+ },
19
+ {
20
+ "epoch": 1.0,
21
+ "eval_loss": 76.47968292236328,
22
+ "eval_runtime": 20.0392,
23
+ "eval_samples_per_second": 202.802,
24
+ "eval_steps_per_second": 3.194,
25
+ "step": 398
26
+ },
27
+ {
28
+ "epoch": 2.0,
29
+ "grad_norm": 661.80322265625,
30
+ "learning_rate": 0.0009673161575995766,
31
+ "loss": 101.9303,
32
+ "step": 796
33
+ },
34
+ {
35
+ "epoch": 2.0,
36
+ "eval_loss": 74.72091674804688,
37
+ "eval_runtime": 16.517,
38
+ "eval_samples_per_second": 246.05,
39
+ "eval_steps_per_second": 3.875,
40
+ "step": 796
41
+ },
42
+ {
43
+ "epoch": 3.0,
44
+ "grad_norm": 233.8292999267578,
45
+ "learning_rate": 0.0016381468554296547,
46
+ "loss": 96.1758,
47
+ "step": 1194
48
+ },
49
+ {
50
+ "epoch": 3.0,
51
+ "eval_loss": 76.53939819335938,
52
+ "eval_runtime": 16.7409,
53
+ "eval_samples_per_second": 242.759,
54
+ "eval_steps_per_second": 3.823,
55
+ "step": 1194
56
+ },
57
+ {
58
+ "epoch": 4.0,
59
+ "grad_norm": 2881.05810546875,
60
+ "learning_rate": 0.001989264776725461,
61
+ "loss": 92.2926,
62
+ "step": 1592
63
+ },
64
+ {
65
+ "epoch": 4.0,
66
+ "eval_loss": 74.74008178710938,
67
+ "eval_runtime": 18.2222,
68
+ "eval_samples_per_second": 223.025,
69
+ "eval_steps_per_second": 3.512,
70
+ "step": 1592
71
+ },
72
+ {
73
+ "epoch": 5.0,
74
+ "grad_norm": 1436.7108154296875,
75
+ "learning_rate": 0.001967294993859577,
76
+ "loss": 91.7396,
77
+ "step": 1990
78
+ },
79
+ {
80
+ "epoch": 5.0,
81
+ "eval_loss": 74.24183654785156,
82
+ "eval_runtime": 16.4508,
83
+ "eval_samples_per_second": 247.04,
84
+ "eval_steps_per_second": 3.89,
85
+ "step": 1990
86
+ },
87
+ {
88
+ "epoch": 6.0,
89
+ "grad_norm": 1180.649169921875,
90
+ "learning_rate": 0.0018380887525394208,
91
+ "loss": 89.8515,
92
+ "step": 2388
93
+ },
94
+ {
95
+ "epoch": 6.0,
96
+ "eval_loss": 79.67572021484375,
97
+ "eval_runtime": 16.8586,
98
+ "eval_samples_per_second": 241.063,
99
+ "eval_steps_per_second": 3.796,
100
+ "step": 2388
101
+ },
102
+ {
103
+ "epoch": 7.0,
104
+ "grad_norm": 1331.4036865234375,
105
+ "learning_rate": 0.0016234913078995259,
106
+ "loss": 87.4604,
107
+ "step": 2786
108
+ },
109
+ {
110
+ "epoch": 7.0,
111
+ "eval_loss": 72.45570373535156,
112
+ "eval_runtime": 16.5884,
113
+ "eval_samples_per_second": 244.991,
114
+ "eval_steps_per_second": 3.858,
115
+ "step": 2786
116
+ }
117
+ ],
118
+ "logging_steps": 500,
119
+ "max_steps": 5572,
120
+ "num_input_tokens_seen": 0,
121
+ "num_train_epochs": 14,
122
+ "save_steps": 500,
123
+ "stateful_callbacks": {
124
+ "EarlyStoppingCallback": {
125
+ "args": {
126
+ "early_stopping_patience": 5,
127
+ "early_stopping_threshold": 0.001
128
+ },
129
+ "attributes": {
130
+ "early_stopping_patience_counter": 0
131
+ }
132
+ },
133
+ "TrainerControl": {
134
+ "args": {
135
+ "should_epoch_stop": false,
136
+ "should_evaluate": false,
137
+ "should_log": false,
138
+ "should_save": true,
139
+ "should_training_stop": false
140
+ },
141
+ "attributes": {}
142
+ }
143
+ },
144
+ "total_flos": 32459312820480.0,
145
+ "train_batch_size": 64,
146
+ "trial_name": null,
147
+ "trial_params": null
148
+ }
output/checkpoint-2786/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a6232fa152053185effc28232db54f86c5b5e5f051938087ee56ad17a1212bf3
3
+ size 5240