best val_rmse 0.0458
Browse files- config.json +3 -1
- pytorch_model.bin +2 -2
config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"best_val_loss": 0.
|
3 |
"comet_ai_key": null,
|
4 |
"context_observations": {
|
5 |
"add_rem": true,
|
@@ -135,6 +135,7 @@
|
|
135 |
"activation": "ReLU",
|
136 |
"aggregator_num_heads": 8,
|
137 |
"aggregator_type": "attention",
|
|
|
138 |
"decoder_hidden_dim": 128,
|
139 |
"decoder_name": "TransformerDecoder",
|
140 |
"decoder_num_layers": 2,
|
@@ -157,6 +158,7 @@
|
|
157 |
"time_obs_encoder_hidden_dim": 128,
|
158 |
"time_obs_encoder_output_dim": 128,
|
159 |
"use_attention": true,
|
|
|
160 |
"zi_latent_dim": 200
|
161 |
},
|
162 |
"run_index": 0,
|
|
|
1 |
{
|
2 |
+
"best_val_loss": 0.045754145830869675,
|
3 |
"comet_ai_key": null,
|
4 |
"context_observations": {
|
5 |
"add_rem": true,
|
|
|
135 |
"activation": "ReLU",
|
136 |
"aggregator_num_heads": 8,
|
137 |
"aggregator_type": "attention",
|
138 |
+
"cov_proj_dim": 16,
|
139 |
"decoder_hidden_dim": 128,
|
140 |
"decoder_name": "TransformerDecoder",
|
141 |
"decoder_num_layers": 2,
|
|
|
158 |
"time_obs_encoder_hidden_dim": 128,
|
159 |
"time_obs_encoder_output_dim": 128,
|
160 |
"use_attention": true,
|
161 |
+
"use_covariance": false,
|
162 |
"zi_latent_dim": 200
|
163 |
},
|
164 |
"run_index": 0,
|
pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:73e03a290675cbe91fbbaebc6cfc35c9dfb2bc5d0b10fd75916fec6781a59546
|
3 |
+
size 4509179
|