End of training
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- README.md +63 -0
- config.json +26 -0
- model.safetensors +3 -0
- run-0/checkpoint-117/config.json +26 -0
- run-0/checkpoint-117/model.safetensors +3 -0
- run-0/checkpoint-117/optimizer.pt +3 -0
- run-0/checkpoint-117/rng_state.pth +3 -0
- run-0/checkpoint-117/scheduler.pt +3 -0
- run-0/checkpoint-117/special_tokens_map.json +7 -0
- run-0/checkpoint-117/tokenizer.json +0 -0
- run-0/checkpoint-117/tokenizer_config.json +56 -0
- run-0/checkpoint-117/trainer_state.json +67 -0
- run-0/checkpoint-117/training_args.bin +3 -0
- run-0/checkpoint-117/vocab.txt +0 -0
- run-0/checkpoint-156/config.json +26 -0
- run-0/checkpoint-156/model.safetensors +3 -0
- run-0/checkpoint-156/optimizer.pt +3 -0
- run-0/checkpoint-156/rng_state.pth +3 -0
- run-0/checkpoint-156/scheduler.pt +3 -0
- run-0/checkpoint-156/special_tokens_map.json +7 -0
- run-0/checkpoint-156/tokenizer.json +0 -0
- run-0/checkpoint-156/tokenizer_config.json +56 -0
- run-0/checkpoint-156/trainer_state.json +76 -0
- run-0/checkpoint-156/training_args.bin +3 -0
- run-0/checkpoint-156/vocab.txt +0 -0
- run-0/checkpoint-195/config.json +26 -0
- run-0/checkpoint-195/model.safetensors +3 -0
- run-0/checkpoint-195/optimizer.pt +3 -0
- run-0/checkpoint-195/rng_state.pth +3 -0
- run-0/checkpoint-195/scheduler.pt +3 -0
- run-0/checkpoint-195/special_tokens_map.json +7 -0
- run-0/checkpoint-195/tokenizer.json +0 -0
- run-0/checkpoint-195/tokenizer_config.json +56 -0
- run-0/checkpoint-195/trainer_state.json +85 -0
- run-0/checkpoint-195/training_args.bin +3 -0
- run-0/checkpoint-195/vocab.txt +0 -0
- run-0/checkpoint-234/config.json +26 -0
- run-0/checkpoint-234/model.safetensors +3 -0
- run-0/checkpoint-234/optimizer.pt +3 -0
- run-0/checkpoint-234/rng_state.pth +3 -0
- run-0/checkpoint-234/scheduler.pt +3 -0
- run-0/checkpoint-234/special_tokens_map.json +7 -0
- run-0/checkpoint-234/tokenizer.json +0 -0
- run-0/checkpoint-234/tokenizer_config.json +56 -0
- run-0/checkpoint-234/trainer_state.json +94 -0
- run-0/checkpoint-234/training_args.bin +3 -0
- run-0/checkpoint-234/vocab.txt +0 -0
- run-0/checkpoint-39/config.json +26 -0
- run-0/checkpoint-39/model.safetensors +3 -0
- run-0/checkpoint-39/optimizer.pt +3 -0
README.md
ADDED
@@ -0,0 +1,63 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
library_name: transformers
|
3 |
+
license: apache-2.0
|
4 |
+
base_model: bert-base-uncased
|
5 |
+
tags:
|
6 |
+
- generated_from_trainer
|
7 |
+
metrics:
|
8 |
+
- accuracy
|
9 |
+
model-index:
|
10 |
+
- name: bert-base-uncased-finetuned-rte-run_3
|
11 |
+
results: []
|
12 |
+
---
|
13 |
+
|
14 |
+
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
15 |
+
should probably proofread and complete it, then remove this comment. -->
|
16 |
+
|
17 |
+
# bert-base-uncased-finetuned-rte-run_3
|
18 |
+
|
19 |
+
This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on an unknown dataset.
|
20 |
+
It achieves the following results on the evaluation set:
|
21 |
+
- Loss: 0.6457
|
22 |
+
- Accuracy: 0.6606
|
23 |
+
|
24 |
+
## Model description
|
25 |
+
|
26 |
+
More information needed
|
27 |
+
|
28 |
+
## Intended uses & limitations
|
29 |
+
|
30 |
+
More information needed
|
31 |
+
|
32 |
+
## Training and evaluation data
|
33 |
+
|
34 |
+
More information needed
|
35 |
+
|
36 |
+
## Training procedure
|
37 |
+
|
38 |
+
### Training hyperparameters
|
39 |
+
|
40 |
+
The following hyperparameters were used during training:
|
41 |
+
- learning_rate: 3.50009902933741e-05
|
42 |
+
- train_batch_size: 64
|
43 |
+
- eval_batch_size: 64
|
44 |
+
- seed: 42
|
45 |
+
- optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
|
46 |
+
- lr_scheduler_type: linear
|
47 |
+
- num_epochs: 3
|
48 |
+
|
49 |
+
### Training results
|
50 |
+
|
51 |
+
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|
52 |
+
|:-------------:|:-----:|:----:|:---------------:|:--------:|
|
53 |
+
| No log | 1.0 | 39 | 0.6918 | 0.5271 |
|
54 |
+
| No log | 2.0 | 78 | 0.6407 | 0.6498 |
|
55 |
+
| No log | 3.0 | 117 | 0.6457 | 0.6606 |
|
56 |
+
|
57 |
+
|
58 |
+
### Framework versions
|
59 |
+
|
60 |
+
- Transformers 4.50.3
|
61 |
+
- Pytorch 2.6.0+cu124
|
62 |
+
- Datasets 3.5.0
|
63 |
+
- Tokenizers 0.21.1
|
config.json
ADDED
@@ -0,0 +1,26 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"architectures": [
|
3 |
+
"BertForSequenceClassification"
|
4 |
+
],
|
5 |
+
"attention_probs_dropout_prob": 0.1,
|
6 |
+
"classifier_dropout": null,
|
7 |
+
"gradient_checkpointing": false,
|
8 |
+
"hidden_act": "gelu",
|
9 |
+
"hidden_dropout_prob": 0.1,
|
10 |
+
"hidden_size": 768,
|
11 |
+
"initializer_range": 0.02,
|
12 |
+
"intermediate_size": 3072,
|
13 |
+
"layer_norm_eps": 1e-12,
|
14 |
+
"max_position_embeddings": 512,
|
15 |
+
"model_type": "bert",
|
16 |
+
"num_attention_heads": 12,
|
17 |
+
"num_hidden_layers": 12,
|
18 |
+
"pad_token_id": 0,
|
19 |
+
"position_embedding_type": "absolute",
|
20 |
+
"problem_type": "single_label_classification",
|
21 |
+
"torch_dtype": "float32",
|
22 |
+
"transformers_version": "4.50.3",
|
23 |
+
"type_vocab_size": 2,
|
24 |
+
"use_cache": true,
|
25 |
+
"vocab_size": 30522
|
26 |
+
}
|
model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e8e82ce999f505702503f5280b6342cbe44c87f8ea6e80ee4f3dd2fa93ac77c7
|
3 |
+
size 437958648
|
run-0/checkpoint-117/config.json
ADDED
@@ -0,0 +1,26 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"architectures": [
|
3 |
+
"BertForSequenceClassification"
|
4 |
+
],
|
5 |
+
"attention_probs_dropout_prob": 0.1,
|
6 |
+
"classifier_dropout": null,
|
7 |
+
"gradient_checkpointing": false,
|
8 |
+
"hidden_act": "gelu",
|
9 |
+
"hidden_dropout_prob": 0.1,
|
10 |
+
"hidden_size": 768,
|
11 |
+
"initializer_range": 0.02,
|
12 |
+
"intermediate_size": 3072,
|
13 |
+
"layer_norm_eps": 1e-12,
|
14 |
+
"max_position_embeddings": 512,
|
15 |
+
"model_type": "bert",
|
16 |
+
"num_attention_heads": 12,
|
17 |
+
"num_hidden_layers": 12,
|
18 |
+
"pad_token_id": 0,
|
19 |
+
"position_embedding_type": "absolute",
|
20 |
+
"problem_type": "single_label_classification",
|
21 |
+
"torch_dtype": "float32",
|
22 |
+
"transformers_version": "4.50.3",
|
23 |
+
"type_vocab_size": 2,
|
24 |
+
"use_cache": true,
|
25 |
+
"vocab_size": 30522
|
26 |
+
}
|
run-0/checkpoint-117/model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:58ac66792066963804151e6788af71377ac80af960ede318603263e8a71fc0f1
|
3 |
+
size 437958648
|
run-0/checkpoint-117/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d63c1630ac509249f218e3274dfdbcccfb5409a8f16cb67a3024a44c7f8d8507
|
3 |
+
size 876038394
|
run-0/checkpoint-117/rng_state.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:66062f09fb79eb516b32f1e74227740c5dc9e0025a015b9853db13b00b33e723
|
3 |
+
size 14244
|
run-0/checkpoint-117/scheduler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2062d22e00ae4657d5a2437cf8cf02f80d3f9a78be8bd9fd5b93e3e8134b5125
|
3 |
+
size 1064
|
run-0/checkpoint-117/special_tokens_map.json
ADDED
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"cls_token": "[CLS]",
|
3 |
+
"mask_token": "[MASK]",
|
4 |
+
"pad_token": "[PAD]",
|
5 |
+
"sep_token": "[SEP]",
|
6 |
+
"unk_token": "[UNK]"
|
7 |
+
}
|
run-0/checkpoint-117/tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
run-0/checkpoint-117/tokenizer_config.json
ADDED
@@ -0,0 +1,56 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"added_tokens_decoder": {
|
3 |
+
"0": {
|
4 |
+
"content": "[PAD]",
|
5 |
+
"lstrip": false,
|
6 |
+
"normalized": false,
|
7 |
+
"rstrip": false,
|
8 |
+
"single_word": false,
|
9 |
+
"special": true
|
10 |
+
},
|
11 |
+
"100": {
|
12 |
+
"content": "[UNK]",
|
13 |
+
"lstrip": false,
|
14 |
+
"normalized": false,
|
15 |
+
"rstrip": false,
|
16 |
+
"single_word": false,
|
17 |
+
"special": true
|
18 |
+
},
|
19 |
+
"101": {
|
20 |
+
"content": "[CLS]",
|
21 |
+
"lstrip": false,
|
22 |
+
"normalized": false,
|
23 |
+
"rstrip": false,
|
24 |
+
"single_word": false,
|
25 |
+
"special": true
|
26 |
+
},
|
27 |
+
"102": {
|
28 |
+
"content": "[SEP]",
|
29 |
+
"lstrip": false,
|
30 |
+
"normalized": false,
|
31 |
+
"rstrip": false,
|
32 |
+
"single_word": false,
|
33 |
+
"special": true
|
34 |
+
},
|
35 |
+
"103": {
|
36 |
+
"content": "[MASK]",
|
37 |
+
"lstrip": false,
|
38 |
+
"normalized": false,
|
39 |
+
"rstrip": false,
|
40 |
+
"single_word": false,
|
41 |
+
"special": true
|
42 |
+
}
|
43 |
+
},
|
44 |
+
"clean_up_tokenization_spaces": false,
|
45 |
+
"cls_token": "[CLS]",
|
46 |
+
"do_lower_case": true,
|
47 |
+
"extra_special_tokens": {},
|
48 |
+
"mask_token": "[MASK]",
|
49 |
+
"model_max_length": 512,
|
50 |
+
"pad_token": "[PAD]",
|
51 |
+
"sep_token": "[SEP]",
|
52 |
+
"strip_accents": null,
|
53 |
+
"tokenize_chinese_chars": true,
|
54 |
+
"tokenizer_class": "BertTokenizer",
|
55 |
+
"unk_token": "[UNK]"
|
56 |
+
}
|
run-0/checkpoint-117/trainer_state.json
ADDED
@@ -0,0 +1,67 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"best_global_step": 117,
|
3 |
+
"best_metric": 0.6642599277978339,
|
4 |
+
"best_model_checkpoint": "bert-base-uncased-finetuned-rte-run_3/run-0/checkpoint-117",
|
5 |
+
"epoch": 3.0,
|
6 |
+
"eval_steps": 500,
|
7 |
+
"global_step": 117,
|
8 |
+
"is_hyper_param_search": true,
|
9 |
+
"is_local_process_zero": true,
|
10 |
+
"is_world_process_zero": true,
|
11 |
+
"log_history": [
|
12 |
+
{
|
13 |
+
"epoch": 1.0,
|
14 |
+
"eval_accuracy": 0.5270758122743683,
|
15 |
+
"eval_loss": 0.6917867064476013,
|
16 |
+
"eval_runtime": 3.2048,
|
17 |
+
"eval_samples_per_second": 86.433,
|
18 |
+
"eval_steps_per_second": 1.56,
|
19 |
+
"step": 39
|
20 |
+
},
|
21 |
+
{
|
22 |
+
"epoch": 2.0,
|
23 |
+
"eval_accuracy": 0.6498194945848376,
|
24 |
+
"eval_loss": 0.6407069563865662,
|
25 |
+
"eval_runtime": 3.1725,
|
26 |
+
"eval_samples_per_second": 87.314,
|
27 |
+
"eval_steps_per_second": 1.576,
|
28 |
+
"step": 78
|
29 |
+
},
|
30 |
+
{
|
31 |
+
"epoch": 3.0,
|
32 |
+
"eval_accuracy": 0.6642599277978339,
|
33 |
+
"eval_loss": 0.6459211707115173,
|
34 |
+
"eval_runtime": 3.1766,
|
35 |
+
"eval_samples_per_second": 87.199,
|
36 |
+
"eval_steps_per_second": 1.574,
|
37 |
+
"step": 117
|
38 |
+
}
|
39 |
+
],
|
40 |
+
"logging_steps": 500,
|
41 |
+
"max_steps": 117,
|
42 |
+
"num_input_tokens_seen": 0,
|
43 |
+
"num_train_epochs": 3,
|
44 |
+
"save_steps": 500,
|
45 |
+
"stateful_callbacks": {
|
46 |
+
"TrainerControl": {
|
47 |
+
"args": {
|
48 |
+
"should_epoch_stop": false,
|
49 |
+
"should_evaluate": false,
|
50 |
+
"should_log": false,
|
51 |
+
"should_save": true,
|
52 |
+
"should_training_stop": true
|
53 |
+
},
|
54 |
+
"attributes": {}
|
55 |
+
}
|
56 |
+
},
|
57 |
+
"total_flos": 0,
|
58 |
+
"train_batch_size": 64,
|
59 |
+
"trial_name": null,
|
60 |
+
"trial_params": {
|
61 |
+
"dropout_rate": 0.17661553519517478,
|
62 |
+
"learning_rate": 3.50009902933741e-05,
|
63 |
+
"max_length": 512,
|
64 |
+
"num_train_epochs": 3,
|
65 |
+
"per_device_train_batch_size": 64
|
66 |
+
}
|
67 |
+
}
|
run-0/checkpoint-117/training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2dc21828eae1c275563be6e7b84b6d4c4c581c53e194ae689da7d5222b7497bf
|
3 |
+
size 5432
|
run-0/checkpoint-117/vocab.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
run-0/checkpoint-156/config.json
ADDED
@@ -0,0 +1,26 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"architectures": [
|
3 |
+
"BertForSequenceClassification"
|
4 |
+
],
|
5 |
+
"attention_probs_dropout_prob": 0.1,
|
6 |
+
"classifier_dropout": null,
|
7 |
+
"gradient_checkpointing": false,
|
8 |
+
"hidden_act": "gelu",
|
9 |
+
"hidden_dropout_prob": 0.1,
|
10 |
+
"hidden_size": 768,
|
11 |
+
"initializer_range": 0.02,
|
12 |
+
"intermediate_size": 3072,
|
13 |
+
"layer_norm_eps": 1e-12,
|
14 |
+
"max_position_embeddings": 512,
|
15 |
+
"model_type": "bert",
|
16 |
+
"num_attention_heads": 12,
|
17 |
+
"num_hidden_layers": 12,
|
18 |
+
"pad_token_id": 0,
|
19 |
+
"position_embedding_type": "absolute",
|
20 |
+
"problem_type": "single_label_classification",
|
21 |
+
"torch_dtype": "float32",
|
22 |
+
"transformers_version": "4.50.3",
|
23 |
+
"type_vocab_size": 2,
|
24 |
+
"use_cache": true,
|
25 |
+
"vocab_size": 30522
|
26 |
+
}
|
run-0/checkpoint-156/model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2ac1eceb00174fe493c079fff1fbd36c37b46fa6b75c678175338eb3832ae867
|
3 |
+
size 437958648
|
run-0/checkpoint-156/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:21a8109b61528f50f43568d61a2809776712bfbc241f5362c07c1f5c4c2bf079
|
3 |
+
size 876038394
|
run-0/checkpoint-156/rng_state.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6f2bd60aa42ae70430738a58c3189fe231933c3b00098640c8eda2327725ddc7
|
3 |
+
size 14244
|
run-0/checkpoint-156/scheduler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6d7bbd17949f959de5a99d2de51c21954f2eaf6f4687a7cdbef223739c23f985
|
3 |
+
size 1064
|
run-0/checkpoint-156/special_tokens_map.json
ADDED
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"cls_token": "[CLS]",
|
3 |
+
"mask_token": "[MASK]",
|
4 |
+
"pad_token": "[PAD]",
|
5 |
+
"sep_token": "[SEP]",
|
6 |
+
"unk_token": "[UNK]"
|
7 |
+
}
|
run-0/checkpoint-156/tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
run-0/checkpoint-156/tokenizer_config.json
ADDED
@@ -0,0 +1,56 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"added_tokens_decoder": {
|
3 |
+
"0": {
|
4 |
+
"content": "[PAD]",
|
5 |
+
"lstrip": false,
|
6 |
+
"normalized": false,
|
7 |
+
"rstrip": false,
|
8 |
+
"single_word": false,
|
9 |
+
"special": true
|
10 |
+
},
|
11 |
+
"100": {
|
12 |
+
"content": "[UNK]",
|
13 |
+
"lstrip": false,
|
14 |
+
"normalized": false,
|
15 |
+
"rstrip": false,
|
16 |
+
"single_word": false,
|
17 |
+
"special": true
|
18 |
+
},
|
19 |
+
"101": {
|
20 |
+
"content": "[CLS]",
|
21 |
+
"lstrip": false,
|
22 |
+
"normalized": false,
|
23 |
+
"rstrip": false,
|
24 |
+
"single_word": false,
|
25 |
+
"special": true
|
26 |
+
},
|
27 |
+
"102": {
|
28 |
+
"content": "[SEP]",
|
29 |
+
"lstrip": false,
|
30 |
+
"normalized": false,
|
31 |
+
"rstrip": false,
|
32 |
+
"single_word": false,
|
33 |
+
"special": true
|
34 |
+
},
|
35 |
+
"103": {
|
36 |
+
"content": "[MASK]",
|
37 |
+
"lstrip": false,
|
38 |
+
"normalized": false,
|
39 |
+
"rstrip": false,
|
40 |
+
"single_word": false,
|
41 |
+
"special": true
|
42 |
+
}
|
43 |
+
},
|
44 |
+
"clean_up_tokenization_spaces": false,
|
45 |
+
"cls_token": "[CLS]",
|
46 |
+
"do_lower_case": true,
|
47 |
+
"extra_special_tokens": {},
|
48 |
+
"mask_token": "[MASK]",
|
49 |
+
"model_max_length": 512,
|
50 |
+
"pad_token": "[PAD]",
|
51 |
+
"sep_token": "[SEP]",
|
52 |
+
"strip_accents": null,
|
53 |
+
"tokenize_chinese_chars": true,
|
54 |
+
"tokenizer_class": "BertTokenizer",
|
55 |
+
"unk_token": "[UNK]"
|
56 |
+
}
|
run-0/checkpoint-156/trainer_state.json
ADDED
@@ -0,0 +1,76 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"best_global_step": 117,
|
3 |
+
"best_metric": 0.6823104693140795,
|
4 |
+
"best_model_checkpoint": "bert-base-uncased-finetuned-rte-run_3/run-0/checkpoint-117",
|
5 |
+
"epoch": 4.0,
|
6 |
+
"eval_steps": 500,
|
7 |
+
"global_step": 156,
|
8 |
+
"is_hyper_param_search": true,
|
9 |
+
"is_local_process_zero": true,
|
10 |
+
"is_world_process_zero": true,
|
11 |
+
"log_history": [
|
12 |
+
{
|
13 |
+
"epoch": 1.0,
|
14 |
+
"eval_accuracy": 0.6064981949458483,
|
15 |
+
"eval_loss": 0.6656323671340942,
|
16 |
+
"eval_runtime": 3.229,
|
17 |
+
"eval_samples_per_second": 85.785,
|
18 |
+
"eval_steps_per_second": 1.548,
|
19 |
+
"step": 39
|
20 |
+
},
|
21 |
+
{
|
22 |
+
"epoch": 2.0,
|
23 |
+
"eval_accuracy": 0.6498194945848376,
|
24 |
+
"eval_loss": 0.6296112537384033,
|
25 |
+
"eval_runtime": 3.2312,
|
26 |
+
"eval_samples_per_second": 85.728,
|
27 |
+
"eval_steps_per_second": 1.547,
|
28 |
+
"step": 78
|
29 |
+
},
|
30 |
+
{
|
31 |
+
"epoch": 3.0,
|
32 |
+
"eval_accuracy": 0.6823104693140795,
|
33 |
+
"eval_loss": 0.6674417853355408,
|
34 |
+
"eval_runtime": 3.1663,
|
35 |
+
"eval_samples_per_second": 87.483,
|
36 |
+
"eval_steps_per_second": 1.579,
|
37 |
+
"step": 117
|
38 |
+
},
|
39 |
+
{
|
40 |
+
"epoch": 4.0,
|
41 |
+
"eval_accuracy": 0.6678700361010831,
|
42 |
+
"eval_loss": 0.7529056072235107,
|
43 |
+
"eval_runtime": 3.1735,
|
44 |
+
"eval_samples_per_second": 87.285,
|
45 |
+
"eval_steps_per_second": 1.576,
|
46 |
+
"step": 156
|
47 |
+
}
|
48 |
+
],
|
49 |
+
"logging_steps": 500,
|
50 |
+
"max_steps": 156,
|
51 |
+
"num_input_tokens_seen": 0,
|
52 |
+
"num_train_epochs": 4,
|
53 |
+
"save_steps": 500,
|
54 |
+
"stateful_callbacks": {
|
55 |
+
"TrainerControl": {
|
56 |
+
"args": {
|
57 |
+
"should_epoch_stop": false,
|
58 |
+
"should_evaluate": false,
|
59 |
+
"should_log": false,
|
60 |
+
"should_save": true,
|
61 |
+
"should_training_stop": true
|
62 |
+
},
|
63 |
+
"attributes": {}
|
64 |
+
}
|
65 |
+
},
|
66 |
+
"total_flos": 0,
|
67 |
+
"train_batch_size": 64,
|
68 |
+
"trial_name": null,
|
69 |
+
"trial_params": {
|
70 |
+
"dropout_rate": 0.32872199298996635,
|
71 |
+
"learning_rate": 4.9534002785729435e-05,
|
72 |
+
"max_length": 256,
|
73 |
+
"num_train_epochs": 4,
|
74 |
+
"per_device_train_batch_size": 64
|
75 |
+
}
|
76 |
+
}
|
run-0/checkpoint-156/training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e79af4227a137db021853d20370fdd1d2ea42b3a338e96a6c3a7f2ad6c7e573c
|
3 |
+
size 5432
|
run-0/checkpoint-156/vocab.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
run-0/checkpoint-195/config.json
ADDED
@@ -0,0 +1,26 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"architectures": [
|
3 |
+
"BertForSequenceClassification"
|
4 |
+
],
|
5 |
+
"attention_probs_dropout_prob": 0.1,
|
6 |
+
"classifier_dropout": null,
|
7 |
+
"gradient_checkpointing": false,
|
8 |
+
"hidden_act": "gelu",
|
9 |
+
"hidden_dropout_prob": 0.1,
|
10 |
+
"hidden_size": 768,
|
11 |
+
"initializer_range": 0.02,
|
12 |
+
"intermediate_size": 3072,
|
13 |
+
"layer_norm_eps": 1e-12,
|
14 |
+
"max_position_embeddings": 512,
|
15 |
+
"model_type": "bert",
|
16 |
+
"num_attention_heads": 12,
|
17 |
+
"num_hidden_layers": 12,
|
18 |
+
"pad_token_id": 0,
|
19 |
+
"position_embedding_type": "absolute",
|
20 |
+
"problem_type": "single_label_classification",
|
21 |
+
"torch_dtype": "float32",
|
22 |
+
"transformers_version": "4.50.3",
|
23 |
+
"type_vocab_size": 2,
|
24 |
+
"use_cache": true,
|
25 |
+
"vocab_size": 30522
|
26 |
+
}
|
run-0/checkpoint-195/model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:95b8d01ea1fa03fecb27d3824ef48321dd292022de2980ec8cd85ee43a3a31cb
|
3 |
+
size 437958648
|
run-0/checkpoint-195/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:32bb59c695dd3a9f11f5c62fe22bb5282124d32d9410ffc137cda9e60aeffb60
|
3 |
+
size 876038394
|
run-0/checkpoint-195/rng_state.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:78e2e34b39683502d4d6915c71fb35228270e1cba7227213b6a9df68b951484b
|
3 |
+
size 14244
|
run-0/checkpoint-195/scheduler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:278bb2692df60b8433969a78547b13bdd26f2d247316b5de7f891cadbc378ef4
|
3 |
+
size 1064
|
run-0/checkpoint-195/special_tokens_map.json
ADDED
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"cls_token": "[CLS]",
|
3 |
+
"mask_token": "[MASK]",
|
4 |
+
"pad_token": "[PAD]",
|
5 |
+
"sep_token": "[SEP]",
|
6 |
+
"unk_token": "[UNK]"
|
7 |
+
}
|
run-0/checkpoint-195/tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
run-0/checkpoint-195/tokenizer_config.json
ADDED
@@ -0,0 +1,56 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"added_tokens_decoder": {
|
3 |
+
"0": {
|
4 |
+
"content": "[PAD]",
|
5 |
+
"lstrip": false,
|
6 |
+
"normalized": false,
|
7 |
+
"rstrip": false,
|
8 |
+
"single_word": false,
|
9 |
+
"special": true
|
10 |
+
},
|
11 |
+
"100": {
|
12 |
+
"content": "[UNK]",
|
13 |
+
"lstrip": false,
|
14 |
+
"normalized": false,
|
15 |
+
"rstrip": false,
|
16 |
+
"single_word": false,
|
17 |
+
"special": true
|
18 |
+
},
|
19 |
+
"101": {
|
20 |
+
"content": "[CLS]",
|
21 |
+
"lstrip": false,
|
22 |
+
"normalized": false,
|
23 |
+
"rstrip": false,
|
24 |
+
"single_word": false,
|
25 |
+
"special": true
|
26 |
+
},
|
27 |
+
"102": {
|
28 |
+
"content": "[SEP]",
|
29 |
+
"lstrip": false,
|
30 |
+
"normalized": false,
|
31 |
+
"rstrip": false,
|
32 |
+
"single_word": false,
|
33 |
+
"special": true
|
34 |
+
},
|
35 |
+
"103": {
|
36 |
+
"content": "[MASK]",
|
37 |
+
"lstrip": false,
|
38 |
+
"normalized": false,
|
39 |
+
"rstrip": false,
|
40 |
+
"single_word": false,
|
41 |
+
"special": true
|
42 |
+
}
|
43 |
+
},
|
44 |
+
"clean_up_tokenization_spaces": false,
|
45 |
+
"cls_token": "[CLS]",
|
46 |
+
"do_lower_case": true,
|
47 |
+
"extra_special_tokens": {},
|
48 |
+
"mask_token": "[MASK]",
|
49 |
+
"model_max_length": 512,
|
50 |
+
"pad_token": "[PAD]",
|
51 |
+
"sep_token": "[SEP]",
|
52 |
+
"strip_accents": null,
|
53 |
+
"tokenize_chinese_chars": true,
|
54 |
+
"tokenizer_class": "BertTokenizer",
|
55 |
+
"unk_token": "[UNK]"
|
56 |
+
}
|
run-0/checkpoint-195/trainer_state.json
ADDED
@@ -0,0 +1,85 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"best_global_step": 156,
|
3 |
+
"best_metric": 0.6931407942238267,
|
4 |
+
"best_model_checkpoint": "bert-base-uncased-finetuned-rte-run_3/run-0/checkpoint-156",
|
5 |
+
"epoch": 5.0,
|
6 |
+
"eval_steps": 500,
|
7 |
+
"global_step": 195,
|
8 |
+
"is_hyper_param_search": true,
|
9 |
+
"is_local_process_zero": true,
|
10 |
+
"is_world_process_zero": true,
|
11 |
+
"log_history": [
|
12 |
+
{
|
13 |
+
"epoch": 1.0,
|
14 |
+
"eval_accuracy": 0.6173285198555957,
|
15 |
+
"eval_loss": 0.6601823568344116,
|
16 |
+
"eval_runtime": 3.2169,
|
17 |
+
"eval_samples_per_second": 86.108,
|
18 |
+
"eval_steps_per_second": 1.554,
|
19 |
+
"step": 39
|
20 |
+
},
|
21 |
+
{
|
22 |
+
"epoch": 2.0,
|
23 |
+
"eval_accuracy": 0.6498194945848376,
|
24 |
+
"eval_loss": 0.6585440635681152,
|
25 |
+
"eval_runtime": 3.167,
|
26 |
+
"eval_samples_per_second": 87.464,
|
27 |
+
"eval_steps_per_second": 1.579,
|
28 |
+
"step": 78
|
29 |
+
},
|
30 |
+
{
|
31 |
+
"epoch": 3.0,
|
32 |
+
"eval_accuracy": 0.6750902527075813,
|
33 |
+
"eval_loss": 0.6770333051681519,
|
34 |
+
"eval_runtime": 3.1901,
|
35 |
+
"eval_samples_per_second": 86.831,
|
36 |
+
"eval_steps_per_second": 1.567,
|
37 |
+
"step": 117
|
38 |
+
},
|
39 |
+
{
|
40 |
+
"epoch": 4.0,
|
41 |
+
"eval_accuracy": 0.6931407942238267,
|
42 |
+
"eval_loss": 0.811002790927887,
|
43 |
+
"eval_runtime": 3.1876,
|
44 |
+
"eval_samples_per_second": 86.9,
|
45 |
+
"eval_steps_per_second": 1.569,
|
46 |
+
"step": 156
|
47 |
+
},
|
48 |
+
{
|
49 |
+
"epoch": 5.0,
|
50 |
+
"eval_accuracy": 0.6823104693140795,
|
51 |
+
"eval_loss": 0.8974205851554871,
|
52 |
+
"eval_runtime": 3.1823,
|
53 |
+
"eval_samples_per_second": 87.045,
|
54 |
+
"eval_steps_per_second": 1.571,
|
55 |
+
"step": 195
|
56 |
+
}
|
57 |
+
],
|
58 |
+
"logging_steps": 500,
|
59 |
+
"max_steps": 234,
|
60 |
+
"num_input_tokens_seen": 0,
|
61 |
+
"num_train_epochs": 6,
|
62 |
+
"save_steps": 500,
|
63 |
+
"stateful_callbacks": {
|
64 |
+
"TrainerControl": {
|
65 |
+
"args": {
|
66 |
+
"should_epoch_stop": false,
|
67 |
+
"should_evaluate": false,
|
68 |
+
"should_log": false,
|
69 |
+
"should_save": true,
|
70 |
+
"should_training_stop": false
|
71 |
+
},
|
72 |
+
"attributes": {}
|
73 |
+
}
|
74 |
+
},
|
75 |
+
"total_flos": 0,
|
76 |
+
"train_batch_size": 64,
|
77 |
+
"trial_name": null,
|
78 |
+
"trial_params": {
|
79 |
+
"dropout_rate": 0.18213454295456433,
|
80 |
+
"learning_rate": 3.6493270590502144e-05,
|
81 |
+
"max_length": 128,
|
82 |
+
"num_train_epochs": 6,
|
83 |
+
"per_device_train_batch_size": 64
|
84 |
+
}
|
85 |
+
}
|
run-0/checkpoint-195/training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:229c1e8ce32265f4785c2f1fc6c5974e8e794c5f13183f0befbbf5834964686e
|
3 |
+
size 5432
|
run-0/checkpoint-195/vocab.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
run-0/checkpoint-234/config.json
ADDED
@@ -0,0 +1,26 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"architectures": [
|
3 |
+
"BertForSequenceClassification"
|
4 |
+
],
|
5 |
+
"attention_probs_dropout_prob": 0.1,
|
6 |
+
"classifier_dropout": null,
|
7 |
+
"gradient_checkpointing": false,
|
8 |
+
"hidden_act": "gelu",
|
9 |
+
"hidden_dropout_prob": 0.1,
|
10 |
+
"hidden_size": 768,
|
11 |
+
"initializer_range": 0.02,
|
12 |
+
"intermediate_size": 3072,
|
13 |
+
"layer_norm_eps": 1e-12,
|
14 |
+
"max_position_embeddings": 512,
|
15 |
+
"model_type": "bert",
|
16 |
+
"num_attention_heads": 12,
|
17 |
+
"num_hidden_layers": 12,
|
18 |
+
"pad_token_id": 0,
|
19 |
+
"position_embedding_type": "absolute",
|
20 |
+
"problem_type": "single_label_classification",
|
21 |
+
"torch_dtype": "float32",
|
22 |
+
"transformers_version": "4.50.3",
|
23 |
+
"type_vocab_size": 2,
|
24 |
+
"use_cache": true,
|
25 |
+
"vocab_size": 30522
|
26 |
+
}
|
run-0/checkpoint-234/model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a3e1f2592db36d89fc4a1436eeb99d8f64588b68868d3187795c220c6fd54ae0
|
3 |
+
size 437958648
|
run-0/checkpoint-234/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ded49ae56a161dea9cc1794f19d43250327ec9b8e5af194d2f03cb961feaf422
|
3 |
+
size 876038394
|
run-0/checkpoint-234/rng_state.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2c335e518260c2aab1d76d67154f45cb6ec7612dd36ed4defaf4739071e2e0de
|
3 |
+
size 14244
|
run-0/checkpoint-234/scheduler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:795acc3a991d2adea7b7d2df172627ff581631ec207324bc6948cc49fceaa143
|
3 |
+
size 1064
|
run-0/checkpoint-234/special_tokens_map.json
ADDED
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"cls_token": "[CLS]",
|
3 |
+
"mask_token": "[MASK]",
|
4 |
+
"pad_token": "[PAD]",
|
5 |
+
"sep_token": "[SEP]",
|
6 |
+
"unk_token": "[UNK]"
|
7 |
+
}
|
run-0/checkpoint-234/tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
run-0/checkpoint-234/tokenizer_config.json
ADDED
@@ -0,0 +1,56 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"added_tokens_decoder": {
|
3 |
+
"0": {
|
4 |
+
"content": "[PAD]",
|
5 |
+
"lstrip": false,
|
6 |
+
"normalized": false,
|
7 |
+
"rstrip": false,
|
8 |
+
"single_word": false,
|
9 |
+
"special": true
|
10 |
+
},
|
11 |
+
"100": {
|
12 |
+
"content": "[UNK]",
|
13 |
+
"lstrip": false,
|
14 |
+
"normalized": false,
|
15 |
+
"rstrip": false,
|
16 |
+
"single_word": false,
|
17 |
+
"special": true
|
18 |
+
},
|
19 |
+
"101": {
|
20 |
+
"content": "[CLS]",
|
21 |
+
"lstrip": false,
|
22 |
+
"normalized": false,
|
23 |
+
"rstrip": false,
|
24 |
+
"single_word": false,
|
25 |
+
"special": true
|
26 |
+
},
|
27 |
+
"102": {
|
28 |
+
"content": "[SEP]",
|
29 |
+
"lstrip": false,
|
30 |
+
"normalized": false,
|
31 |
+
"rstrip": false,
|
32 |
+
"single_word": false,
|
33 |
+
"special": true
|
34 |
+
},
|
35 |
+
"103": {
|
36 |
+
"content": "[MASK]",
|
37 |
+
"lstrip": false,
|
38 |
+
"normalized": false,
|
39 |
+
"rstrip": false,
|
40 |
+
"single_word": false,
|
41 |
+
"special": true
|
42 |
+
}
|
43 |
+
},
|
44 |
+
"clean_up_tokenization_spaces": false,
|
45 |
+
"cls_token": "[CLS]",
|
46 |
+
"do_lower_case": true,
|
47 |
+
"extra_special_tokens": {},
|
48 |
+
"mask_token": "[MASK]",
|
49 |
+
"model_max_length": 512,
|
50 |
+
"pad_token": "[PAD]",
|
51 |
+
"sep_token": "[SEP]",
|
52 |
+
"strip_accents": null,
|
53 |
+
"tokenize_chinese_chars": true,
|
54 |
+
"tokenizer_class": "BertTokenizer",
|
55 |
+
"unk_token": "[UNK]"
|
56 |
+
}
|
run-0/checkpoint-234/trainer_state.json
ADDED
@@ -0,0 +1,94 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"best_global_step": 156,
|
3 |
+
"best_metric": 0.6931407942238267,
|
4 |
+
"best_model_checkpoint": "bert-base-uncased-finetuned-rte-run_3/run-0/checkpoint-156",
|
5 |
+
"epoch": 6.0,
|
6 |
+
"eval_steps": 500,
|
7 |
+
"global_step": 234,
|
8 |
+
"is_hyper_param_search": true,
|
9 |
+
"is_local_process_zero": true,
|
10 |
+
"is_world_process_zero": true,
|
11 |
+
"log_history": [
|
12 |
+
{
|
13 |
+
"epoch": 1.0,
|
14 |
+
"eval_accuracy": 0.6173285198555957,
|
15 |
+
"eval_loss": 0.6601823568344116,
|
16 |
+
"eval_runtime": 3.2169,
|
17 |
+
"eval_samples_per_second": 86.108,
|
18 |
+
"eval_steps_per_second": 1.554,
|
19 |
+
"step": 39
|
20 |
+
},
|
21 |
+
{
|
22 |
+
"epoch": 2.0,
|
23 |
+
"eval_accuracy": 0.6498194945848376,
|
24 |
+
"eval_loss": 0.6585440635681152,
|
25 |
+
"eval_runtime": 3.167,
|
26 |
+
"eval_samples_per_second": 87.464,
|
27 |
+
"eval_steps_per_second": 1.579,
|
28 |
+
"step": 78
|
29 |
+
},
|
30 |
+
{
|
31 |
+
"epoch": 3.0,
|
32 |
+
"eval_accuracy": 0.6750902527075813,
|
33 |
+
"eval_loss": 0.6770333051681519,
|
34 |
+
"eval_runtime": 3.1901,
|
35 |
+
"eval_samples_per_second": 86.831,
|
36 |
+
"eval_steps_per_second": 1.567,
|
37 |
+
"step": 117
|
38 |
+
},
|
39 |
+
{
|
40 |
+
"epoch": 4.0,
|
41 |
+
"eval_accuracy": 0.6931407942238267,
|
42 |
+
"eval_loss": 0.811002790927887,
|
43 |
+
"eval_runtime": 3.1876,
|
44 |
+
"eval_samples_per_second": 86.9,
|
45 |
+
"eval_steps_per_second": 1.569,
|
46 |
+
"step": 156
|
47 |
+
},
|
48 |
+
{
|
49 |
+
"epoch": 5.0,
|
50 |
+
"eval_accuracy": 0.6823104693140795,
|
51 |
+
"eval_loss": 0.8974205851554871,
|
52 |
+
"eval_runtime": 3.1823,
|
53 |
+
"eval_samples_per_second": 87.045,
|
54 |
+
"eval_steps_per_second": 1.571,
|
55 |
+
"step": 195
|
56 |
+
},
|
57 |
+
{
|
58 |
+
"epoch": 6.0,
|
59 |
+
"eval_accuracy": 0.6931407942238267,
|
60 |
+
"eval_loss": 0.9125683307647705,
|
61 |
+
"eval_runtime": 3.182,
|
62 |
+
"eval_samples_per_second": 87.053,
|
63 |
+
"eval_steps_per_second": 1.571,
|
64 |
+
"step": 234
|
65 |
+
}
|
66 |
+
],
|
67 |
+
"logging_steps": 500,
|
68 |
+
"max_steps": 234,
|
69 |
+
"num_input_tokens_seen": 0,
|
70 |
+
"num_train_epochs": 6,
|
71 |
+
"save_steps": 500,
|
72 |
+
"stateful_callbacks": {
|
73 |
+
"TrainerControl": {
|
74 |
+
"args": {
|
75 |
+
"should_epoch_stop": false,
|
76 |
+
"should_evaluate": false,
|
77 |
+
"should_log": false,
|
78 |
+
"should_save": true,
|
79 |
+
"should_training_stop": true
|
80 |
+
},
|
81 |
+
"attributes": {}
|
82 |
+
}
|
83 |
+
},
|
84 |
+
"total_flos": 0,
|
85 |
+
"train_batch_size": 64,
|
86 |
+
"trial_name": null,
|
87 |
+
"trial_params": {
|
88 |
+
"dropout_rate": 0.18213454295456433,
|
89 |
+
"learning_rate": 3.6493270590502144e-05,
|
90 |
+
"max_length": 128,
|
91 |
+
"num_train_epochs": 6,
|
92 |
+
"per_device_train_batch_size": 64
|
93 |
+
}
|
94 |
+
}
|
run-0/checkpoint-234/training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:229c1e8ce32265f4785c2f1fc6c5974e8e794c5f13183f0befbbf5834964686e
|
3 |
+
size 5432
|
run-0/checkpoint-234/vocab.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
run-0/checkpoint-39/config.json
ADDED
@@ -0,0 +1,26 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"architectures": [
|
3 |
+
"BertForSequenceClassification"
|
4 |
+
],
|
5 |
+
"attention_probs_dropout_prob": 0.1,
|
6 |
+
"classifier_dropout": null,
|
7 |
+
"gradient_checkpointing": false,
|
8 |
+
"hidden_act": "gelu",
|
9 |
+
"hidden_dropout_prob": 0.1,
|
10 |
+
"hidden_size": 768,
|
11 |
+
"initializer_range": 0.02,
|
12 |
+
"intermediate_size": 3072,
|
13 |
+
"layer_norm_eps": 1e-12,
|
14 |
+
"max_position_embeddings": 512,
|
15 |
+
"model_type": "bert",
|
16 |
+
"num_attention_heads": 12,
|
17 |
+
"num_hidden_layers": 12,
|
18 |
+
"pad_token_id": 0,
|
19 |
+
"position_embedding_type": "absolute",
|
20 |
+
"problem_type": "single_label_classification",
|
21 |
+
"torch_dtype": "float32",
|
22 |
+
"transformers_version": "4.50.3",
|
23 |
+
"type_vocab_size": 2,
|
24 |
+
"use_cache": true,
|
25 |
+
"vocab_size": 30522
|
26 |
+
}
|
run-0/checkpoint-39/model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f5666a40d81e971b7e328323dbaf5b586dc7d08eb3b80e5d286b3aa1f944fa1b
|
3 |
+
size 437958648
|
run-0/checkpoint-39/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:704af016b770a116f5c042e898063e2e8eec492a362fd0c63bfc5270a583ac44
|
3 |
+
size 876038394
|