DunnBC22 commited on
Commit
1a669c1
·
1 Parent(s): baaa797

All Dunn!!!

Browse files
last-checkpoint/config.json DELETED
@@ -1,35 +0,0 @@
1
- {
2
- "_name_or_path": "distilbert-base-uncased",
3
- "activation": "gelu",
4
- "architectures": [
5
- "DistilBertForSequenceClassification"
6
- ],
7
- "attention_dropout": 0.1,
8
- "dim": 768,
9
- "dropout": 0.1,
10
- "hidden_dim": 3072,
11
- "id2label": {
12
- "0": "LABEL_0",
13
- "1": "LABEL_1",
14
- "2": "LABEL_2"
15
- },
16
- "initializer_range": 0.02,
17
- "label2id": {
18
- "LABEL_0": 0,
19
- "LABEL_1": 1,
20
- "LABEL_2": 2
21
- },
22
- "max_position_embeddings": 512,
23
- "model_type": "distilbert",
24
- "n_heads": 12,
25
- "n_layers": 6,
26
- "pad_token_id": 0,
27
- "problem_type": "single_label_classification",
28
- "qa_dropout": 0.1,
29
- "seq_classif_dropout": 0.2,
30
- "sinusoidal_pos_embds": false,
31
- "tie_weights_": true,
32
- "torch_dtype": "float32",
33
- "transformers_version": "4.22.1",
34
- "vocab_size": 30522
35
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
last-checkpoint/optimizer.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:b357056f2accea2293c55790128b087fbf757fe7277557393925b49d2718f3c5
3
- size 535703265
 
 
 
 
last-checkpoint/pytorch_model.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:948a176b366086ca08ca3221df606c276b297992ebf8465411aeda21e91538e1
3
- size 267855985
 
 
 
 
last-checkpoint/rng_state.pth DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:238dfc24b3e3697078838e1aadf0fcdbdd30de36f27f5cb292ca17c74e1f813a
3
- size 13547
 
 
 
 
last-checkpoint/scheduler.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:35e659be997fa1542e5dfcf324806e102690c463edef7e011a08b6cc931fce20
3
- size 623
 
 
 
 
last-checkpoint/special_tokens_map.json DELETED
@@ -1,7 +0,0 @@
1
- {
2
- "cls_token": "[CLS]",
3
- "mask_token": "[MASK]",
4
- "pad_token": "[PAD]",
5
- "sep_token": "[SEP]",
6
- "unk_token": "[UNK]"
7
- }
 
 
 
 
 
 
 
 
last-checkpoint/tokenizer.json DELETED
The diff for this file is too large to render. See raw diff
 
last-checkpoint/tokenizer_config.json DELETED
@@ -1,14 +0,0 @@
1
- {
2
- "cls_token": "[CLS]",
3
- "do_lower_case": true,
4
- "mask_token": "[MASK]",
5
- "model_max_length": 512,
6
- "name_or_path": "distilbert-base-uncased",
7
- "pad_token": "[PAD]",
8
- "sep_token": "[SEP]",
9
- "special_tokens_map_file": null,
10
- "strip_accents": null,
11
- "tokenize_chinese_chars": true,
12
- "tokenizer_class": "DistilBertTokenizer",
13
- "unk_token": "[UNK]"
14
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
last-checkpoint/trainer_state.json DELETED
@@ -1,102 +0,0 @@
1
- {
2
- "best_metric": 0.8564085217556289,
3
- "best_model_checkpoint": "distilbert-base-uncased-Finanacial_Sentiment_Analysis_5e/checkpoint-670",
4
- "epoch": 5.0,
5
- "global_step": 670,
6
- "is_hyper_param_search": false,
7
- "is_local_process_zero": true,
8
- "is_world_process_zero": true,
9
- "log_history": [
10
- {
11
- "epoch": 0.01,
12
- "learning_rate": 1.9970149253731346e-05,
13
- "loss": 1.1124,
14
- "step": 1
15
- },
16
- {
17
- "epoch": 1.0,
18
- "learning_rate": 1.6000000000000003e-05,
19
- "loss": 0.5569,
20
- "step": 134
21
- },
22
- {
23
- "epoch": 1.0,
24
- "eval_accuracy": 0.7591377694470478,
25
- "eval_f1_score": 0.7558879300287058,
26
- "eval_loss": 0.39540889859199524,
27
- "eval_runtime": 231.6275,
28
- "eval_samples_per_second": 4.607,
29
- "eval_steps_per_second": 0.073,
30
- "step": 134
31
- },
32
- {
33
- "epoch": 2.0,
34
- "learning_rate": 1.2e-05,
35
- "loss": 0.3177,
36
- "step": 268
37
- },
38
- {
39
- "epoch": 2.0,
40
- "eval_accuracy": 0.8134957825679475,
41
- "eval_f1_score": 0.8151397951439218,
42
- "eval_loss": 0.3391183614730835,
43
- "eval_runtime": 231.1509,
44
- "eval_samples_per_second": 4.616,
45
- "eval_steps_per_second": 0.074,
46
- "step": 268
47
- },
48
- {
49
- "epoch": 3.0,
50
- "learning_rate": 8.000000000000001e-06,
51
- "loss": 0.2479,
52
- "step": 402
53
- },
54
- {
55
- "epoch": 3.0,
56
- "eval_accuracy": 0.8322399250234301,
57
- "eval_f1_score": 0.8353417132688705,
58
- "eval_loss": 0.3210773766040802,
59
- "eval_runtime": 231.1395,
60
- "eval_samples_per_second": 4.616,
61
- "eval_steps_per_second": 0.074,
62
- "step": 402
63
- },
64
- {
65
- "epoch": 4.0,
66
- "learning_rate": 4.000000000000001e-06,
67
- "loss": 0.2049,
68
- "step": 536
69
- },
70
- {
71
- "epoch": 4.0,
72
- "eval_accuracy": 0.8462980318650422,
73
- "eval_f1_score": 0.8506075560174677,
74
- "eval_loss": 0.3065849840641022,
75
- "eval_runtime": 231.1636,
76
- "eval_samples_per_second": 4.616,
77
- "eval_steps_per_second": 0.074,
78
- "step": 536
79
- },
80
- {
81
- "epoch": 5.0,
82
- "learning_rate": 0.0,
83
- "loss": 0.1802,
84
- "step": 670
85
- },
86
- {
87
- "epoch": 5.0,
88
- "eval_accuracy": 0.8528584817244611,
89
- "eval_f1_score": 0.8564085217556289,
90
- "eval_loss": 0.3078767955303192,
91
- "eval_runtime": 231.121,
92
- "eval_samples_per_second": 4.617,
93
- "eval_steps_per_second": 0.074,
94
- "step": 670
95
- }
96
- ],
97
- "max_steps": 670,
98
- "num_train_epochs": 5,
99
- "total_flos": 1148430034477440.0,
100
- "trial_name": null,
101
- "trial_params": null
102
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
last-checkpoint/training_args.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:b98d51052aee56b46aee98973da08c124641dbff311dcd0b662226571845cf68
3
- size 3439
 
 
 
 
last-checkpoint/vocab.txt DELETED
The diff for this file is too large to render. See raw diff
 
runs/Sep25_00-44-45_Brians-Mac-mini/events.out.tfevents.1664084692.Brians-Mac-mini.3612.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:22ae9bce003b666ba353afb6bb139fab69cb90c1d8e53dc1ff56e2aced17c35d
3
- size 6671
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b9760dcb2268ddc3731769652a6fadebe292d7a7e140db1383ac9b5daa374b92
3
+ size 7025