Finerio commited on
Commit
4ae4baf
·
verified ·
1 Parent(s): 5010f7b

Upload folder using huggingface_hub

Browse files
best/added_tokens.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "$START": 50265
3
+ }
best/config.json ADDED
The diff for this file is too large to render. See raw diff
 
best/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
best/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:32514302c01cab5a24931df3ead6c76cff11b7a557fcf2b49d5f240666708d3a
3
+ size 511632936
best/special_tokens_map.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ {
4
+ "content": "$START",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false
9
+ }
10
+ ],
11
+ "bos_token": "<s>",
12
+ "cls_token": "<s>",
13
+ "eos_token": "</s>",
14
+ "mask_token": {
15
+ "content": "<mask>",
16
+ "lstrip": true,
17
+ "normalized": false,
18
+ "rstrip": false,
19
+ "single_word": false
20
+ },
21
+ "pad_token": "<pad>",
22
+ "sep_token": "</s>",
23
+ "unk_token": "<unk>"
24
+ }
best/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
best/tokenizer_config.json ADDED
@@ -0,0 +1,69 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": true,
3
+ "added_tokens_decoder": {
4
+ "0": {
5
+ "content": "<s>",
6
+ "lstrip": false,
7
+ "normalized": true,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "1": {
13
+ "content": "<pad>",
14
+ "lstrip": false,
15
+ "normalized": true,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "2": {
21
+ "content": "</s>",
22
+ "lstrip": false,
23
+ "normalized": true,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ },
28
+ "3": {
29
+ "content": "<unk>",
30
+ "lstrip": false,
31
+ "normalized": true,
32
+ "rstrip": false,
33
+ "single_word": false,
34
+ "special": true
35
+ },
36
+ "50264": {
37
+ "content": "<mask>",
38
+ "lstrip": true,
39
+ "normalized": false,
40
+ "rstrip": false,
41
+ "single_word": false,
42
+ "special": true
43
+ },
44
+ "50265": {
45
+ "content": "$START",
46
+ "lstrip": false,
47
+ "normalized": false,
48
+ "rstrip": false,
49
+ "single_word": false,
50
+ "special": true
51
+ }
52
+ },
53
+ "additional_special_tokens": [
54
+ "$START"
55
+ ],
56
+ "bos_token": "<s>",
57
+ "clean_up_tokenization_spaces": false,
58
+ "cls_token": "<s>",
59
+ "eos_token": "</s>",
60
+ "errors": "replace",
61
+ "extra_special_tokens": {},
62
+ "mask_token": "<mask>",
63
+ "model_max_length": 512,
64
+ "pad_token": "<pad>",
65
+ "sep_token": "</s>",
66
+ "tokenizer_class": "RobertaTokenizer",
67
+ "trim_offsets": true,
68
+ "unk_token": "<unk>"
69
+ }
best/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
last/added_tokens.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "$START": 50265
3
+ }
last/config.json ADDED
The diff for this file is too large to render. See raw diff
 
last/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
last/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:32514302c01cab5a24931df3ead6c76cff11b7a557fcf2b49d5f240666708d3a
3
+ size 511632936
last/special_tokens_map.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ {
4
+ "content": "$START",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false
9
+ }
10
+ ],
11
+ "bos_token": "<s>",
12
+ "cls_token": "<s>",
13
+ "eos_token": "</s>",
14
+ "mask_token": {
15
+ "content": "<mask>",
16
+ "lstrip": true,
17
+ "normalized": false,
18
+ "rstrip": false,
19
+ "single_word": false
20
+ },
21
+ "pad_token": "<pad>",
22
+ "sep_token": "</s>",
23
+ "unk_token": "<unk>"
24
+ }
last/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
last/tokenizer_config.json ADDED
@@ -0,0 +1,69 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": true,
3
+ "added_tokens_decoder": {
4
+ "0": {
5
+ "content": "<s>",
6
+ "lstrip": false,
7
+ "normalized": true,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "1": {
13
+ "content": "<pad>",
14
+ "lstrip": false,
15
+ "normalized": true,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "2": {
21
+ "content": "</s>",
22
+ "lstrip": false,
23
+ "normalized": true,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ },
28
+ "3": {
29
+ "content": "<unk>",
30
+ "lstrip": false,
31
+ "normalized": true,
32
+ "rstrip": false,
33
+ "single_word": false,
34
+ "special": true
35
+ },
36
+ "50264": {
37
+ "content": "<mask>",
38
+ "lstrip": true,
39
+ "normalized": false,
40
+ "rstrip": false,
41
+ "single_word": false,
42
+ "special": true
43
+ },
44
+ "50265": {
45
+ "content": "$START",
46
+ "lstrip": false,
47
+ "normalized": false,
48
+ "rstrip": false,
49
+ "single_word": false,
50
+ "special": true
51
+ }
52
+ },
53
+ "additional_special_tokens": [
54
+ "$START"
55
+ ],
56
+ "bos_token": "<s>",
57
+ "clean_up_tokenization_spaces": false,
58
+ "cls_token": "<s>",
59
+ "eos_token": "</s>",
60
+ "errors": "replace",
61
+ "extra_special_tokens": {},
62
+ "mask_token": "<mask>",
63
+ "model_max_length": 512,
64
+ "pad_token": "<pad>",
65
+ "sep_token": "</s>",
66
+ "tokenizer_class": "RobertaTokenizer",
67
+ "trim_offsets": true,
68
+ "unk_token": "<unk>"
69
+ }
last/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
log.json ADDED
@@ -0,0 +1,194 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "argparse": {
3
+ "train_file": "gector/data/train_preprocessed.pt",
4
+ "valid_file": "gector/data/valid_preprocessed.pt",
5
+ "model_id": "FacebookAI/roberta-base",
6
+ "batch_size": 32,
7
+ "early_stopper": 8,
8
+ "delimeter": "SEPL|||SEPR",
9
+ "additional_delimeter": "SEPL__SEPR",
10
+ "restore_dir": null,
11
+ "restore_vocab": null,
12
+ "restore_vocab_official": null,
13
+ "save_dir": "drive/MyDrive/outputs/gector_finerio",
14
+ "max_len": 128,
15
+ "n_max_labels": 5000,
16
+ "n_epochs": 120,
17
+ "p_dropout": 0.0,
18
+ "lr": 1e-05,
19
+ "cold_lr": 0.001,
20
+ "accumulation": 1,
21
+ "seed": 10,
22
+ "label_smoothing": 0.0,
23
+ "n_cold_epochs": 2,
24
+ "num_warmup_steps": 500,
25
+ "lr_scheduler_type": "constant"
26
+ },
27
+ "Epoch 0": {
28
+ "train_log": {
29
+ "loss": 1.7032910907026595,
30
+ "accuracy": 0.803969903675253,
31
+ "accuracy_d": 0.8724265860173506
32
+ },
33
+ "valid_log": {
34
+ "loss": 1.1267204428979771,
35
+ "accuracy": 0.8362742538576002,
36
+ "accuracy_d": 0.8833048800369362,
37
+ "message": "The best checkpoint has been updated."
38
+ }
39
+ },
40
+ "Epoch 1": {
41
+ "train_log": {
42
+ "loss": 1.2184834499734831,
43
+ "accuracy": 0.8329997152576303,
44
+ "accuracy_d": 0.8792239682540636
45
+ },
46
+ "valid_log": {
47
+ "loss": 0.9992180193510037,
48
+ "accuracy": 0.8476613973421157,
49
+ "accuracy_d": 0.8882333797736159,
50
+ "message": "The best checkpoint has been updated."
51
+ }
52
+ },
53
+ "Epoch 2": {
54
+ "train_log": {
55
+ "loss": 0.8153748830096256,
56
+ "accuracy": 0.8644381554725395,
57
+ "accuracy_d": 0.9125891849665446
58
+ },
59
+ "valid_log": {
60
+ "loss": 0.708008555180509,
61
+ "accuracy": 0.8810764396345461,
62
+ "accuracy_d": 0.9304811779775956,
63
+ "message": "The best checkpoint has been updated."
64
+ }
65
+ },
66
+ "Epoch 3": {
67
+ "train_log": {
68
+ "loss": 0.670869435280431,
69
+ "accuracy": 0.8802149062891003,
70
+ "accuracy_d": 0.9280548943165902
71
+ },
72
+ "valid_log": {
73
+ "loss": 0.6924886661192924,
74
+ "accuracy": 0.8869327425072056,
75
+ "accuracy_d": 0.9310040617696872,
76
+ "message": "The best checkpoint has been updated."
77
+ }
78
+ },
79
+ "Epoch 4": {
80
+ "train_log": {
81
+ "loss": 0.5957314862146635,
82
+ "accuracy": 0.88956841407807,
83
+ "accuracy_d": 0.9349972145328757
84
+ },
85
+ "valid_log": {
86
+ "loss": 0.6383009447624599,
87
+ "accuracy": 0.8902179040360318,
88
+ "accuracy_d": 0.9384833052339713,
89
+ "message": "The best checkpoint has been updated."
90
+ }
91
+ },
92
+ "Epoch 5": {
93
+ "train_log": {
94
+ "loss": 0.5390387795334833,
95
+ "accuracy": 0.8971076633320166,
96
+ "accuracy_d": 0.940645343271613
97
+ },
98
+ "valid_log": {
99
+ "loss": 0.6460947684376934,
100
+ "accuracy": 0.8933471329579327,
101
+ "accuracy_d": 0.9373614872099075,
102
+ "message": "The best checkpoint has been updated."
103
+ }
104
+ },
105
+ "Epoch 6": {
106
+ "train_log": {
107
+ "loss": 0.48871075836634,
108
+ "accuracy": 0.9041179819406858,
109
+ "accuracy_d": 0.9453216991100091
110
+ },
111
+ "valid_log": {
112
+ "loss": 0.6075716740235769,
113
+ "accuracy": 0.894183147373801,
114
+ "accuracy_d": 0.9401003544759662,
115
+ "message": "The best checkpoint has been updated."
116
+ }
117
+ },
118
+ "Epoch 7": {
119
+ "train_log": {
120
+ "loss": 0.4453746993235511,
121
+ "accuracy": 0.9112622401081515,
122
+ "accuracy_d": 0.9495536471997605
123
+ },
124
+ "valid_log": {
125
+ "loss": 0.620118053674919,
126
+ "accuracy": 0.8942456968640131,
127
+ "accuracy_d": 0.9408376017837666,
128
+ "message": "The best checkpoint has been updated."
129
+ }
130
+ },
131
+ "Epoch 8": {
132
+ "train_log": {
133
+ "loss": 0.4053901939949712,
134
+ "accuracy": 0.9178540091138886,
135
+ "accuracy_d": 0.9538734333239084
136
+ },
137
+ "valid_log": {
138
+ "loss": 0.6320258723727405,
139
+ "accuracy": 0.8951096773589918,
140
+ "accuracy_d": 0.9403002296396443,
141
+ "message": "The best checkpoint has been updated."
142
+ }
143
+ },
144
+ "Epoch 9": {
145
+ "train_log": {
146
+ "loss": 0.3688604742395323,
147
+ "accuracy": 0.9238926816214331,
148
+ "accuracy_d": 0.9574667460833989
149
+ },
150
+ "valid_log": {
151
+ "loss": 0.6419418054962865,
152
+ "accuracy": 0.8946705572237995,
153
+ "accuracy_d": 0.9425146041199537
154
+ }
155
+ },
156
+ "Epoch 10": {
157
+ "train_log": {
158
+ "loss": 0.33416874041144135,
159
+ "accuracy": 0.9297098048028365,
160
+ "accuracy_d": 0.9619453058886044
161
+ },
162
+ "valid_log": {
163
+ "loss": 0.6550916678000912,
164
+ "accuracy": 0.8971033769768554,
165
+ "accuracy_d": 0.9415456193718706,
166
+ "message": "The best checkpoint has been updated."
167
+ }
168
+ },
169
+ "Epoch 11": {
170
+ "train_log": {
171
+ "loss": 0.303828910638019,
172
+ "accuracy": 0.935022974531346,
173
+ "accuracy_d": 0.9651207187191879
174
+ },
175
+ "valid_log": {
176
+ "loss": 0.6778709775386361,
177
+ "accuracy": 0.8962904538869415,
178
+ "accuracy_d": 0.9426721779004099
179
+ }
180
+ },
181
+ "Epoch 12": {
182
+ "train_log": {
183
+ "loss": 0.2758491554175579,
184
+ "accuracy": 0.9401938989966029,
185
+ "accuracy_d": 0.9684101530294043
186
+ },
187
+ "valid_log": {
188
+ "loss": 0.695371216380751,
189
+ "accuracy": 0.897200845168119,
190
+ "accuracy_d": 0.9425573480593694,
191
+ "message": "The best checkpoint has been updated."
192
+ }
193
+ }
194
+ }