vietan32 commited on
Commit
4063540
·
verified ·
1 Parent(s): ced9653

Upload 9 files

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
config.json ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "XLMRobertaForQuestionAnswering"
4
+ ],
5
+ "attention_probs_dropout_prob": 0.1,
6
+ "bos_token_id": 0,
7
+ "classifier_dropout": null,
8
+ "eos_token_id": 2,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 3072,
14
+ "layer_norm_eps": 1e-05,
15
+ "max_position_embeddings": 514,
16
+ "model_type": "xlm-roberta",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 12,
19
+ "output_past": true,
20
+ "pad_token_id": 1,
21
+ "position_embedding_type": "absolute",
22
+ "torch_dtype": "float32",
23
+ "transformers_version": "4.51.3",
24
+ "type_vocab_size": 1,
25
+ "use_cache": true,
26
+ "vocab_size": 250002
27
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e0de0a8a597dcf25f892cb6e00f52f7b2f737c36247e4b26465913cfa26c2a52
3
+ size 1109842416
rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a7dd53f0df8b1d80033430fdedb0514af4adfe1f4a44b465e893e90f436b8150
3
+ size 14645
sentencepiece.bpe.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cfc8146abe2a0488e9e2a0c56de7952f7c11ab059eca145a0a727afce0db2865
3
+ size 5069051
special_tokens_map.json ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "cls_token": {
10
+ "content": "<s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "eos_token": {
17
+ "content": "</s>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "mask_token": {
24
+ "content": "<mask>",
25
+ "lstrip": true,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "pad_token": {
31
+ "content": "<pad>",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ },
37
+ "sep_token": {
38
+ "content": "</s>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false
43
+ },
44
+ "unk_token": {
45
+ "content": "<unk>",
46
+ "lstrip": false,
47
+ "normalized": false,
48
+ "rstrip": false,
49
+ "single_word": false
50
+ }
51
+ }
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e9092d8b5989337626af7001a7a97b88471e2da7c59dce83aba82a0ef90cad1f
3
+ size 17082999
tokenizer_config.json ADDED
@@ -0,0 +1,55 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "<s>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "<pad>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "</s>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "<unk>",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "250001": {
36
+ "content": "<mask>",
37
+ "lstrip": true,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "bos_token": "<s>",
45
+ "clean_up_tokenization_spaces": false,
46
+ "cls_token": "<s>",
47
+ "eos_token": "</s>",
48
+ "extra_special_tokens": {},
49
+ "mask_token": "<mask>",
50
+ "model_max_length": 512,
51
+ "pad_token": "<pad>",
52
+ "sep_token": "</s>",
53
+ "tokenizer_class": "XLMRobertaTokenizer",
54
+ "unk_token": "<unk>"
55
+ }
trainer_state.json ADDED
@@ -0,0 +1,270 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 5.0,
6
+ "eval_steps": 500,
7
+ "global_step": 14425,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.17335529167027824,
14
+ "grad_norm": 11.912790298461914,
15
+ "learning_rate": 1.0812567713976167e-06,
16
+ "loss": 5.3489,
17
+ "step": 500
18
+ },
19
+ {
20
+ "epoch": 0.3467105833405565,
21
+ "grad_norm": 26.164979934692383,
22
+ "learning_rate": 2.164680390032503e-06,
23
+ "loss": 2.4606,
24
+ "step": 1000
25
+ },
26
+ {
27
+ "epoch": 0.5200658750108347,
28
+ "grad_norm": 22.109607696533203,
29
+ "learning_rate": 3.248104008667389e-06,
30
+ "loss": 1.6107,
31
+ "step": 1500
32
+ },
33
+ {
34
+ "epoch": 0.693421166681113,
35
+ "grad_norm": 11.122232437133789,
36
+ "learning_rate": 4.331527627302276e-06,
37
+ "loss": 1.3092,
38
+ "step": 2000
39
+ },
40
+ {
41
+ "epoch": 0.8667764583513912,
42
+ "grad_norm": 9.637592315673828,
43
+ "learning_rate": 5.414951245937162e-06,
44
+ "loss": 1.1522,
45
+ "step": 2500
46
+ },
47
+ {
48
+ "epoch": 1.0,
49
+ "eval_loss": 0.9044516682624817,
50
+ "eval_runtime": 73.1068,
51
+ "eval_samples_per_second": 1322.489,
52
+ "eval_steps_per_second": 82.66,
53
+ "step": 2885
54
+ },
55
+ {
56
+ "epoch": 1.039871717084164,
57
+ "grad_norm": 8.937479972839355,
58
+ "learning_rate": 6.498374864572048e-06,
59
+ "loss": 1.0359,
60
+ "step": 3000
61
+ },
62
+ {
63
+ "epoch": 1.2132270087544423,
64
+ "grad_norm": 9.546417236328125,
65
+ "learning_rate": 7.581798483206935e-06,
66
+ "loss": 0.9387,
67
+ "step": 3500
68
+ },
69
+ {
70
+ "epoch": 1.3865823004247204,
71
+ "grad_norm": 12.781045913696289,
72
+ "learning_rate": 8.66522210184182e-06,
73
+ "loss": 0.8925,
74
+ "step": 4000
75
+ },
76
+ {
77
+ "epoch": 1.5599375920949987,
78
+ "grad_norm": 16.61449432373047,
79
+ "learning_rate": 9.748645720476708e-06,
80
+ "loss": 0.8351,
81
+ "step": 4500
82
+ },
83
+ {
84
+ "epoch": 1.733292883765277,
85
+ "grad_norm": 8.786149024963379,
86
+ "learning_rate": 9.791948854093299e-06,
87
+ "loss": 0.804,
88
+ "step": 5000
89
+ },
90
+ {
91
+ "epoch": 1.906648175435555,
92
+ "grad_norm": 8.35647201538086,
93
+ "learning_rate": 9.52104892452728e-06,
94
+ "loss": 0.7741,
95
+ "step": 5500
96
+ },
97
+ {
98
+ "epoch": 2.0,
99
+ "eval_loss": 0.681246817111969,
100
+ "eval_runtime": 73.1325,
101
+ "eval_samples_per_second": 1322.025,
102
+ "eval_steps_per_second": 82.631,
103
+ "step": 5770
104
+ },
105
+ {
106
+ "epoch": 2.079743434168328,
107
+ "grad_norm": 8.843521118164062,
108
+ "learning_rate": 9.250148994961262e-06,
109
+ "loss": 0.7289,
110
+ "step": 6000
111
+ },
112
+ {
113
+ "epoch": 2.253098725838606,
114
+ "grad_norm": 7.018795490264893,
115
+ "learning_rate": 8.979249065395244e-06,
116
+ "loss": 0.6891,
117
+ "step": 6500
118
+ },
119
+ {
120
+ "epoch": 2.4264540175088847,
121
+ "grad_norm": 7.121385097503662,
122
+ "learning_rate": 8.708349135829224e-06,
123
+ "loss": 0.6894,
124
+ "step": 7000
125
+ },
126
+ {
127
+ "epoch": 2.5998093091791628,
128
+ "grad_norm": 7.603145599365234,
129
+ "learning_rate": 8.437449206263208e-06,
130
+ "loss": 0.6573,
131
+ "step": 7500
132
+ },
133
+ {
134
+ "epoch": 2.773164600849441,
135
+ "grad_norm": 7.313769817352295,
136
+ "learning_rate": 8.16654927669719e-06,
137
+ "loss": 0.6599,
138
+ "step": 8000
139
+ },
140
+ {
141
+ "epoch": 2.946519892519719,
142
+ "grad_norm": 12.033502578735352,
143
+ "learning_rate": 7.89564934713117e-06,
144
+ "loss": 0.6459,
145
+ "step": 8500
146
+ },
147
+ {
148
+ "epoch": 3.0,
149
+ "eval_loss": 0.6194610595703125,
150
+ "eval_runtime": 73.1152,
151
+ "eval_samples_per_second": 1322.339,
152
+ "eval_steps_per_second": 82.65,
153
+ "step": 8655
154
+ },
155
+ {
156
+ "epoch": 3.119615151252492,
157
+ "grad_norm": 8.768052101135254,
158
+ "learning_rate": 7.6247494175651514e-06,
159
+ "loss": 0.6049,
160
+ "step": 9000
161
+ },
162
+ {
163
+ "epoch": 3.29297044292277,
164
+ "grad_norm": 8.126376152038574,
165
+ "learning_rate": 7.353849487999134e-06,
166
+ "loss": 0.5915,
167
+ "step": 9500
168
+ },
169
+ {
170
+ "epoch": 3.4663257345930485,
171
+ "grad_norm": 8.753697395324707,
172
+ "learning_rate": 7.082949558433115e-06,
173
+ "loss": 0.5856,
174
+ "step": 10000
175
+ },
176
+ {
177
+ "epoch": 3.6396810262633266,
178
+ "grad_norm": 8.999127388000488,
179
+ "learning_rate": 6.812049628867097e-06,
180
+ "loss": 0.5807,
181
+ "step": 10500
182
+ },
183
+ {
184
+ "epoch": 3.813036317933605,
185
+ "grad_norm": 8.415000915527344,
186
+ "learning_rate": 6.5411496993010795e-06,
187
+ "loss": 0.575,
188
+ "step": 11000
189
+ },
190
+ {
191
+ "epoch": 3.986391609603883,
192
+ "grad_norm": 9.230948448181152,
193
+ "learning_rate": 6.27024976973506e-06,
194
+ "loss": 0.5722,
195
+ "step": 11500
196
+ },
197
+ {
198
+ "epoch": 4.0,
199
+ "eval_loss": 0.6000474691390991,
200
+ "eval_runtime": 73.0477,
201
+ "eval_samples_per_second": 1323.56,
202
+ "eval_steps_per_second": 82.727,
203
+ "step": 11540
204
+ },
205
+ {
206
+ "epoch": 4.159486868336656,
207
+ "grad_norm": 10.552839279174805,
208
+ "learning_rate": 5.999349840169042e-06,
209
+ "loss": 0.5376,
210
+ "step": 12000
211
+ },
212
+ {
213
+ "epoch": 4.332842160006934,
214
+ "grad_norm": 9.041687965393066,
215
+ "learning_rate": 5.728449910603023e-06,
216
+ "loss": 0.5222,
217
+ "step": 12500
218
+ },
219
+ {
220
+ "epoch": 4.506197451677212,
221
+ "grad_norm": 11.044148445129395,
222
+ "learning_rate": 5.457549981037006e-06,
223
+ "loss": 0.5183,
224
+ "step": 13000
225
+ },
226
+ {
227
+ "epoch": 4.67955274334749,
228
+ "grad_norm": 8.885095596313477,
229
+ "learning_rate": 5.186650051470988e-06,
230
+ "loss": 0.522,
231
+ "step": 13500
232
+ },
233
+ {
234
+ "epoch": 4.852908035017769,
235
+ "grad_norm": 9.691495895385742,
236
+ "learning_rate": 4.9157501219049685e-06,
237
+ "loss": 0.5134,
238
+ "step": 14000
239
+ },
240
+ {
241
+ "epoch": 5.0,
242
+ "eval_loss": 0.5658867359161377,
243
+ "eval_runtime": 73.1854,
244
+ "eval_samples_per_second": 1321.07,
245
+ "eval_steps_per_second": 82.571,
246
+ "step": 14425
247
+ }
248
+ ],
249
+ "logging_steps": 500,
250
+ "max_steps": 23072,
251
+ "num_input_tokens_seen": 0,
252
+ "num_train_epochs": 8,
253
+ "save_steps": 500,
254
+ "stateful_callbacks": {
255
+ "TrainerControl": {
256
+ "args": {
257
+ "should_epoch_stop": false,
258
+ "should_evaluate": false,
259
+ "should_log": false,
260
+ "should_save": true,
261
+ "should_training_stop": false
262
+ },
263
+ "attributes": {}
264
+ }
265
+ },
266
+ "total_flos": 3.617438027185613e+17,
267
+ "train_batch_size": 16,
268
+ "trial_name": null,
269
+ "trial_params": null
270
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:017afcf1fd9814bcd9376ebaf89bdb7d98c1c3bc6e014b869bb77889a39234b7
3
+ size 5841