Upload folder using huggingface_hub
Browse files- .gitattributes +2 -0
- checkpoint-169000/config.json +79 -0
- checkpoint-169000/generation_config.json +6 -0
- checkpoint-169000/model.safetensors +3 -0
- checkpoint-169000/optimizer.pt +3 -0
- checkpoint-169000/rng_state_0.pth +3 -0
- checkpoint-169000/rng_state_1.pth +3 -0
- checkpoint-169000/rng_state_10.pth +3 -0
- checkpoint-169000/rng_state_11.pth +3 -0
- checkpoint-169000/rng_state_12.pth +3 -0
- checkpoint-169000/rng_state_13.pth +3 -0
- checkpoint-169000/rng_state_14.pth +3 -0
- checkpoint-169000/rng_state_15.pth +3 -0
- checkpoint-169000/rng_state_2.pth +3 -0
- checkpoint-169000/rng_state_3.pth +3 -0
- checkpoint-169000/rng_state_4.pth +3 -0
- checkpoint-169000/rng_state_5.pth +3 -0
- checkpoint-169000/rng_state_6.pth +3 -0
- checkpoint-169000/rng_state_7.pth +3 -0
- checkpoint-169000/rng_state_8.pth +3 -0
- checkpoint-169000/rng_state_9.pth +3 -0
- checkpoint-169000/scheduler.pt +3 -0
- checkpoint-169000/trainer_state.json +3 -0
- checkpoint-169000/training_args.bin +3 -0
- checkpoint-170000/config.json +79 -0
- checkpoint-170000/generation_config.json +6 -0
- checkpoint-170000/model.safetensors +3 -0
- checkpoint-170000/optimizer.pt +3 -0
- checkpoint-170000/rng_state_0.pth +3 -0
- checkpoint-170000/rng_state_1.pth +3 -0
- checkpoint-170000/rng_state_10.pth +3 -0
- checkpoint-170000/rng_state_11.pth +3 -0
- checkpoint-170000/rng_state_12.pth +3 -0
- checkpoint-170000/rng_state_13.pth +3 -0
- checkpoint-170000/rng_state_14.pth +3 -0
- checkpoint-170000/rng_state_15.pth +3 -0
- checkpoint-170000/rng_state_2.pth +3 -0
- checkpoint-170000/rng_state_3.pth +3 -0
- checkpoint-170000/rng_state_4.pth +3 -0
- checkpoint-170000/rng_state_5.pth +3 -0
- checkpoint-170000/rng_state_6.pth +3 -0
- checkpoint-170000/rng_state_7.pth +3 -0
- checkpoint-170000/rng_state_8.pth +3 -0
- checkpoint-170000/rng_state_9.pth +3 -0
- checkpoint-170000/scheduler.pt +3 -0
- checkpoint-170000/trainer_state.json +3 -0
- checkpoint-170000/training_args.bin +3 -0
.gitattributes
CHANGED
@@ -69,3 +69,5 @@ checkpoint-159000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
|
|
69 |
checkpoint-160000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
|
70 |
checkpoint-165000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
|
71 |
checkpoint-166000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
69 |
checkpoint-160000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
|
70 |
checkpoint-165000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
|
71 |
checkpoint-166000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
|
72 |
+
checkpoint-169000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
|
73 |
+
checkpoint-170000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
|
checkpoint-169000/config.json
ADDED
@@ -0,0 +1,79 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"ar_steps": 1,
|
3 |
+
"architectures": [
|
4 |
+
"DiffVLMBaseline"
|
5 |
+
],
|
6 |
+
"attention_dropout": 0.0,
|
7 |
+
"bos_token_id": 151643,
|
8 |
+
"condition_layer": -1,
|
9 |
+
"eos_token_id": 151645,
|
10 |
+
"hidden_act": "silu",
|
11 |
+
"hidden_size": 1536,
|
12 |
+
"image_token_id": 151655,
|
13 |
+
"img_cross_attention_dim": 2048,
|
14 |
+
"img_diffuser_depth": 2,
|
15 |
+
"img_ffn_dim_multiplier": null,
|
16 |
+
"img_hidden_size": 1536,
|
17 |
+
"img_multiple_of": 256,
|
18 |
+
"img_norm_eps": 1e-05,
|
19 |
+
"img_num_attention_heads": 12,
|
20 |
+
"img_num_kv_heads": 12,
|
21 |
+
"img_qk_norm": true,
|
22 |
+
"in_channels": 32,
|
23 |
+
"initializer_range": 0.02,
|
24 |
+
"inject_img_diffuser": false,
|
25 |
+
"input_size": 32,
|
26 |
+
"intermediate_size": 8960,
|
27 |
+
"layer_group_size": 7,
|
28 |
+
"layerwise_start_idx": 0,
|
29 |
+
"lora_alpha": 128,
|
30 |
+
"lora_bias": "none",
|
31 |
+
"lora_dropout": 0.05,
|
32 |
+
"lora_enable": false,
|
33 |
+
"lora_r": 64,
|
34 |
+
"max_position_embeddings": 32768,
|
35 |
+
"max_window_layers": 28,
|
36 |
+
"model_type": "qwen2_vl",
|
37 |
+
"non_linearity": 1,
|
38 |
+
"norm_elementwise_affine": true,
|
39 |
+
"num_attention_heads": 12,
|
40 |
+
"num_hidden_layers": 28,
|
41 |
+
"num_key_value_heads": 2,
|
42 |
+
"patch_size": 1,
|
43 |
+
"repa_coeff": 0.1,
|
44 |
+
"repa_layers": "2",
|
45 |
+
"repa_shared": false,
|
46 |
+
"rms_norm_eps": 1e-06,
|
47 |
+
"rope_scaling": {
|
48 |
+
"mrope_section": [
|
49 |
+
16,
|
50 |
+
24,
|
51 |
+
24
|
52 |
+
],
|
53 |
+
"rope_type": "default",
|
54 |
+
"type": "default"
|
55 |
+
},
|
56 |
+
"rope_theta": 1000000.0,
|
57 |
+
"sample_size": 128,
|
58 |
+
"sampling_steps": 28,
|
59 |
+
"sliding_window": null,
|
60 |
+
"tie_word_embeddings": true,
|
61 |
+
"torch_dtype": "bfloat16",
|
62 |
+
"transformers_version": "4.47.0",
|
63 |
+
"use_cache": true,
|
64 |
+
"use_repa": false,
|
65 |
+
"use_residual_attn": false,
|
66 |
+
"use_sliding_window": false,
|
67 |
+
"vae_path": "mit-han-lab/dc-ae-f32c32-in-1.0-diffusers",
|
68 |
+
"video_token_id": 151656,
|
69 |
+
"vision_config": {
|
70 |
+
"hidden_size": 1536,
|
71 |
+
"in_chans": 3,
|
72 |
+
"model_type": "qwen2_vl",
|
73 |
+
"spatial_patch_size": 14
|
74 |
+
},
|
75 |
+
"vision_end_token_id": 151653,
|
76 |
+
"vision_start_token_id": 151652,
|
77 |
+
"vision_token_id": 151654,
|
78 |
+
"vocab_size": 151936
|
79 |
+
}
|
checkpoint-169000/generation_config.json
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_from_model_config": true,
|
3 |
+
"bos_token_id": 151643,
|
4 |
+
"eos_token_id": 151645,
|
5 |
+
"transformers_version": "4.47.0"
|
6 |
+
}
|
checkpoint-169000/model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ba281ac0401d40141a001182ca9effb079940f500ee68827cd89ff90fbe6e54b
|
3 |
+
size 4410723984
|
checkpoint-169000/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:972cbc3ee09da9e57301ae25087d89b82aa944392c8614be66b30a467d82445c
|
3 |
+
size 6330255386
|
checkpoint-169000/rng_state_0.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d65b068c445273015aff639e0140b5291c9928c6925cb25f76ea08110792b570
|
3 |
+
size 15984
|
checkpoint-169000/rng_state_1.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d39cf304928695686a88543679f5cbd0ad13e29faf0ed22c27a68c1f44f62092
|
3 |
+
size 15984
|
checkpoint-169000/rng_state_10.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9d4084d8aa6ffdc45108c951524de7babd6c1d809d4fc115fccc6fda9290d64f
|
3 |
+
size 15997
|
checkpoint-169000/rng_state_11.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bbabfd154f333f07c7c242694e034579f273f6caea04d1525acd64746f6d4fed
|
3 |
+
size 15997
|
checkpoint-169000/rng_state_12.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:245ad9b4d9e08b9d954c657ad86a1c8f71bf398ea89b32ab920d2db8e7438c10
|
3 |
+
size 15997
|
checkpoint-169000/rng_state_13.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d3b363883c58b576883cdaedd1ac1575fa7efb500a1c151568a73785ca6b6807
|
3 |
+
size 15997
|
checkpoint-169000/rng_state_14.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0bd3dfa319f56fcfa2ce2a2aff7db553493589fd8cf7c2b8e4f5fe3aa2a7d91e
|
3 |
+
size 15997
|
checkpoint-169000/rng_state_15.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9103d24ac96367df0bbc95e545de2f1f7385988e8eab415735017dabf6cdeefc
|
3 |
+
size 15997
|
checkpoint-169000/rng_state_2.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:898829f08fbb97b3d609c393906795bc5a354eae3110eaf3fadced55f1c3faa1
|
3 |
+
size 15984
|
checkpoint-169000/rng_state_3.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3e30690d34bea2711084f01fa62b12d8da8f9f73dcc44c75fc0f9a9cd1233b64
|
3 |
+
size 15984
|
checkpoint-169000/rng_state_4.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7d1b8a6176c38a50c222e1d983cd0f7cff62e35408e3d3dfa6b258cfae4cb0c2
|
3 |
+
size 15984
|
checkpoint-169000/rng_state_5.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:605cb6cadb294b327cfda902273750754f98e123eda2e856b145ea5fb27cd274
|
3 |
+
size 15984
|
checkpoint-169000/rng_state_6.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0a639a9b58f319bb85d2ab94a35e66250f7d74b273438294c8b0a2a3b1f42a77
|
3 |
+
size 15984
|
checkpoint-169000/rng_state_7.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:61b7255c2bdece5a1124b7f6a52c7b3182745c86a91d2c2ad47ad2a0878f9a89
|
3 |
+
size 15984
|
checkpoint-169000/rng_state_8.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:816907dea154d707f5d2c2f09eef3b5c72c19c2650c5a3eb847bbff371d32dc0
|
3 |
+
size 15984
|
checkpoint-169000/rng_state_9.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:186aba890ab1b2305469af17dc345a88529cd6341544ba4b337a79d8a65d8994
|
3 |
+
size 15984
|
checkpoint-169000/scheduler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:83de56ae994f16992f17aaac3675983d0052ef337b3aec07d00c998a5092c000
|
3 |
+
size 1064
|
checkpoint-169000/trainer_state.json
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:41644007e5ee42de699c99e21aa8982f5aee1b1eb1fa79bedb5a85454bb6ff39
|
3 |
+
size 25820728
|
checkpoint-169000/training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:78baaaf0e002052f7ca00642775f5bb891516dfb2c2aa21690f3e01074eb4505
|
3 |
+
size 5944
|
checkpoint-170000/config.json
ADDED
@@ -0,0 +1,79 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"ar_steps": 1,
|
3 |
+
"architectures": [
|
4 |
+
"DiffVLMBaseline"
|
5 |
+
],
|
6 |
+
"attention_dropout": 0.0,
|
7 |
+
"bos_token_id": 151643,
|
8 |
+
"condition_layer": -1,
|
9 |
+
"eos_token_id": 151645,
|
10 |
+
"hidden_act": "silu",
|
11 |
+
"hidden_size": 1536,
|
12 |
+
"image_token_id": 151655,
|
13 |
+
"img_cross_attention_dim": 2048,
|
14 |
+
"img_diffuser_depth": 2,
|
15 |
+
"img_ffn_dim_multiplier": null,
|
16 |
+
"img_hidden_size": 1536,
|
17 |
+
"img_multiple_of": 256,
|
18 |
+
"img_norm_eps": 1e-05,
|
19 |
+
"img_num_attention_heads": 12,
|
20 |
+
"img_num_kv_heads": 12,
|
21 |
+
"img_qk_norm": true,
|
22 |
+
"in_channels": 32,
|
23 |
+
"initializer_range": 0.02,
|
24 |
+
"inject_img_diffuser": false,
|
25 |
+
"input_size": 32,
|
26 |
+
"intermediate_size": 8960,
|
27 |
+
"layer_group_size": 7,
|
28 |
+
"layerwise_start_idx": 0,
|
29 |
+
"lora_alpha": 128,
|
30 |
+
"lora_bias": "none",
|
31 |
+
"lora_dropout": 0.05,
|
32 |
+
"lora_enable": false,
|
33 |
+
"lora_r": 64,
|
34 |
+
"max_position_embeddings": 32768,
|
35 |
+
"max_window_layers": 28,
|
36 |
+
"model_type": "qwen2_vl",
|
37 |
+
"non_linearity": 1,
|
38 |
+
"norm_elementwise_affine": true,
|
39 |
+
"num_attention_heads": 12,
|
40 |
+
"num_hidden_layers": 28,
|
41 |
+
"num_key_value_heads": 2,
|
42 |
+
"patch_size": 1,
|
43 |
+
"repa_coeff": 0.1,
|
44 |
+
"repa_layers": "2",
|
45 |
+
"repa_shared": false,
|
46 |
+
"rms_norm_eps": 1e-06,
|
47 |
+
"rope_scaling": {
|
48 |
+
"mrope_section": [
|
49 |
+
16,
|
50 |
+
24,
|
51 |
+
24
|
52 |
+
],
|
53 |
+
"rope_type": "default",
|
54 |
+
"type": "default"
|
55 |
+
},
|
56 |
+
"rope_theta": 1000000.0,
|
57 |
+
"sample_size": 128,
|
58 |
+
"sampling_steps": 28,
|
59 |
+
"sliding_window": null,
|
60 |
+
"tie_word_embeddings": true,
|
61 |
+
"torch_dtype": "bfloat16",
|
62 |
+
"transformers_version": "4.47.0",
|
63 |
+
"use_cache": true,
|
64 |
+
"use_repa": false,
|
65 |
+
"use_residual_attn": false,
|
66 |
+
"use_sliding_window": false,
|
67 |
+
"vae_path": "mit-han-lab/dc-ae-f32c32-in-1.0-diffusers",
|
68 |
+
"video_token_id": 151656,
|
69 |
+
"vision_config": {
|
70 |
+
"hidden_size": 1536,
|
71 |
+
"in_chans": 3,
|
72 |
+
"model_type": "qwen2_vl",
|
73 |
+
"spatial_patch_size": 14
|
74 |
+
},
|
75 |
+
"vision_end_token_id": 151653,
|
76 |
+
"vision_start_token_id": 151652,
|
77 |
+
"vision_token_id": 151654,
|
78 |
+
"vocab_size": 151936
|
79 |
+
}
|
checkpoint-170000/generation_config.json
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_from_model_config": true,
|
3 |
+
"bos_token_id": 151643,
|
4 |
+
"eos_token_id": 151645,
|
5 |
+
"transformers_version": "4.47.0"
|
6 |
+
}
|
checkpoint-170000/model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:73d4ce449c19f75cdd21c226d4baea75e9155a1db69b6e9d9bd3b13eb25bfa84
|
3 |
+
size 4410723984
|
checkpoint-170000/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0f3985e3e233da6969b98ed63c87fe8a823eedaff9e86c8bc989b8b2bd8762cf
|
3 |
+
size 6330255386
|
checkpoint-170000/rng_state_0.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:81142f0631ca973174c405c57e5d963212b19cedf8f21109e4e79656d58fedd7
|
3 |
+
size 15984
|
checkpoint-170000/rng_state_1.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1dfb43dce8ae40db3177be1793c096ca8258cc4da99b128c9e17146ebd9ec4af
|
3 |
+
size 15984
|
checkpoint-170000/rng_state_10.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9b64a6f13c96f95c4b65ddd463c6adad9c982318754adf497267274c39ce0f2c
|
3 |
+
size 15997
|
checkpoint-170000/rng_state_11.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2c091badd3838b5b68e409d4cfc48661105b86c51634fa0416046a7af7494529
|
3 |
+
size 15997
|
checkpoint-170000/rng_state_12.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fec19c41bb03ba94266dd82dc16cf9a09018c6cc4a8b4c0b0b5dff49c01392c0
|
3 |
+
size 15997
|
checkpoint-170000/rng_state_13.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a98f3544f4a0dbf1a574690618c190eaf41d409d7a7d68d16d28f56ce6bfbe98
|
3 |
+
size 15997
|
checkpoint-170000/rng_state_14.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:254743c6a556ef01af55db3f31c2b7a5a9563ae8da3c39f991a4735c3f715874
|
3 |
+
size 15997
|
checkpoint-170000/rng_state_15.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5e21a71ce27a989e3d7b16af8d45c6cad02af8c5b760c78a52aa9df68700cfed
|
3 |
+
size 15997
|
checkpoint-170000/rng_state_2.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:abfedb0c070f079bf38d81d178543dd487e855f4d902c1df532391bd4325b30b
|
3 |
+
size 15984
|
checkpoint-170000/rng_state_3.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:803db2891b0e13b7a5b76bca385b54cd45e7a506bd345ba9b0e9dd8c12e2039e
|
3 |
+
size 15984
|
checkpoint-170000/rng_state_4.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e43d46b1deb17c9c811ba4df651c84836ad4ef3c1c40806fc33d87c806f5feaa
|
3 |
+
size 15984
|
checkpoint-170000/rng_state_5.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7b4f6930049cc6386ad42b383aba0656ddf57770127d4de764e1e96664bac4a8
|
3 |
+
size 15984
|
checkpoint-170000/rng_state_6.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f3f71bbe3d75f7252546a60532734245fecca9492bae678fcb0595f387dbc5d4
|
3 |
+
size 15984
|
checkpoint-170000/rng_state_7.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:05df5a17f56dec711f9c195581f678ea0bbccffa467e08d2b219e482c5d6f978
|
3 |
+
size 15984
|
checkpoint-170000/rng_state_8.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:78b8b1f953141d0c2af64378da72d92b45fe1ad12d6b8e33a7eb6c29cf1766da
|
3 |
+
size 15984
|
checkpoint-170000/rng_state_9.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f6cb12436d41d2d19158ed1a9548bc8ebf672c9ccbb87f3bebe2479c459b51c1
|
3 |
+
size 15984
|
checkpoint-170000/scheduler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:53418078886f31c0993feeddb7d41f245589575ed7c2ae4e89cfde013fe46589
|
3 |
+
size 1064
|
checkpoint-170000/trainer_state.json
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:32f993885481ffe6e9a04cb5b0a7a124433272be1d18f38c9797393089bb7aa3
|
3 |
+
size 25974108
|
checkpoint-170000/training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:78baaaf0e002052f7ca00642775f5bb891516dfb2c2aa21690f3e01074eb4505
|
3 |
+
size 5944
|