Upload folder using huggingface_hub
Browse files- .gitattributes +2 -0
- checkpoint-96000/config.json +79 -0
- checkpoint-96000/generation_config.json +6 -0
- checkpoint-96000/model.safetensors +3 -0
- checkpoint-96000/optimizer.pt +3 -0
- checkpoint-96000/rng_state_0.pth +3 -0
- checkpoint-96000/rng_state_1.pth +3 -0
- checkpoint-96000/rng_state_10.pth +3 -0
- checkpoint-96000/rng_state_11.pth +3 -0
- checkpoint-96000/rng_state_12.pth +3 -0
- checkpoint-96000/rng_state_13.pth +3 -0
- checkpoint-96000/rng_state_14.pth +3 -0
- checkpoint-96000/rng_state_15.pth +3 -0
- checkpoint-96000/rng_state_2.pth +3 -0
- checkpoint-96000/rng_state_3.pth +3 -0
- checkpoint-96000/rng_state_4.pth +3 -0
- checkpoint-96000/rng_state_5.pth +3 -0
- checkpoint-96000/rng_state_6.pth +3 -0
- checkpoint-96000/rng_state_7.pth +3 -0
- checkpoint-96000/rng_state_8.pth +3 -0
- checkpoint-96000/rng_state_9.pth +3 -0
- checkpoint-96000/scheduler.pt +3 -0
- checkpoint-96000/trainer_state.json +3 -0
- checkpoint-96000/training_args.bin +3 -0
- checkpoint-97000/config.json +79 -0
- checkpoint-97000/generation_config.json +6 -0
- checkpoint-97000/model.safetensors +3 -0
- checkpoint-97000/optimizer.pt +3 -0
- checkpoint-97000/rng_state_0.pth +3 -0
- checkpoint-97000/rng_state_1.pth +3 -0
- checkpoint-97000/rng_state_10.pth +3 -0
- checkpoint-97000/rng_state_11.pth +3 -0
- checkpoint-97000/rng_state_12.pth +3 -0
- checkpoint-97000/rng_state_13.pth +3 -0
- checkpoint-97000/rng_state_14.pth +3 -0
- checkpoint-97000/rng_state_15.pth +3 -0
- checkpoint-97000/rng_state_2.pth +3 -0
- checkpoint-97000/rng_state_3.pth +3 -0
- checkpoint-97000/rng_state_4.pth +3 -0
- checkpoint-97000/rng_state_5.pth +3 -0
- checkpoint-97000/rng_state_6.pth +3 -0
- checkpoint-97000/rng_state_7.pth +3 -0
- checkpoint-97000/rng_state_8.pth +3 -0
- checkpoint-97000/rng_state_9.pth +3 -0
- checkpoint-97000/scheduler.pt +3 -0
- checkpoint-97000/trainer_state.json +3 -0
- checkpoint-97000/training_args.bin +3 -0
.gitattributes
CHANGED
@@ -43,3 +43,5 @@ checkpoint-81000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
|
|
43 |
checkpoint-88000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
|
44 |
checkpoint-89000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
|
45 |
checkpoint-90000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
43 |
checkpoint-88000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
|
44 |
checkpoint-89000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
|
45 |
checkpoint-90000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
|
46 |
+
checkpoint-96000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
|
47 |
+
checkpoint-97000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
|
checkpoint-96000/config.json
ADDED
@@ -0,0 +1,79 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"ar_steps": 1,
|
3 |
+
"architectures": [
|
4 |
+
"DiffVLMDiffusion"
|
5 |
+
],
|
6 |
+
"attention_dropout": 0.0,
|
7 |
+
"bos_token_id": 151643,
|
8 |
+
"condition_layer": -1,
|
9 |
+
"eos_token_id": 151645,
|
10 |
+
"hidden_act": "silu",
|
11 |
+
"hidden_size": 1536,
|
12 |
+
"image_token_id": 151655,
|
13 |
+
"img_cross_attention_dim": 2048,
|
14 |
+
"img_diffuser_depth": 2,
|
15 |
+
"img_ffn_dim_multiplier": null,
|
16 |
+
"img_hidden_size": 1536,
|
17 |
+
"img_multiple_of": 256,
|
18 |
+
"img_norm_eps": 1e-05,
|
19 |
+
"img_num_attention_heads": 12,
|
20 |
+
"img_num_kv_heads": 12,
|
21 |
+
"img_qk_norm": true,
|
22 |
+
"in_channels": 32,
|
23 |
+
"initializer_range": 0.02,
|
24 |
+
"inject_img_diffuser": false,
|
25 |
+
"input_size": 32,
|
26 |
+
"intermediate_size": 8960,
|
27 |
+
"layer_group_size": 7,
|
28 |
+
"layerwise_start_idx": 0,
|
29 |
+
"lora_alpha": 128,
|
30 |
+
"lora_bias": "none",
|
31 |
+
"lora_dropout": 0.05,
|
32 |
+
"lora_enable": false,
|
33 |
+
"lora_r": 64,
|
34 |
+
"max_position_embeddings": 32768,
|
35 |
+
"max_window_layers": 28,
|
36 |
+
"model_type": "qwen2_vl",
|
37 |
+
"non_linearity": 1,
|
38 |
+
"norm_elementwise_affine": true,
|
39 |
+
"num_attention_heads": 12,
|
40 |
+
"num_hidden_layers": 28,
|
41 |
+
"num_key_value_heads": 2,
|
42 |
+
"patch_size": 2,
|
43 |
+
"repa_coeff": 0.1,
|
44 |
+
"repa_layers": "2",
|
45 |
+
"repa_shared": false,
|
46 |
+
"rms_norm_eps": 1e-06,
|
47 |
+
"rope_scaling": {
|
48 |
+
"mrope_section": [
|
49 |
+
16,
|
50 |
+
24,
|
51 |
+
24
|
52 |
+
],
|
53 |
+
"rope_type": "default",
|
54 |
+
"type": "default"
|
55 |
+
},
|
56 |
+
"rope_theta": 1000000.0,
|
57 |
+
"sample_size": 128,
|
58 |
+
"sampling_steps": 28,
|
59 |
+
"sliding_window": null,
|
60 |
+
"tie_word_embeddings": true,
|
61 |
+
"torch_dtype": "bfloat16",
|
62 |
+
"transformers_version": "4.47.0",
|
63 |
+
"use_cache": true,
|
64 |
+
"use_repa": false,
|
65 |
+
"use_residual_attn": true,
|
66 |
+
"use_sliding_window": false,
|
67 |
+
"vae_path": "mit-han-lab/dc-ae-f32c32-sana-1.1-diffusers",
|
68 |
+
"video_token_id": 151656,
|
69 |
+
"vision_config": {
|
70 |
+
"hidden_size": 1536,
|
71 |
+
"in_chans": 3,
|
72 |
+
"model_type": "qwen2_vl",
|
73 |
+
"spatial_patch_size": 14
|
74 |
+
},
|
75 |
+
"vision_end_token_id": 151653,
|
76 |
+
"vision_start_token_id": 151652,
|
77 |
+
"vision_token_id": 151654,
|
78 |
+
"vocab_size": 151936
|
79 |
+
}
|
checkpoint-96000/generation_config.json
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_from_model_config": true,
|
3 |
+
"bos_token_id": 151643,
|
4 |
+
"eos_token_id": 151645,
|
5 |
+
"transformers_version": "4.47.0"
|
6 |
+
}
|
checkpoint-96000/model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cb35f8670f8773dce9a7fab20d8f8bd3c15585d97e3c4d7fe4f87519a02de59f
|
3 |
+
size 4539487822
|
checkpoint-96000/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ddd6d30db29fd906ec11222954bf14321d4694d8653b659014c5e754e647b65b
|
3 |
+
size 6632345482
|
checkpoint-96000/rng_state_0.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ea608ce571e3c16d9a15a35803481c8331de16619d9c6404402d5f7aa3e5e430
|
3 |
+
size 15984
|
checkpoint-96000/rng_state_1.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d326fae37874a05c8bd2c5dbbd8ca86c646f126787ed60e1d3c93363c431232d
|
3 |
+
size 15984
|
checkpoint-96000/rng_state_10.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2ea5650ae7baa5c2704842b84735bbdda96821daf5afdabfed3340d5b135da1c
|
3 |
+
size 15997
|
checkpoint-96000/rng_state_11.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:85aca8feeb63884d92f63eee57f423392952ed15ca483bc8964ec53853a5fc29
|
3 |
+
size 15997
|
checkpoint-96000/rng_state_12.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9df76de5d0954fc3891b6a2b5cd0127121da32474a4fb6bdd4aaed4994be0358
|
3 |
+
size 15997
|
checkpoint-96000/rng_state_13.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b691ce87a8ad9c4bb9a51b148d00b72d8e8491424dcd6ef80f95eb69d90550fe
|
3 |
+
size 15997
|
checkpoint-96000/rng_state_14.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ba822e72d6a757eb818848f03ed6e712d60d0953a0925f7df6b083ae8f76fd89
|
3 |
+
size 15997
|
checkpoint-96000/rng_state_15.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e3c17c93cda153f7bd74c77593f5f023329a13dae4a33a1c6d36109587efc135
|
3 |
+
size 15997
|
checkpoint-96000/rng_state_2.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0cc5604b281ef723b0b0d79c5c36a199deb52f9687a45d17376f2ed746026ebf
|
3 |
+
size 15984
|
checkpoint-96000/rng_state_3.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:423499cb5b8be67b2122694fb4742144a1c14a73dad94f434bf85fcfd50c3c8c
|
3 |
+
size 15984
|
checkpoint-96000/rng_state_4.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:de8fb339c42e566e5ff9556f6477de13b47c59d0150c6dcf24cdeb47898884f1
|
3 |
+
size 15984
|
checkpoint-96000/rng_state_5.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5092fe67bfc9ceb81aeed558e9118e3cd9717353c994b7ef566958990920dca8
|
3 |
+
size 15984
|
checkpoint-96000/rng_state_6.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c5965b09ef123aa6614bba5fe421f80ac01648c7642cbe64dd1a96b976df0ddf
|
3 |
+
size 15984
|
checkpoint-96000/rng_state_7.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:79dc4963e4820a34cf71a16d06a35b4995cad9a6d6a32c4b8798b467b5c12b2a
|
3 |
+
size 15984
|
checkpoint-96000/rng_state_8.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:608d6b8e24bf87516a4f8fa2f47bb427dff3666c5f5cb453347d821945460bfe
|
3 |
+
size 15984
|
checkpoint-96000/rng_state_9.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dcae70a389ee8fff81b0d01b9dfffa4d5804b764461240ed8fee36d8630e9cf0
|
3 |
+
size 15984
|
checkpoint-96000/scheduler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6508cf08d65cb093e0e7a03f42a0294aa212e1212fdacb7ee4e24ee00f689a71
|
3 |
+
size 1064
|
checkpoint-96000/trainer_state.json
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:77a26ba232f3efbab997171a0120533dcb11b92fb8d177da942401eb7d35c8b3
|
3 |
+
size 14810877
|
checkpoint-96000/training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:06e6094b82e63a7b07faa7fdb70755edd334ec1d1dc7c89804c0c08312fdff85
|
3 |
+
size 6008
|
checkpoint-97000/config.json
ADDED
@@ -0,0 +1,79 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"ar_steps": 1,
|
3 |
+
"architectures": [
|
4 |
+
"DiffVLMDiffusion"
|
5 |
+
],
|
6 |
+
"attention_dropout": 0.0,
|
7 |
+
"bos_token_id": 151643,
|
8 |
+
"condition_layer": -1,
|
9 |
+
"eos_token_id": 151645,
|
10 |
+
"hidden_act": "silu",
|
11 |
+
"hidden_size": 1536,
|
12 |
+
"image_token_id": 151655,
|
13 |
+
"img_cross_attention_dim": 2048,
|
14 |
+
"img_diffuser_depth": 2,
|
15 |
+
"img_ffn_dim_multiplier": null,
|
16 |
+
"img_hidden_size": 1536,
|
17 |
+
"img_multiple_of": 256,
|
18 |
+
"img_norm_eps": 1e-05,
|
19 |
+
"img_num_attention_heads": 12,
|
20 |
+
"img_num_kv_heads": 12,
|
21 |
+
"img_qk_norm": true,
|
22 |
+
"in_channels": 32,
|
23 |
+
"initializer_range": 0.02,
|
24 |
+
"inject_img_diffuser": false,
|
25 |
+
"input_size": 32,
|
26 |
+
"intermediate_size": 8960,
|
27 |
+
"layer_group_size": 7,
|
28 |
+
"layerwise_start_idx": 0,
|
29 |
+
"lora_alpha": 128,
|
30 |
+
"lora_bias": "none",
|
31 |
+
"lora_dropout": 0.05,
|
32 |
+
"lora_enable": false,
|
33 |
+
"lora_r": 64,
|
34 |
+
"max_position_embeddings": 32768,
|
35 |
+
"max_window_layers": 28,
|
36 |
+
"model_type": "qwen2_vl",
|
37 |
+
"non_linearity": 1,
|
38 |
+
"norm_elementwise_affine": true,
|
39 |
+
"num_attention_heads": 12,
|
40 |
+
"num_hidden_layers": 28,
|
41 |
+
"num_key_value_heads": 2,
|
42 |
+
"patch_size": 2,
|
43 |
+
"repa_coeff": 0.1,
|
44 |
+
"repa_layers": "2",
|
45 |
+
"repa_shared": false,
|
46 |
+
"rms_norm_eps": 1e-06,
|
47 |
+
"rope_scaling": {
|
48 |
+
"mrope_section": [
|
49 |
+
16,
|
50 |
+
24,
|
51 |
+
24
|
52 |
+
],
|
53 |
+
"rope_type": "default",
|
54 |
+
"type": "default"
|
55 |
+
},
|
56 |
+
"rope_theta": 1000000.0,
|
57 |
+
"sample_size": 128,
|
58 |
+
"sampling_steps": 28,
|
59 |
+
"sliding_window": null,
|
60 |
+
"tie_word_embeddings": true,
|
61 |
+
"torch_dtype": "bfloat16",
|
62 |
+
"transformers_version": "4.47.0",
|
63 |
+
"use_cache": true,
|
64 |
+
"use_repa": false,
|
65 |
+
"use_residual_attn": true,
|
66 |
+
"use_sliding_window": false,
|
67 |
+
"vae_path": "mit-han-lab/dc-ae-f32c32-sana-1.1-diffusers",
|
68 |
+
"video_token_id": 151656,
|
69 |
+
"vision_config": {
|
70 |
+
"hidden_size": 1536,
|
71 |
+
"in_chans": 3,
|
72 |
+
"model_type": "qwen2_vl",
|
73 |
+
"spatial_patch_size": 14
|
74 |
+
},
|
75 |
+
"vision_end_token_id": 151653,
|
76 |
+
"vision_start_token_id": 151652,
|
77 |
+
"vision_token_id": 151654,
|
78 |
+
"vocab_size": 151936
|
79 |
+
}
|
checkpoint-97000/generation_config.json
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_from_model_config": true,
|
3 |
+
"bos_token_id": 151643,
|
4 |
+
"eos_token_id": 151645,
|
5 |
+
"transformers_version": "4.47.0"
|
6 |
+
}
|
checkpoint-97000/model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:523ed9be5d1ac5cc0065a74fc357daf157fe994bf5f67084c91887e9891fbc15
|
3 |
+
size 4539487822
|
checkpoint-97000/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a935e46e532effd4b4679e3cce782a369dbcbf7672275c00df5c04af8fce6952
|
3 |
+
size 6632345482
|
checkpoint-97000/rng_state_0.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1c674e086090aafc0825d5ce0957b97e6d8a31f154b3f3bf6a1558050a422abd
|
3 |
+
size 15984
|
checkpoint-97000/rng_state_1.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8da56175e4c6354aa2328b9e7fb87b22a19d65a77aa0530d8c34e9c0ede01175
|
3 |
+
size 15984
|
checkpoint-97000/rng_state_10.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:de28bc031fda4999cf2604e76b01c0984e38f15ad8e01bb7d87577107a981ee7
|
3 |
+
size 15997
|
checkpoint-97000/rng_state_11.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2f316cc706c8a58b737aead43ea50cdc97681578518b6901fdd3c9217e5a75b7
|
3 |
+
size 15997
|
checkpoint-97000/rng_state_12.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:952034ba3fb86c975a0b8d4ef6de9f0b3ee9877a447ef8c0ffc7cfc3031239ed
|
3 |
+
size 15997
|
checkpoint-97000/rng_state_13.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a21aca12e142c3c442eca98cb6264d8831facb07d3e02810539d46cf615bafd2
|
3 |
+
size 15997
|
checkpoint-97000/rng_state_14.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e6111d4b520cab3feb90486dd5035a3f4d1d5f1b8911ebd8a7512ba75b654697
|
3 |
+
size 15997
|
checkpoint-97000/rng_state_15.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9c2b107a34a7d00b1a5aebfd18e820b354dd0505418e38f754a0430b9e2ec230
|
3 |
+
size 15997
|
checkpoint-97000/rng_state_2.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:af88b28e652d1bdfaf91e366a13de77ca4e70e7c9a263dee61c3aeb6a70f4de7
|
3 |
+
size 15984
|
checkpoint-97000/rng_state_3.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:72b5380f2c2d33d5a2f1079f58f8e13ccc001bbdb438a3c98153cbf2446aee22
|
3 |
+
size 15984
|
checkpoint-97000/rng_state_4.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3979f7f2a5fe3612d89209c1d034b1aae41fa5557dcb8a561d728e8f275b0112
|
3 |
+
size 15984
|
checkpoint-97000/rng_state_5.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8d2461f09b464d5f3b76a23f03b7d352850b92c0527fd512926ed1ca62860cf9
|
3 |
+
size 15984
|
checkpoint-97000/rng_state_6.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:29097e1d1492900158e46a5e975a78da43c5ef378ded66065b2b7770d46aea48
|
3 |
+
size 15984
|
checkpoint-97000/rng_state_7.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d079a74022981dba0787c228c1c76a81dbfecf7d165570f82786da7bb6b0f500
|
3 |
+
size 15984
|
checkpoint-97000/rng_state_8.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bd445b9088de180b4165825e82aeb2b327a3c12089f176ceb6d3b314723a0789
|
3 |
+
size 15984
|
checkpoint-97000/rng_state_9.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0fa9275d901ad2ca80513194aaf97a7bef3b6be82e5cd4e290ad2deb6839e209
|
3 |
+
size 15984
|
checkpoint-97000/scheduler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d414d13208710d65a3ed2174c99cea6d8bdf1efa677c19812b1c4eae248023f9
|
3 |
+
size 1064
|
checkpoint-97000/trainer_state.json
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e3a9ed8772220c0a19aa05d37988e905c4ea4bc514ff1ef1709b0f44e1c07d3c
|
3 |
+
size 14965077
|
checkpoint-97000/training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f7bdbe13f8ba23ff4ccf4cbb5e986869ee3ac9a83b6804b3d1d7d56369ddc48b
|
3 |
+
size 6008
|