zhiyang1 commited on
Commit
3a2f0b4
·
verified ·
1 Parent(s): b9bce1a

Upload folder using huggingface_hub

Browse files
Files changed (47) hide show
  1. .gitattributes +2 -0
  2. checkpoint-159000/config.json +79 -0
  3. checkpoint-159000/generation_config.json +6 -0
  4. checkpoint-159000/model.safetensors +3 -0
  5. checkpoint-159000/optimizer.pt +3 -0
  6. checkpoint-159000/rng_state_0.pth +3 -0
  7. checkpoint-159000/rng_state_1.pth +3 -0
  8. checkpoint-159000/rng_state_10.pth +3 -0
  9. checkpoint-159000/rng_state_11.pth +3 -0
  10. checkpoint-159000/rng_state_12.pth +3 -0
  11. checkpoint-159000/rng_state_13.pth +3 -0
  12. checkpoint-159000/rng_state_14.pth +3 -0
  13. checkpoint-159000/rng_state_15.pth +3 -0
  14. checkpoint-159000/rng_state_2.pth +3 -0
  15. checkpoint-159000/rng_state_3.pth +3 -0
  16. checkpoint-159000/rng_state_4.pth +3 -0
  17. checkpoint-159000/rng_state_5.pth +3 -0
  18. checkpoint-159000/rng_state_6.pth +3 -0
  19. checkpoint-159000/rng_state_7.pth +3 -0
  20. checkpoint-159000/rng_state_8.pth +3 -0
  21. checkpoint-159000/rng_state_9.pth +3 -0
  22. checkpoint-159000/scheduler.pt +3 -0
  23. checkpoint-159000/trainer_state.json +3 -0
  24. checkpoint-159000/training_args.bin +3 -0
  25. checkpoint-160000/config.json +79 -0
  26. checkpoint-160000/generation_config.json +6 -0
  27. checkpoint-160000/model.safetensors +3 -0
  28. checkpoint-160000/optimizer.pt +3 -0
  29. checkpoint-160000/rng_state_0.pth +3 -0
  30. checkpoint-160000/rng_state_1.pth +3 -0
  31. checkpoint-160000/rng_state_10.pth +3 -0
  32. checkpoint-160000/rng_state_11.pth +3 -0
  33. checkpoint-160000/rng_state_12.pth +3 -0
  34. checkpoint-160000/rng_state_13.pth +3 -0
  35. checkpoint-160000/rng_state_14.pth +3 -0
  36. checkpoint-160000/rng_state_15.pth +3 -0
  37. checkpoint-160000/rng_state_2.pth +3 -0
  38. checkpoint-160000/rng_state_3.pth +3 -0
  39. checkpoint-160000/rng_state_4.pth +3 -0
  40. checkpoint-160000/rng_state_5.pth +3 -0
  41. checkpoint-160000/rng_state_6.pth +3 -0
  42. checkpoint-160000/rng_state_7.pth +3 -0
  43. checkpoint-160000/rng_state_8.pth +3 -0
  44. checkpoint-160000/rng_state_9.pth +3 -0
  45. checkpoint-160000/scheduler.pt +3 -0
  46. checkpoint-160000/trainer_state.json +3 -0
  47. checkpoint-160000/training_args.bin +3 -0
.gitattributes CHANGED
@@ -65,3 +65,5 @@ checkpoint-149000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
65
  checkpoint-150000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
66
  checkpoint-154000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
67
  checkpoint-155000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
 
 
 
65
  checkpoint-150000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
66
  checkpoint-154000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
67
  checkpoint-155000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
68
+ checkpoint-159000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
69
+ checkpoint-160000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
checkpoint-159000/config.json ADDED
@@ -0,0 +1,79 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "ar_steps": 1,
3
+ "architectures": [
4
+ "DiffVLMBaseline"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 151643,
8
+ "condition_layer": -1,
9
+ "eos_token_id": 151645,
10
+ "hidden_act": "silu",
11
+ "hidden_size": 1536,
12
+ "image_token_id": 151655,
13
+ "img_cross_attention_dim": 2048,
14
+ "img_diffuser_depth": 2,
15
+ "img_ffn_dim_multiplier": null,
16
+ "img_hidden_size": 1536,
17
+ "img_multiple_of": 256,
18
+ "img_norm_eps": 1e-05,
19
+ "img_num_attention_heads": 12,
20
+ "img_num_kv_heads": 12,
21
+ "img_qk_norm": true,
22
+ "in_channels": 32,
23
+ "initializer_range": 0.02,
24
+ "inject_img_diffuser": false,
25
+ "input_size": 32,
26
+ "intermediate_size": 8960,
27
+ "layer_group_size": 7,
28
+ "layerwise_start_idx": 0,
29
+ "lora_alpha": 128,
30
+ "lora_bias": "none",
31
+ "lora_dropout": 0.05,
32
+ "lora_enable": false,
33
+ "lora_r": 64,
34
+ "max_position_embeddings": 32768,
35
+ "max_window_layers": 28,
36
+ "model_type": "qwen2_vl",
37
+ "non_linearity": 1,
38
+ "norm_elementwise_affine": true,
39
+ "num_attention_heads": 12,
40
+ "num_hidden_layers": 28,
41
+ "num_key_value_heads": 2,
42
+ "patch_size": 1,
43
+ "repa_coeff": 0.1,
44
+ "repa_layers": "2",
45
+ "repa_shared": false,
46
+ "rms_norm_eps": 1e-06,
47
+ "rope_scaling": {
48
+ "mrope_section": [
49
+ 16,
50
+ 24,
51
+ 24
52
+ ],
53
+ "rope_type": "default",
54
+ "type": "default"
55
+ },
56
+ "rope_theta": 1000000.0,
57
+ "sample_size": 128,
58
+ "sampling_steps": 28,
59
+ "sliding_window": null,
60
+ "tie_word_embeddings": true,
61
+ "torch_dtype": "bfloat16",
62
+ "transformers_version": "4.47.0",
63
+ "use_cache": true,
64
+ "use_repa": false,
65
+ "use_residual_attn": false,
66
+ "use_sliding_window": false,
67
+ "vae_path": "mit-han-lab/dc-ae-f32c32-in-1.0-diffusers",
68
+ "video_token_id": 151656,
69
+ "vision_config": {
70
+ "hidden_size": 1536,
71
+ "in_chans": 3,
72
+ "model_type": "qwen2_vl",
73
+ "spatial_patch_size": 14
74
+ },
75
+ "vision_end_token_id": 151653,
76
+ "vision_start_token_id": 151652,
77
+ "vision_token_id": 151654,
78
+ "vocab_size": 151936
79
+ }
checkpoint-159000/generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 151643,
4
+ "eos_token_id": 151645,
5
+ "transformers_version": "4.47.0"
6
+ }
checkpoint-159000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:392819bb7cb12ad79c15e633d39b1c4309cbf1c6921ef95e91a2118977ab4cc0
3
+ size 4410723984
checkpoint-159000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c3ae9a9d0c2a5e5b1e82edbd2cb45fb862476b2a453742402bf68fea6732e887
3
+ size 6330255386
checkpoint-159000/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d66f48d98e047bd9d9ee55d558fdfac36f42134f481edcfeceb8817c132f97e7
3
+ size 15984
checkpoint-159000/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:63901ec65085fc6c249ea66052c90de04dee04f86b31f61e9661bf517959cbfa
3
+ size 15984
checkpoint-159000/rng_state_10.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dae4ac4bc04590bfb52f5e823139ae8c42253a9a40a89ffb97d47d3a0e9eeabf
3
+ size 15997
checkpoint-159000/rng_state_11.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8802a73c774c811f72c1fbafe6d2fd1277576a72807553a2e6b148b8651594a2
3
+ size 15997
checkpoint-159000/rng_state_12.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af24d76dca39aea94538f74014bfa427d502d399c1cf53dfe8524cd27f5dba2f
3
+ size 15997
checkpoint-159000/rng_state_13.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6f550df771abd1a31c55ab60274b1a3464da216471c987b71b121481847550da
3
+ size 15997
checkpoint-159000/rng_state_14.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:245f7963a89a1f0477b6febc6a24649e26064e13d23cacb2757eb65bfb637c0c
3
+ size 15997
checkpoint-159000/rng_state_15.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:152dcb773614b61b6422adabdc819f7bde22ffeec5d25f4d2f69ba23684e1145
3
+ size 15997
checkpoint-159000/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fd745ccc39de953f1c8ea694639ae3d0075acdea1c4e66bf42dcbe0b85c52fed
3
+ size 15984
checkpoint-159000/rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:394563f18bfab77bffe1dadeadf7aae94318d62e954eaa6a5cdf6756fa91e1fb
3
+ size 15984
checkpoint-159000/rng_state_4.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c1f1ad8e46d3b00293d666e42903b68ee559311000fd42f4b0f373d5f511be80
3
+ size 15984
checkpoint-159000/rng_state_5.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:541c42fda12002f15e0452b554c9817e2ef5ad7464d92fc504641a3becc7d171
3
+ size 15984
checkpoint-159000/rng_state_6.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b8e401695a6af4ebfef64ab41eeb2745c60c93d3920135b6f71443ffca71967f
3
+ size 15984
checkpoint-159000/rng_state_7.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:12c52da9f4881fde9afb979d87066712fbf38e3b15f9cf8519aba27243cfc263
3
+ size 15984
checkpoint-159000/rng_state_8.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:849528ad4c916f7cba3865df2ef4e7f39f65a16d091fc7ead11d053279f49088
3
+ size 15984
checkpoint-159000/rng_state_9.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d0887baa6b1087e78d4a2fc12b05a306392e5879ac29a47e79c765c530a37d86
3
+ size 15984
checkpoint-159000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:65f3f10b56514e2b2bf01a2d84b809dfcc93eeda9b052f74ac6d0a6e36e26048
3
+ size 1064
checkpoint-159000/trainer_state.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b4b74164761f560284fb3e6ba78de829fd7533973f59f4f312ad98a61ec00db5
3
+ size 24286882
checkpoint-159000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cbdc2f5e33fa177202c8ac4451a86e0d0fe996e7c7165fc6a0c0279ddb60243b
3
+ size 5944
checkpoint-160000/config.json ADDED
@@ -0,0 +1,79 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "ar_steps": 1,
3
+ "architectures": [
4
+ "DiffVLMBaseline"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 151643,
8
+ "condition_layer": -1,
9
+ "eos_token_id": 151645,
10
+ "hidden_act": "silu",
11
+ "hidden_size": 1536,
12
+ "image_token_id": 151655,
13
+ "img_cross_attention_dim": 2048,
14
+ "img_diffuser_depth": 2,
15
+ "img_ffn_dim_multiplier": null,
16
+ "img_hidden_size": 1536,
17
+ "img_multiple_of": 256,
18
+ "img_norm_eps": 1e-05,
19
+ "img_num_attention_heads": 12,
20
+ "img_num_kv_heads": 12,
21
+ "img_qk_norm": true,
22
+ "in_channels": 32,
23
+ "initializer_range": 0.02,
24
+ "inject_img_diffuser": false,
25
+ "input_size": 32,
26
+ "intermediate_size": 8960,
27
+ "layer_group_size": 7,
28
+ "layerwise_start_idx": 0,
29
+ "lora_alpha": 128,
30
+ "lora_bias": "none",
31
+ "lora_dropout": 0.05,
32
+ "lora_enable": false,
33
+ "lora_r": 64,
34
+ "max_position_embeddings": 32768,
35
+ "max_window_layers": 28,
36
+ "model_type": "qwen2_vl",
37
+ "non_linearity": 1,
38
+ "norm_elementwise_affine": true,
39
+ "num_attention_heads": 12,
40
+ "num_hidden_layers": 28,
41
+ "num_key_value_heads": 2,
42
+ "patch_size": 1,
43
+ "repa_coeff": 0.1,
44
+ "repa_layers": "2",
45
+ "repa_shared": false,
46
+ "rms_norm_eps": 1e-06,
47
+ "rope_scaling": {
48
+ "mrope_section": [
49
+ 16,
50
+ 24,
51
+ 24
52
+ ],
53
+ "rope_type": "default",
54
+ "type": "default"
55
+ },
56
+ "rope_theta": 1000000.0,
57
+ "sample_size": 128,
58
+ "sampling_steps": 28,
59
+ "sliding_window": null,
60
+ "tie_word_embeddings": true,
61
+ "torch_dtype": "bfloat16",
62
+ "transformers_version": "4.47.0",
63
+ "use_cache": true,
64
+ "use_repa": false,
65
+ "use_residual_attn": false,
66
+ "use_sliding_window": false,
67
+ "vae_path": "mit-han-lab/dc-ae-f32c32-in-1.0-diffusers",
68
+ "video_token_id": 151656,
69
+ "vision_config": {
70
+ "hidden_size": 1536,
71
+ "in_chans": 3,
72
+ "model_type": "qwen2_vl",
73
+ "spatial_patch_size": 14
74
+ },
75
+ "vision_end_token_id": 151653,
76
+ "vision_start_token_id": 151652,
77
+ "vision_token_id": 151654,
78
+ "vocab_size": 151936
79
+ }
checkpoint-160000/generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 151643,
4
+ "eos_token_id": 151645,
5
+ "transformers_version": "4.47.0"
6
+ }
checkpoint-160000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:941e26c90add0c06ec5a61871a4899425cb87df3283403c6e7e0f9d14cf83c74
3
+ size 4410723984
checkpoint-160000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bbc289c0f198b8473f44fec20476697eedac85990285a353d80633ea8e9c04dd
3
+ size 6330255386
checkpoint-160000/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5bb33a6c5c0c1ea4be9a85ee3772cdc5ebbf9d19c9299b702f59f71b7d66df69
3
+ size 15984
checkpoint-160000/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:30b5d58c2540cb0abf7bf961431a20996eb6811a28fb473fc4d2c8da05330f72
3
+ size 15984
checkpoint-160000/rng_state_10.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b95ebe5b30e6e9cf1c40c561bea1bf64226ec9d416c8d0317f6aa3e797e3ce9c
3
+ size 15997
checkpoint-160000/rng_state_11.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7a9d360d313f1dbe5d09285e4e9ef45691452015e046dc5da99848a9bc4d10e3
3
+ size 15997
checkpoint-160000/rng_state_12.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:98991b27524ba521bd502ea7d9482b1dd0f092e4fdf23295d62c6eeccbf64e1f
3
+ size 15997
checkpoint-160000/rng_state_13.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c51375b206d8584f05be00e8dffa31617073ca1f1613829a26a4830ea1113fc
3
+ size 15997
checkpoint-160000/rng_state_14.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2fd1afd9a8e5da32da13ed15918a4991a29f10448dfa5c9b00e8c23cd5698b4b
3
+ size 15997
checkpoint-160000/rng_state_15.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:676374a1fccbb3ca9e6a4172a89113cdfc48a1512daac11c590a336b8ed7ec2a
3
+ size 15997
checkpoint-160000/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3c4bb5935566c38f71b34f719978b39eed78a2242850f1f2a93e78cb7bdccf21
3
+ size 15984
checkpoint-160000/rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d3cfaed3cabf7c2411a99db1eba5b791d927307d89e8f65cef8fde9d3c163330
3
+ size 15984
checkpoint-160000/rng_state_4.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:de47aea2d169c93f213791a1f3e01fc4912424510d3de80bc08a93f96c48b033
3
+ size 15984
checkpoint-160000/rng_state_5.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bf16ff3c4a3a6bbb6cda22ecc18240f2a8d1acc3b582fc3d5b6fb7b2aa7e097f
3
+ size 15984
checkpoint-160000/rng_state_6.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d52e5454a975e953f4155276589f72daaffd35418580d60ee52ba37f968f41d4
3
+ size 15984
checkpoint-160000/rng_state_7.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d78bc52f63f7edf1351a84a402ed70170694f2548754df3dd8bb67bf8f1a7fea
3
+ size 15984
checkpoint-160000/rng_state_8.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bb91bfed4df8a698a2e5055088dcf3b34a67d7910de027275b68406a966f7197
3
+ size 15984
checkpoint-160000/rng_state_9.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2a63e55ec3b0e10ebb57f1073e50f105a4d6106f4d33b619924fc5bef5a90f5c
3
+ size 15984
checkpoint-160000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef899c54efc5f24248b766856ac317dedee8e1748a3fbecae1adf8d0a0c2d417
3
+ size 1064
checkpoint-160000/trainer_state.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a4b9c68c0d49fe04a22ec37f4969469865dd6d0c89304cc706f22e6dc43ebfa
3
+ size 24440530
checkpoint-160000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cbdc2f5e33fa177202c8ac4451a86e0d0fe996e7c7165fc6a0c0279ddb60243b
3
+ size 5944