zhiyang1 commited on
Commit
d8c21c6
·
verified ·
1 Parent(s): 871945a

Upload folder using huggingface_hub

Browse files
Files changed (47) hide show
  1. .gitattributes +2 -0
  2. checkpoint-179000/config.json +79 -0
  3. checkpoint-179000/generation_config.json +6 -0
  4. checkpoint-179000/model.safetensors +3 -0
  5. checkpoint-179000/optimizer.pt +3 -0
  6. checkpoint-179000/rng_state_0.pth +3 -0
  7. checkpoint-179000/rng_state_1.pth +3 -0
  8. checkpoint-179000/rng_state_10.pth +3 -0
  9. checkpoint-179000/rng_state_11.pth +3 -0
  10. checkpoint-179000/rng_state_12.pth +3 -0
  11. checkpoint-179000/rng_state_13.pth +3 -0
  12. checkpoint-179000/rng_state_14.pth +3 -0
  13. checkpoint-179000/rng_state_15.pth +3 -0
  14. checkpoint-179000/rng_state_2.pth +3 -0
  15. checkpoint-179000/rng_state_3.pth +3 -0
  16. checkpoint-179000/rng_state_4.pth +3 -0
  17. checkpoint-179000/rng_state_5.pth +3 -0
  18. checkpoint-179000/rng_state_6.pth +3 -0
  19. checkpoint-179000/rng_state_7.pth +3 -0
  20. checkpoint-179000/rng_state_8.pth +3 -0
  21. checkpoint-179000/rng_state_9.pth +3 -0
  22. checkpoint-179000/scheduler.pt +3 -0
  23. checkpoint-179000/trainer_state.json +3 -0
  24. checkpoint-179000/training_args.bin +3 -0
  25. checkpoint-180000/config.json +79 -0
  26. checkpoint-180000/generation_config.json +6 -0
  27. checkpoint-180000/model.safetensors +3 -0
  28. checkpoint-180000/optimizer.pt +3 -0
  29. checkpoint-180000/rng_state_0.pth +3 -0
  30. checkpoint-180000/rng_state_1.pth +3 -0
  31. checkpoint-180000/rng_state_10.pth +3 -0
  32. checkpoint-180000/rng_state_11.pth +3 -0
  33. checkpoint-180000/rng_state_12.pth +3 -0
  34. checkpoint-180000/rng_state_13.pth +3 -0
  35. checkpoint-180000/rng_state_14.pth +3 -0
  36. checkpoint-180000/rng_state_15.pth +3 -0
  37. checkpoint-180000/rng_state_2.pth +3 -0
  38. checkpoint-180000/rng_state_3.pth +3 -0
  39. checkpoint-180000/rng_state_4.pth +3 -0
  40. checkpoint-180000/rng_state_5.pth +3 -0
  41. checkpoint-180000/rng_state_6.pth +3 -0
  42. checkpoint-180000/rng_state_7.pth +3 -0
  43. checkpoint-180000/rng_state_8.pth +3 -0
  44. checkpoint-180000/rng_state_9.pth +3 -0
  45. checkpoint-180000/scheduler.pt +3 -0
  46. checkpoint-180000/trainer_state.json +3 -0
  47. checkpoint-180000/training_args.bin +3 -0
.gitattributes CHANGED
@@ -73,3 +73,5 @@ checkpoint-169000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
73
  checkpoint-170000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
74
  checkpoint-177000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
75
  checkpoint-178000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
 
 
 
73
  checkpoint-170000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
74
  checkpoint-177000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
75
  checkpoint-178000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
76
+ checkpoint-179000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
77
+ checkpoint-180000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
checkpoint-179000/config.json ADDED
@@ -0,0 +1,79 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "ar_steps": 1,
3
+ "architectures": [
4
+ "DiffVLMBaseline"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 151643,
8
+ "condition_layer": -1,
9
+ "eos_token_id": 151645,
10
+ "hidden_act": "silu",
11
+ "hidden_size": 1536,
12
+ "image_token_id": 151655,
13
+ "img_cross_attention_dim": 2048,
14
+ "img_diffuser_depth": 2,
15
+ "img_ffn_dim_multiplier": null,
16
+ "img_hidden_size": 1536,
17
+ "img_multiple_of": 256,
18
+ "img_norm_eps": 1e-05,
19
+ "img_num_attention_heads": 12,
20
+ "img_num_kv_heads": 12,
21
+ "img_qk_norm": true,
22
+ "in_channels": 32,
23
+ "initializer_range": 0.02,
24
+ "inject_img_diffuser": false,
25
+ "input_size": 32,
26
+ "intermediate_size": 8960,
27
+ "layer_group_size": 7,
28
+ "layerwise_start_idx": 0,
29
+ "lora_alpha": 128,
30
+ "lora_bias": "none",
31
+ "lora_dropout": 0.05,
32
+ "lora_enable": false,
33
+ "lora_r": 64,
34
+ "max_position_embeddings": 32768,
35
+ "max_window_layers": 28,
36
+ "model_type": "qwen2_vl",
37
+ "non_linearity": 1,
38
+ "norm_elementwise_affine": true,
39
+ "num_attention_heads": 12,
40
+ "num_hidden_layers": 28,
41
+ "num_key_value_heads": 2,
42
+ "patch_size": 1,
43
+ "repa_coeff": 0.1,
44
+ "repa_layers": "2",
45
+ "repa_shared": false,
46
+ "rms_norm_eps": 1e-06,
47
+ "rope_scaling": {
48
+ "mrope_section": [
49
+ 16,
50
+ 24,
51
+ 24
52
+ ],
53
+ "rope_type": "default",
54
+ "type": "default"
55
+ },
56
+ "rope_theta": 1000000.0,
57
+ "sample_size": 128,
58
+ "sampling_steps": 28,
59
+ "sliding_window": null,
60
+ "tie_word_embeddings": true,
61
+ "torch_dtype": "bfloat16",
62
+ "transformers_version": "4.47.0",
63
+ "use_cache": true,
64
+ "use_repa": false,
65
+ "use_residual_attn": false,
66
+ "use_sliding_window": false,
67
+ "vae_path": "mit-han-lab/dc-ae-f32c32-in-1.0-diffusers",
68
+ "video_token_id": 151656,
69
+ "vision_config": {
70
+ "hidden_size": 1536,
71
+ "in_chans": 3,
72
+ "model_type": "qwen2_vl",
73
+ "spatial_patch_size": 14
74
+ },
75
+ "vision_end_token_id": 151653,
76
+ "vision_start_token_id": 151652,
77
+ "vision_token_id": 151654,
78
+ "vocab_size": 151936
79
+ }
checkpoint-179000/generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 151643,
4
+ "eos_token_id": 151645,
5
+ "transformers_version": "4.47.0"
6
+ }
checkpoint-179000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8cb4cc3919ce9bf5503ea62c857046dd46d42e8e65ec663bbb9712d3d25be164
3
+ size 4410723984
checkpoint-179000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:06ac17a4c93ab52cdf58edab3cbbd38ad674abf0f40fc2bf23e2d799cf8c4931
3
+ size 6330255386
checkpoint-179000/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:78760004bafcc73ec6fcf2af193094b5cd4e608e13e61442d57366f1f4da7a89
3
+ size 15984
checkpoint-179000/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6fe034dd3820f335e171e86f6355a3f26737bfbb156ec7a37160d2f4f460f878
3
+ size 15984
checkpoint-179000/rng_state_10.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b7e1bdb0f189e550f4f266ab4ba6ed8668e8650543e6dd16801f6f04a759fc38
3
+ size 15997
checkpoint-179000/rng_state_11.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c6b16f50d9d77ed1efbb402b1159e32f25e487cf9a012c8729ddf9fb656e3379
3
+ size 15997
checkpoint-179000/rng_state_12.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e8e94ac8ab9112e1c0fd3e7f839fe0a60773f491735e514a1fc3033fe7831a17
3
+ size 15997
checkpoint-179000/rng_state_13.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a5528a93f016bc1f5cc1f3c9abd7699ab7030c53142a2c1e0af0ab2846ed7f97
3
+ size 15997
checkpoint-179000/rng_state_14.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:73f012e58c4b6dbb0269941e09927b0612d46b7ac1cc0236d155b04fa2283475
3
+ size 15997
checkpoint-179000/rng_state_15.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:48c56cae0c4f472f62ede5dfa41a89824234a48b36f71549e4d5165628864404
3
+ size 15997
checkpoint-179000/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c96c70a66e5d783942375148799a7ba41a7637962920fdc6b152aec88a3336cb
3
+ size 15984
checkpoint-179000/rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c38c8427c0c2659ad0109b21b19a9207251422ed5e3b495cbb45bf57a0da8d2c
3
+ size 15984
checkpoint-179000/rng_state_4.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3653f56141301de5c32d113a72ebbead4f7703e913a5286c31f78785931d64dd
3
+ size 15984
checkpoint-179000/rng_state_5.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f90c8a019d1316ac4c55467d1959c9264894896ecef6a36eaa98af213f0d1a00
3
+ size 15984
checkpoint-179000/rng_state_6.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1ed5033ff914c22a7f6c2fa172fad826cc1939ce38be6dcac9c9920b61787a0e
3
+ size 15984
checkpoint-179000/rng_state_7.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f27aef31a6c78f80cb6c31de9062c420adb21c627303e99a71ef116c8cbb4ef
3
+ size 15984
checkpoint-179000/rng_state_8.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ced3284428d6893c0152d122aa3f4c48ff21ffa0567036dbd5b88227866fd4eb
3
+ size 15984
checkpoint-179000/rng_state_9.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:89557be00e080bda908572457c01b7b0aa4655969512bb3e9bcfdda949f0ff19
3
+ size 15984
checkpoint-179000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e71d2f64115ce74e7a25f255a1a83a4f34da5f6833a6b7236b021bad9ea2e9f8
3
+ size 1064
checkpoint-179000/trainer_state.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:db66a7340fb5c827810cda685e847e9b528ad28b58c11e04b03c20fa387299df
3
+ size 27354252
checkpoint-179000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5e655117e0b364a441f6b722a259c297f7b5ccaf2bbd399241653a57f1f11bf2
3
+ size 5944
checkpoint-180000/config.json ADDED
@@ -0,0 +1,79 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "ar_steps": 1,
3
+ "architectures": [
4
+ "DiffVLMBaseline"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 151643,
8
+ "condition_layer": -1,
9
+ "eos_token_id": 151645,
10
+ "hidden_act": "silu",
11
+ "hidden_size": 1536,
12
+ "image_token_id": 151655,
13
+ "img_cross_attention_dim": 2048,
14
+ "img_diffuser_depth": 2,
15
+ "img_ffn_dim_multiplier": null,
16
+ "img_hidden_size": 1536,
17
+ "img_multiple_of": 256,
18
+ "img_norm_eps": 1e-05,
19
+ "img_num_attention_heads": 12,
20
+ "img_num_kv_heads": 12,
21
+ "img_qk_norm": true,
22
+ "in_channels": 32,
23
+ "initializer_range": 0.02,
24
+ "inject_img_diffuser": false,
25
+ "input_size": 32,
26
+ "intermediate_size": 8960,
27
+ "layer_group_size": 7,
28
+ "layerwise_start_idx": 0,
29
+ "lora_alpha": 128,
30
+ "lora_bias": "none",
31
+ "lora_dropout": 0.05,
32
+ "lora_enable": false,
33
+ "lora_r": 64,
34
+ "max_position_embeddings": 32768,
35
+ "max_window_layers": 28,
36
+ "model_type": "qwen2_vl",
37
+ "non_linearity": 1,
38
+ "norm_elementwise_affine": true,
39
+ "num_attention_heads": 12,
40
+ "num_hidden_layers": 28,
41
+ "num_key_value_heads": 2,
42
+ "patch_size": 1,
43
+ "repa_coeff": 0.1,
44
+ "repa_layers": "2",
45
+ "repa_shared": false,
46
+ "rms_norm_eps": 1e-06,
47
+ "rope_scaling": {
48
+ "mrope_section": [
49
+ 16,
50
+ 24,
51
+ 24
52
+ ],
53
+ "rope_type": "default",
54
+ "type": "default"
55
+ },
56
+ "rope_theta": 1000000.0,
57
+ "sample_size": 128,
58
+ "sampling_steps": 28,
59
+ "sliding_window": null,
60
+ "tie_word_embeddings": true,
61
+ "torch_dtype": "bfloat16",
62
+ "transformers_version": "4.47.0",
63
+ "use_cache": true,
64
+ "use_repa": false,
65
+ "use_residual_attn": false,
66
+ "use_sliding_window": false,
67
+ "vae_path": "mit-han-lab/dc-ae-f32c32-in-1.0-diffusers",
68
+ "video_token_id": 151656,
69
+ "vision_config": {
70
+ "hidden_size": 1536,
71
+ "in_chans": 3,
72
+ "model_type": "qwen2_vl",
73
+ "spatial_patch_size": 14
74
+ },
75
+ "vision_end_token_id": 151653,
76
+ "vision_start_token_id": 151652,
77
+ "vision_token_id": 151654,
78
+ "vocab_size": 151936
79
+ }
checkpoint-180000/generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 151643,
4
+ "eos_token_id": 151645,
5
+ "transformers_version": "4.47.0"
6
+ }
checkpoint-180000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b30d0696eb719ed46a048d9e7d5da4df084ffe7f4428b4aec79dfdd98d7e636e
3
+ size 4410723984
checkpoint-180000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:369ff6ecc6ff15acc0ecc9fea933d9c17be59a47f559c15defd4d75769e385f1
3
+ size 6330255386
checkpoint-180000/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9d1f5edf71e59cead7b9cffe852410a60fb63b64d721106f1d8960bda717ff7e
3
+ size 15984
checkpoint-180000/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5e24daf584eca81d100d11103fff1ab689ca1f2a66c9f7330ee9a327f5f78b7a
3
+ size 15984
checkpoint-180000/rng_state_10.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ec5df7b2c67fe14424f88bff3cb564d6a2fed42c22faf35389a7294a637c350a
3
+ size 15997
checkpoint-180000/rng_state_11.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e371021dbfa0d26f29d034a04b14b037c32453d7fb028d92b8c42ad14459975
3
+ size 15997
checkpoint-180000/rng_state_12.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3acfec454898f94fbc4351d4fa24f82c91adc227ae9ff06f7262d9e3439ecfcd
3
+ size 15997
checkpoint-180000/rng_state_13.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:08aed2930016c865803de2ef272deec0a51e6ab24c23d8f5503f53af691baf1b
3
+ size 15997
checkpoint-180000/rng_state_14.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:607fa9e53801f43b946c046bbd660e62524f5059a2f5e157893914539724d917
3
+ size 15997
checkpoint-180000/rng_state_15.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:97c99551d6a22b3113d42838c853a2c6f83a4eae51ca857d75aff7074bcd6b7b
3
+ size 15997
checkpoint-180000/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:55cbe28052bcc019effa5d2deadc0db73493a11ba08c9b85660398dab1f65c0d
3
+ size 15984
checkpoint-180000/rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3328057414685471318c9db113d6e865bb55fb7dd3cb96d656334dcbcb4d91b0
3
+ size 15984
checkpoint-180000/rng_state_4.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f4c943b9c6b536e907ac4e31bf18a16e9f05a74b2c5ce37aae05f5c3ecdd86a6
3
+ size 15984
checkpoint-180000/rng_state_5.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:182362d14f3a455edb110b6792b81c360c75e66ee00e48a26906c5ba692b2155
3
+ size 15984
checkpoint-180000/rng_state_6.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0f984d1f23a4782454c11c3e7078e710cb68274373be6b5932d43fe4d616564c
3
+ size 15984
checkpoint-180000/rng_state_7.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4b4e53eca45cd813823166d763109893941b2de14d31b8c95180a0b63379a181
3
+ size 15984
checkpoint-180000/rng_state_8.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:671a0f4404da65862093491fbf97f89b55ff23ba445f7d58ab1c9f0b4dc87cd5
3
+ size 15984
checkpoint-180000/rng_state_9.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4e5dd9038f5332a4899d6b127df726759204e883a4236b2aab6d22c172bebec9
3
+ size 15984
checkpoint-180000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:28776a9c3de5a32cbf30f46040d1cd3eed38c068e90db3841c12b18e0a5bf5f0
3
+ size 1064
checkpoint-180000/trainer_state.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a605492971901c5cfcf725661c8b7afc066ffc43ff7b492aaf30944b982cfbb3
3
+ size 27507590
checkpoint-180000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5e655117e0b364a441f6b722a259c297f7b5ccaf2bbd399241653a57f1f11bf2
3
+ size 5944