zhiyang1 commited on
Commit
48eb35f
·
verified ·
1 Parent(s): cc2740c

Upload folder using huggingface_hub

Browse files
Files changed (47) hide show
  1. .gitattributes +2 -0
  2. checkpoint-169000/config.json +79 -0
  3. checkpoint-169000/generation_config.json +6 -0
  4. checkpoint-169000/model.safetensors +3 -0
  5. checkpoint-169000/optimizer.pt +3 -0
  6. checkpoint-169000/rng_state_0.pth +3 -0
  7. checkpoint-169000/rng_state_1.pth +3 -0
  8. checkpoint-169000/rng_state_10.pth +3 -0
  9. checkpoint-169000/rng_state_11.pth +3 -0
  10. checkpoint-169000/rng_state_12.pth +3 -0
  11. checkpoint-169000/rng_state_13.pth +3 -0
  12. checkpoint-169000/rng_state_14.pth +3 -0
  13. checkpoint-169000/rng_state_15.pth +3 -0
  14. checkpoint-169000/rng_state_2.pth +3 -0
  15. checkpoint-169000/rng_state_3.pth +3 -0
  16. checkpoint-169000/rng_state_4.pth +3 -0
  17. checkpoint-169000/rng_state_5.pth +3 -0
  18. checkpoint-169000/rng_state_6.pth +3 -0
  19. checkpoint-169000/rng_state_7.pth +3 -0
  20. checkpoint-169000/rng_state_8.pth +3 -0
  21. checkpoint-169000/rng_state_9.pth +3 -0
  22. checkpoint-169000/scheduler.pt +3 -0
  23. checkpoint-169000/trainer_state.json +3 -0
  24. checkpoint-169000/training_args.bin +3 -0
  25. checkpoint-170000/config.json +79 -0
  26. checkpoint-170000/generation_config.json +6 -0
  27. checkpoint-170000/model.safetensors +3 -0
  28. checkpoint-170000/optimizer.pt +3 -0
  29. checkpoint-170000/rng_state_0.pth +3 -0
  30. checkpoint-170000/rng_state_1.pth +3 -0
  31. checkpoint-170000/rng_state_10.pth +3 -0
  32. checkpoint-170000/rng_state_11.pth +3 -0
  33. checkpoint-170000/rng_state_12.pth +3 -0
  34. checkpoint-170000/rng_state_13.pth +3 -0
  35. checkpoint-170000/rng_state_14.pth +3 -0
  36. checkpoint-170000/rng_state_15.pth +3 -0
  37. checkpoint-170000/rng_state_2.pth +3 -0
  38. checkpoint-170000/rng_state_3.pth +3 -0
  39. checkpoint-170000/rng_state_4.pth +3 -0
  40. checkpoint-170000/rng_state_5.pth +3 -0
  41. checkpoint-170000/rng_state_6.pth +3 -0
  42. checkpoint-170000/rng_state_7.pth +3 -0
  43. checkpoint-170000/rng_state_8.pth +3 -0
  44. checkpoint-170000/rng_state_9.pth +3 -0
  45. checkpoint-170000/scheduler.pt +3 -0
  46. checkpoint-170000/trainer_state.json +3 -0
  47. checkpoint-170000/training_args.bin +3 -0
.gitattributes CHANGED
@@ -69,3 +69,5 @@ checkpoint-159000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
69
  checkpoint-160000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
70
  checkpoint-165000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
71
  checkpoint-166000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
 
 
 
69
  checkpoint-160000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
70
  checkpoint-165000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
71
  checkpoint-166000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
72
+ checkpoint-169000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
73
+ checkpoint-170000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
checkpoint-169000/config.json ADDED
@@ -0,0 +1,79 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "ar_steps": 1,
3
+ "architectures": [
4
+ "DiffVLMBaseline"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 151643,
8
+ "condition_layer": -1,
9
+ "eos_token_id": 151645,
10
+ "hidden_act": "silu",
11
+ "hidden_size": 1536,
12
+ "image_token_id": 151655,
13
+ "img_cross_attention_dim": 2048,
14
+ "img_diffuser_depth": 2,
15
+ "img_ffn_dim_multiplier": null,
16
+ "img_hidden_size": 1536,
17
+ "img_multiple_of": 256,
18
+ "img_norm_eps": 1e-05,
19
+ "img_num_attention_heads": 12,
20
+ "img_num_kv_heads": 12,
21
+ "img_qk_norm": true,
22
+ "in_channels": 32,
23
+ "initializer_range": 0.02,
24
+ "inject_img_diffuser": false,
25
+ "input_size": 32,
26
+ "intermediate_size": 8960,
27
+ "layer_group_size": 7,
28
+ "layerwise_start_idx": 0,
29
+ "lora_alpha": 128,
30
+ "lora_bias": "none",
31
+ "lora_dropout": 0.05,
32
+ "lora_enable": false,
33
+ "lora_r": 64,
34
+ "max_position_embeddings": 32768,
35
+ "max_window_layers": 28,
36
+ "model_type": "qwen2_vl",
37
+ "non_linearity": 1,
38
+ "norm_elementwise_affine": true,
39
+ "num_attention_heads": 12,
40
+ "num_hidden_layers": 28,
41
+ "num_key_value_heads": 2,
42
+ "patch_size": 1,
43
+ "repa_coeff": 0.1,
44
+ "repa_layers": "2",
45
+ "repa_shared": false,
46
+ "rms_norm_eps": 1e-06,
47
+ "rope_scaling": {
48
+ "mrope_section": [
49
+ 16,
50
+ 24,
51
+ 24
52
+ ],
53
+ "rope_type": "default",
54
+ "type": "default"
55
+ },
56
+ "rope_theta": 1000000.0,
57
+ "sample_size": 128,
58
+ "sampling_steps": 28,
59
+ "sliding_window": null,
60
+ "tie_word_embeddings": true,
61
+ "torch_dtype": "bfloat16",
62
+ "transformers_version": "4.47.0",
63
+ "use_cache": true,
64
+ "use_repa": false,
65
+ "use_residual_attn": false,
66
+ "use_sliding_window": false,
67
+ "vae_path": "mit-han-lab/dc-ae-f32c32-in-1.0-diffusers",
68
+ "video_token_id": 151656,
69
+ "vision_config": {
70
+ "hidden_size": 1536,
71
+ "in_chans": 3,
72
+ "model_type": "qwen2_vl",
73
+ "spatial_patch_size": 14
74
+ },
75
+ "vision_end_token_id": 151653,
76
+ "vision_start_token_id": 151652,
77
+ "vision_token_id": 151654,
78
+ "vocab_size": 151936
79
+ }
checkpoint-169000/generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 151643,
4
+ "eos_token_id": 151645,
5
+ "transformers_version": "4.47.0"
6
+ }
checkpoint-169000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ba281ac0401d40141a001182ca9effb079940f500ee68827cd89ff90fbe6e54b
3
+ size 4410723984
checkpoint-169000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:972cbc3ee09da9e57301ae25087d89b82aa944392c8614be66b30a467d82445c
3
+ size 6330255386
checkpoint-169000/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d65b068c445273015aff639e0140b5291c9928c6925cb25f76ea08110792b570
3
+ size 15984
checkpoint-169000/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d39cf304928695686a88543679f5cbd0ad13e29faf0ed22c27a68c1f44f62092
3
+ size 15984
checkpoint-169000/rng_state_10.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9d4084d8aa6ffdc45108c951524de7babd6c1d809d4fc115fccc6fda9290d64f
3
+ size 15997
checkpoint-169000/rng_state_11.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bbabfd154f333f07c7c242694e034579f273f6caea04d1525acd64746f6d4fed
3
+ size 15997
checkpoint-169000/rng_state_12.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:245ad9b4d9e08b9d954c657ad86a1c8f71bf398ea89b32ab920d2db8e7438c10
3
+ size 15997
checkpoint-169000/rng_state_13.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d3b363883c58b576883cdaedd1ac1575fa7efb500a1c151568a73785ca6b6807
3
+ size 15997
checkpoint-169000/rng_state_14.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0bd3dfa319f56fcfa2ce2a2aff7db553493589fd8cf7c2b8e4f5fe3aa2a7d91e
3
+ size 15997
checkpoint-169000/rng_state_15.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9103d24ac96367df0bbc95e545de2f1f7385988e8eab415735017dabf6cdeefc
3
+ size 15997
checkpoint-169000/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:898829f08fbb97b3d609c393906795bc5a354eae3110eaf3fadced55f1c3faa1
3
+ size 15984
checkpoint-169000/rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3e30690d34bea2711084f01fa62b12d8da8f9f73dcc44c75fc0f9a9cd1233b64
3
+ size 15984
checkpoint-169000/rng_state_4.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7d1b8a6176c38a50c222e1d983cd0f7cff62e35408e3d3dfa6b258cfae4cb0c2
3
+ size 15984
checkpoint-169000/rng_state_5.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:605cb6cadb294b327cfda902273750754f98e123eda2e856b145ea5fb27cd274
3
+ size 15984
checkpoint-169000/rng_state_6.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0a639a9b58f319bb85d2ab94a35e66250f7d74b273438294c8b0a2a3b1f42a77
3
+ size 15984
checkpoint-169000/rng_state_7.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61b7255c2bdece5a1124b7f6a52c7b3182745c86a91d2c2ad47ad2a0878f9a89
3
+ size 15984
checkpoint-169000/rng_state_8.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:816907dea154d707f5d2c2f09eef3b5c72c19c2650c5a3eb847bbff371d32dc0
3
+ size 15984
checkpoint-169000/rng_state_9.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:186aba890ab1b2305469af17dc345a88529cd6341544ba4b337a79d8a65d8994
3
+ size 15984
checkpoint-169000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:83de56ae994f16992f17aaac3675983d0052ef337b3aec07d00c998a5092c000
3
+ size 1064
checkpoint-169000/trainer_state.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:41644007e5ee42de699c99e21aa8982f5aee1b1eb1fa79bedb5a85454bb6ff39
3
+ size 25820728
checkpoint-169000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:78baaaf0e002052f7ca00642775f5bb891516dfb2c2aa21690f3e01074eb4505
3
+ size 5944
checkpoint-170000/config.json ADDED
@@ -0,0 +1,79 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "ar_steps": 1,
3
+ "architectures": [
4
+ "DiffVLMBaseline"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 151643,
8
+ "condition_layer": -1,
9
+ "eos_token_id": 151645,
10
+ "hidden_act": "silu",
11
+ "hidden_size": 1536,
12
+ "image_token_id": 151655,
13
+ "img_cross_attention_dim": 2048,
14
+ "img_diffuser_depth": 2,
15
+ "img_ffn_dim_multiplier": null,
16
+ "img_hidden_size": 1536,
17
+ "img_multiple_of": 256,
18
+ "img_norm_eps": 1e-05,
19
+ "img_num_attention_heads": 12,
20
+ "img_num_kv_heads": 12,
21
+ "img_qk_norm": true,
22
+ "in_channels": 32,
23
+ "initializer_range": 0.02,
24
+ "inject_img_diffuser": false,
25
+ "input_size": 32,
26
+ "intermediate_size": 8960,
27
+ "layer_group_size": 7,
28
+ "layerwise_start_idx": 0,
29
+ "lora_alpha": 128,
30
+ "lora_bias": "none",
31
+ "lora_dropout": 0.05,
32
+ "lora_enable": false,
33
+ "lora_r": 64,
34
+ "max_position_embeddings": 32768,
35
+ "max_window_layers": 28,
36
+ "model_type": "qwen2_vl",
37
+ "non_linearity": 1,
38
+ "norm_elementwise_affine": true,
39
+ "num_attention_heads": 12,
40
+ "num_hidden_layers": 28,
41
+ "num_key_value_heads": 2,
42
+ "patch_size": 1,
43
+ "repa_coeff": 0.1,
44
+ "repa_layers": "2",
45
+ "repa_shared": false,
46
+ "rms_norm_eps": 1e-06,
47
+ "rope_scaling": {
48
+ "mrope_section": [
49
+ 16,
50
+ 24,
51
+ 24
52
+ ],
53
+ "rope_type": "default",
54
+ "type": "default"
55
+ },
56
+ "rope_theta": 1000000.0,
57
+ "sample_size": 128,
58
+ "sampling_steps": 28,
59
+ "sliding_window": null,
60
+ "tie_word_embeddings": true,
61
+ "torch_dtype": "bfloat16",
62
+ "transformers_version": "4.47.0",
63
+ "use_cache": true,
64
+ "use_repa": false,
65
+ "use_residual_attn": false,
66
+ "use_sliding_window": false,
67
+ "vae_path": "mit-han-lab/dc-ae-f32c32-in-1.0-diffusers",
68
+ "video_token_id": 151656,
69
+ "vision_config": {
70
+ "hidden_size": 1536,
71
+ "in_chans": 3,
72
+ "model_type": "qwen2_vl",
73
+ "spatial_patch_size": 14
74
+ },
75
+ "vision_end_token_id": 151653,
76
+ "vision_start_token_id": 151652,
77
+ "vision_token_id": 151654,
78
+ "vocab_size": 151936
79
+ }
checkpoint-170000/generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 151643,
4
+ "eos_token_id": 151645,
5
+ "transformers_version": "4.47.0"
6
+ }
checkpoint-170000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:73d4ce449c19f75cdd21c226d4baea75e9155a1db69b6e9d9bd3b13eb25bfa84
3
+ size 4410723984
checkpoint-170000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0f3985e3e233da6969b98ed63c87fe8a823eedaff9e86c8bc989b8b2bd8762cf
3
+ size 6330255386
checkpoint-170000/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:81142f0631ca973174c405c57e5d963212b19cedf8f21109e4e79656d58fedd7
3
+ size 15984
checkpoint-170000/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1dfb43dce8ae40db3177be1793c096ca8258cc4da99b128c9e17146ebd9ec4af
3
+ size 15984
checkpoint-170000/rng_state_10.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9b64a6f13c96f95c4b65ddd463c6adad9c982318754adf497267274c39ce0f2c
3
+ size 15997
checkpoint-170000/rng_state_11.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2c091badd3838b5b68e409d4cfc48661105b86c51634fa0416046a7af7494529
3
+ size 15997
checkpoint-170000/rng_state_12.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fec19c41bb03ba94266dd82dc16cf9a09018c6cc4a8b4c0b0b5dff49c01392c0
3
+ size 15997
checkpoint-170000/rng_state_13.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a98f3544f4a0dbf1a574690618c190eaf41d409d7a7d68d16d28f56ce6bfbe98
3
+ size 15997
checkpoint-170000/rng_state_14.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:254743c6a556ef01af55db3f31c2b7a5a9563ae8da3c39f991a4735c3f715874
3
+ size 15997
checkpoint-170000/rng_state_15.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5e21a71ce27a989e3d7b16af8d45c6cad02af8c5b760c78a52aa9df68700cfed
3
+ size 15997
checkpoint-170000/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:abfedb0c070f079bf38d81d178543dd487e855f4d902c1df532391bd4325b30b
3
+ size 15984
checkpoint-170000/rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:803db2891b0e13b7a5b76bca385b54cd45e7a506bd345ba9b0e9dd8c12e2039e
3
+ size 15984
checkpoint-170000/rng_state_4.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e43d46b1deb17c9c811ba4df651c84836ad4ef3c1c40806fc33d87c806f5feaa
3
+ size 15984
checkpoint-170000/rng_state_5.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7b4f6930049cc6386ad42b383aba0656ddf57770127d4de764e1e96664bac4a8
3
+ size 15984
checkpoint-170000/rng_state_6.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f3f71bbe3d75f7252546a60532734245fecca9492bae678fcb0595f387dbc5d4
3
+ size 15984
checkpoint-170000/rng_state_7.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:05df5a17f56dec711f9c195581f678ea0bbccffa467e08d2b219e482c5d6f978
3
+ size 15984
checkpoint-170000/rng_state_8.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:78b8b1f953141d0c2af64378da72d92b45fe1ad12d6b8e33a7eb6c29cf1766da
3
+ size 15984
checkpoint-170000/rng_state_9.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f6cb12436d41d2d19158ed1a9548bc8ebf672c9ccbb87f3bebe2479c459b51c1
3
+ size 15984
checkpoint-170000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:53418078886f31c0993feeddb7d41f245589575ed7c2ae4e89cfde013fe46589
3
+ size 1064
checkpoint-170000/trainer_state.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:32f993885481ffe6e9a04cb5b0a7a124433272be1d18f38c9797393089bb7aa3
3
+ size 25974108
checkpoint-170000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:78baaaf0e002052f7ca00642775f5bb891516dfb2c2aa21690f3e01074eb4505
3
+ size 5944