zhiyang1 commited on
Commit
c512e29
·
verified ·
1 Parent(s): 6cd0528

Upload folder using huggingface_hub

Browse files
Files changed (47) hide show
  1. .gitattributes +2 -0
  2. checkpoint-209000/config.json +79 -0
  3. checkpoint-209000/generation_config.json +6 -0
  4. checkpoint-209000/model.safetensors +3 -0
  5. checkpoint-209000/optimizer.pt +3 -0
  6. checkpoint-209000/rng_state_0.pth +3 -0
  7. checkpoint-209000/rng_state_1.pth +3 -0
  8. checkpoint-209000/rng_state_10.pth +3 -0
  9. checkpoint-209000/rng_state_11.pth +3 -0
  10. checkpoint-209000/rng_state_12.pth +3 -0
  11. checkpoint-209000/rng_state_13.pth +3 -0
  12. checkpoint-209000/rng_state_14.pth +3 -0
  13. checkpoint-209000/rng_state_15.pth +3 -0
  14. checkpoint-209000/rng_state_2.pth +3 -0
  15. checkpoint-209000/rng_state_3.pth +3 -0
  16. checkpoint-209000/rng_state_4.pth +3 -0
  17. checkpoint-209000/rng_state_5.pth +3 -0
  18. checkpoint-209000/rng_state_6.pth +3 -0
  19. checkpoint-209000/rng_state_7.pth +3 -0
  20. checkpoint-209000/rng_state_8.pth +3 -0
  21. checkpoint-209000/rng_state_9.pth +3 -0
  22. checkpoint-209000/scheduler.pt +3 -0
  23. checkpoint-209000/trainer_state.json +3 -0
  24. checkpoint-209000/training_args.bin +3 -0
  25. checkpoint-210000/config.json +79 -0
  26. checkpoint-210000/generation_config.json +6 -0
  27. checkpoint-210000/model.safetensors +3 -0
  28. checkpoint-210000/optimizer.pt +3 -0
  29. checkpoint-210000/rng_state_0.pth +3 -0
  30. checkpoint-210000/rng_state_1.pth +3 -0
  31. checkpoint-210000/rng_state_10.pth +3 -0
  32. checkpoint-210000/rng_state_11.pth +3 -0
  33. checkpoint-210000/rng_state_12.pth +3 -0
  34. checkpoint-210000/rng_state_13.pth +3 -0
  35. checkpoint-210000/rng_state_14.pth +3 -0
  36. checkpoint-210000/rng_state_15.pth +3 -0
  37. checkpoint-210000/rng_state_2.pth +3 -0
  38. checkpoint-210000/rng_state_3.pth +3 -0
  39. checkpoint-210000/rng_state_4.pth +3 -0
  40. checkpoint-210000/rng_state_5.pth +3 -0
  41. checkpoint-210000/rng_state_6.pth +3 -0
  42. checkpoint-210000/rng_state_7.pth +3 -0
  43. checkpoint-210000/rng_state_8.pth +3 -0
  44. checkpoint-210000/rng_state_9.pth +3 -0
  45. checkpoint-210000/scheduler.pt +3 -0
  46. checkpoint-210000/trainer_state.json +3 -0
  47. checkpoint-210000/training_args.bin +3 -0
.gitattributes CHANGED
@@ -81,3 +81,5 @@ checkpoint-199000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
81
  checkpoint-200000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
82
  checkpoint-201000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
83
  checkpoint-202000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
 
 
 
81
  checkpoint-200000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
82
  checkpoint-201000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
83
  checkpoint-202000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
84
+ checkpoint-209000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
85
+ checkpoint-210000/trainer_state.json filter=lfs diff=lfs merge=lfs -text
checkpoint-209000/config.json ADDED
@@ -0,0 +1,79 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "ar_steps": 1,
3
+ "architectures": [
4
+ "DiffVLMBaseline"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 151643,
8
+ "condition_layer": -1,
9
+ "eos_token_id": 151645,
10
+ "hidden_act": "silu",
11
+ "hidden_size": 1536,
12
+ "image_token_id": 151655,
13
+ "img_cross_attention_dim": 2048,
14
+ "img_diffuser_depth": 2,
15
+ "img_ffn_dim_multiplier": null,
16
+ "img_hidden_size": 1536,
17
+ "img_multiple_of": 256,
18
+ "img_norm_eps": 1e-05,
19
+ "img_num_attention_heads": 12,
20
+ "img_num_kv_heads": 12,
21
+ "img_qk_norm": true,
22
+ "in_channels": 32,
23
+ "initializer_range": 0.02,
24
+ "inject_img_diffuser": false,
25
+ "input_size": 32,
26
+ "intermediate_size": 8960,
27
+ "layer_group_size": 7,
28
+ "layerwise_start_idx": 0,
29
+ "lora_alpha": 128,
30
+ "lora_bias": "none",
31
+ "lora_dropout": 0.05,
32
+ "lora_enable": false,
33
+ "lora_r": 64,
34
+ "max_position_embeddings": 32768,
35
+ "max_window_layers": 28,
36
+ "model_type": "qwen2_vl",
37
+ "non_linearity": 1,
38
+ "norm_elementwise_affine": true,
39
+ "num_attention_heads": 12,
40
+ "num_hidden_layers": 28,
41
+ "num_key_value_heads": 2,
42
+ "patch_size": 1,
43
+ "repa_coeff": 0.1,
44
+ "repa_layers": "2",
45
+ "repa_shared": false,
46
+ "rms_norm_eps": 1e-06,
47
+ "rope_scaling": {
48
+ "mrope_section": [
49
+ 16,
50
+ 24,
51
+ 24
52
+ ],
53
+ "rope_type": "default",
54
+ "type": "default"
55
+ },
56
+ "rope_theta": 1000000.0,
57
+ "sample_size": 128,
58
+ "sampling_steps": 28,
59
+ "sliding_window": null,
60
+ "tie_word_embeddings": true,
61
+ "torch_dtype": "bfloat16",
62
+ "transformers_version": "4.47.0",
63
+ "use_cache": true,
64
+ "use_repa": false,
65
+ "use_residual_attn": false,
66
+ "use_sliding_window": false,
67
+ "vae_path": "mit-han-lab/dc-ae-f32c32-in-1.0-diffusers",
68
+ "video_token_id": 151656,
69
+ "vision_config": {
70
+ "hidden_size": 1536,
71
+ "in_chans": 3,
72
+ "model_type": "qwen2_vl",
73
+ "spatial_patch_size": 14
74
+ },
75
+ "vision_end_token_id": 151653,
76
+ "vision_start_token_id": 151652,
77
+ "vision_token_id": 151654,
78
+ "vocab_size": 151936
79
+ }
checkpoint-209000/generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 151643,
4
+ "eos_token_id": 151645,
5
+ "transformers_version": "4.47.0"
6
+ }
checkpoint-209000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a920db154a449370d1180d5053132706ab67cda0f84d86812ad9e24a414e55ac
3
+ size 4410723984
checkpoint-209000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cea6e2d3acefafbe0bdf35e80c7391a5a14523fac0486608101659eedb2cdebb
3
+ size 6330255386
checkpoint-209000/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aa835494b6f2d8ef502fe82b24b64b5e1f208b1dff98c5957f0ac0391872f5bf
3
+ size 15984
checkpoint-209000/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:33f46fcb1aef2d8921dc4763bc634c500b44224235f780c6232d4d8cf3cdc7df
3
+ size 15984
checkpoint-209000/rng_state_10.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d0bfa12e23f1686c744b0b8ec6f33be60394a315d04082b79c90a907875379a2
3
+ size 15997
checkpoint-209000/rng_state_11.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5611c4fee62d37d64e1f6787a0c3d8959bd7c1a2e23c65df559bfdf57b3f8ed0
3
+ size 15997
checkpoint-209000/rng_state_12.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e8c27cea49a60e21b033c8e5eba5f6c9ec0edb794fb544d1fcb79fb231fcee7e
3
+ size 15997
checkpoint-209000/rng_state_13.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b4b0218718203612029434206ed2ec33815131f6c931d28aeb27b19ad2b7860e
3
+ size 15997
checkpoint-209000/rng_state_14.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dccc073a36216ec383d7d73e033f05a0668a7e37ab06b4b6e46bdbb9da136526
3
+ size 15997
checkpoint-209000/rng_state_15.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ab847b519ebe89e38850d7d40f081cf1a86290fe5c898677692d4086af8dc1ac
3
+ size 15997
checkpoint-209000/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3438fd67a90d9b6e434ef9879db7dea548e5291b38a71634323db3e4f2c9c96a
3
+ size 15984
checkpoint-209000/rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cc70d06a1c6a994cd88dfe970099eb3868c5ced11b86191517af9f0071d4a54d
3
+ size 15984
checkpoint-209000/rng_state_4.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c5c2a1b9e18681e5a84a3f76eba507c67e515fdaba966187d57ee8fc50ae628e
3
+ size 15984
checkpoint-209000/rng_state_5.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8bd976fd34cb0cda44f35bd0746a20dc87d371ad0f855f555cec11506f3b68f8
3
+ size 15984
checkpoint-209000/rng_state_6.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0f01614ce294233dce276e9a8b3939c9cadbbc1104e0eb3f496bc0569c4b069f
3
+ size 15984
checkpoint-209000/rng_state_7.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:58c540da3e2f50884781099c3ae081c272db1a94c3e20ff6393ab04596e683f9
3
+ size 15984
checkpoint-209000/rng_state_8.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a8f3dbd98b67039522beb9cd0e97144531ce30df74abe5657ab100e27d51f88e
3
+ size 15984
checkpoint-209000/rng_state_9.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d007a1c0f13e0c2ab7811ef7a02bf633c534f64851a3fcec6ecb9d04b4136690
3
+ size 15984
checkpoint-209000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0dd09743ab140a4f3c180ed7e6d9ed3337fd98fa4d210fbf2ce94151002fdebe
3
+ size 1064
checkpoint-209000/trainer_state.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:434a8e3447150fa158334b52f4a361df72f8a7ee7c139394a570760fa20775f0
3
+ size 31954419
checkpoint-209000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d354d6ad462de64b8718f9432522cc71b3d2dd3d61aad54c27b688902f9b142c
3
+ size 5944
checkpoint-210000/config.json ADDED
@@ -0,0 +1,79 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "ar_steps": 1,
3
+ "architectures": [
4
+ "DiffVLMBaseline"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 151643,
8
+ "condition_layer": -1,
9
+ "eos_token_id": 151645,
10
+ "hidden_act": "silu",
11
+ "hidden_size": 1536,
12
+ "image_token_id": 151655,
13
+ "img_cross_attention_dim": 2048,
14
+ "img_diffuser_depth": 2,
15
+ "img_ffn_dim_multiplier": null,
16
+ "img_hidden_size": 1536,
17
+ "img_multiple_of": 256,
18
+ "img_norm_eps": 1e-05,
19
+ "img_num_attention_heads": 12,
20
+ "img_num_kv_heads": 12,
21
+ "img_qk_norm": true,
22
+ "in_channels": 32,
23
+ "initializer_range": 0.02,
24
+ "inject_img_diffuser": false,
25
+ "input_size": 32,
26
+ "intermediate_size": 8960,
27
+ "layer_group_size": 7,
28
+ "layerwise_start_idx": 0,
29
+ "lora_alpha": 128,
30
+ "lora_bias": "none",
31
+ "lora_dropout": 0.05,
32
+ "lora_enable": false,
33
+ "lora_r": 64,
34
+ "max_position_embeddings": 32768,
35
+ "max_window_layers": 28,
36
+ "model_type": "qwen2_vl",
37
+ "non_linearity": 1,
38
+ "norm_elementwise_affine": true,
39
+ "num_attention_heads": 12,
40
+ "num_hidden_layers": 28,
41
+ "num_key_value_heads": 2,
42
+ "patch_size": 1,
43
+ "repa_coeff": 0.1,
44
+ "repa_layers": "2",
45
+ "repa_shared": false,
46
+ "rms_norm_eps": 1e-06,
47
+ "rope_scaling": {
48
+ "mrope_section": [
49
+ 16,
50
+ 24,
51
+ 24
52
+ ],
53
+ "rope_type": "default",
54
+ "type": "default"
55
+ },
56
+ "rope_theta": 1000000.0,
57
+ "sample_size": 128,
58
+ "sampling_steps": 28,
59
+ "sliding_window": null,
60
+ "tie_word_embeddings": true,
61
+ "torch_dtype": "bfloat16",
62
+ "transformers_version": "4.47.0",
63
+ "use_cache": true,
64
+ "use_repa": false,
65
+ "use_residual_attn": false,
66
+ "use_sliding_window": false,
67
+ "vae_path": "mit-han-lab/dc-ae-f32c32-in-1.0-diffusers",
68
+ "video_token_id": 151656,
69
+ "vision_config": {
70
+ "hidden_size": 1536,
71
+ "in_chans": 3,
72
+ "model_type": "qwen2_vl",
73
+ "spatial_patch_size": 14
74
+ },
75
+ "vision_end_token_id": 151653,
76
+ "vision_start_token_id": 151652,
77
+ "vision_token_id": 151654,
78
+ "vocab_size": 151936
79
+ }
checkpoint-210000/generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 151643,
4
+ "eos_token_id": 151645,
5
+ "transformers_version": "4.47.0"
6
+ }
checkpoint-210000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:27c8902981dd3f44f2b632be108b5498167b0eb28e8d0bb9a7bed1ca7bc93490
3
+ size 4410723984
checkpoint-210000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:42094de8eb340b38e74f6211e1bb63cab26341b50473b946ca47dd203df5e5a4
3
+ size 6330255386
checkpoint-210000/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3fce1276c976e8dac991b0f96a921ab4d6cbe4dde454c8d99ee4d31049fef645
3
+ size 15984
checkpoint-210000/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5622490620a15a5e76c58f123e2ccf22902a314455c92986bebb6ee46f81715f
3
+ size 15984
checkpoint-210000/rng_state_10.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f09adc42810be0f0bd588667faa499d6d449569118c11c06774ef38f7c3d169
3
+ size 15997
checkpoint-210000/rng_state_11.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3d9735b0c82db0c36ea367cf4498681d13321d51fdd834149b55d5d96da861f1
3
+ size 15997
checkpoint-210000/rng_state_12.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:748dfc496a452cbbe9882f925c497a4e07ec93227b9b3721ec45a9bea4f9f2f9
3
+ size 15997
checkpoint-210000/rng_state_13.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:126cb617fb4b7ae0c55826eb762566e9280feb174013ae40d34da63c3aa37379
3
+ size 15997
checkpoint-210000/rng_state_14.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:52daa6aba0beadd0fd5fb7e4cd720cb694fa2034f63a20c3096771bb9574640c
3
+ size 15997
checkpoint-210000/rng_state_15.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:499231a5abcf7efd5c3f608bf39e2a675112f1c4d92fc9fab6f9b5edc2ee6f86
3
+ size 15997
checkpoint-210000/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2679a6a6cafe9bcb21e3462c757b79272456f50ec090fe0289245bd74422ea50
3
+ size 15984
checkpoint-210000/rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6ddb532b67905c6d43141d262b7122f630c84a7dc3831f9ba560d3155f6fd607
3
+ size 15984
checkpoint-210000/rng_state_4.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a981231724c89a8492987c698e889ba8203adb1ddab32bb239c3e23559b857b
3
+ size 15984
checkpoint-210000/rng_state_5.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ab564359cdbf37925fe4c4b46b33159f63d8c7d07787bdd1c70d720f6790ed87
3
+ size 15984
checkpoint-210000/rng_state_6.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:942bf9227ca7e80f74a40567d278081a542ca529b8b2cad48414f044903e3349
3
+ size 15984
checkpoint-210000/rng_state_7.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a9cb81b9ea1a2174f7f4f4c8633ad9d1039a29e4ee3456d3746069d7091e796f
3
+ size 15984
checkpoint-210000/rng_state_8.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0fd158b103344cdf42b88ee5a8310ef02f7deabec8a1b4c5df88b5a65bf08876
3
+ size 15984
checkpoint-210000/rng_state_9.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7524377c180915d0d8d7bf18c937642c561a0e99db9e7edf417d793c79660b8d
3
+ size 15984
checkpoint-210000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a79bba9bb724889141d1efbc2bb27deffe931f0394f23daf4d46c563104990a2
3
+ size 1064
checkpoint-210000/trainer_state.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:50ea14c5e7126e9ef718116049d1a0c67fd3343138a4f0fe12e82474ce486280
3
+ size 32107817
checkpoint-210000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d354d6ad462de64b8718f9432522cc71b3d2dd3d61aad54c27b688902f9b142c
3
+ size 5944