Upload folder using huggingface_hub
Browse files- .gitattributes +1 -0
- README.md +21 -3
- added_tokens.json +24 -0
- chat_template.jinja +7 -0
- config.json +109 -0
- generation_config.json +7 -0
- merges.txt +0 -0
- model-00001-of-00004.safetensors +3 -0
- model-00002-of-00004.safetensors +3 -0
- model-00003-of-00004.safetensors +3 -0
- model-00004-of-00004.safetensors +3 -0
- model.safetensors.index.json +0 -0
- preprocessor_config.json +31 -0
- special_tokens_map.json +38 -0
- tokenizer.json +3 -0
- tokenizer_config.json +226 -0
- vocab.json +0 -0
.gitattributes
CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
+
tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
README.md
CHANGED
@@ -1,3 +1,21 @@
|
|
1 |
-
|
2 |
-
|
3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# HumanOmniV2: From Understanding to Omni-Modal Reasoning with Context
|
2 |
+
- Paper: [Arxiv](https://arxiv.org/abs/2506.21277)
|
3 |
+
|
4 |
+
|
5 |
+
With the rapid evolution of multimodal large language models, the capacity to deeply understand and interpret human intentions has emerged as a critical capability, which demands detailed and thoughtful reasoning. In recent studies, Reinforcement Learning (RL) has demonstrated potential in enhancing the reasoning capabilities of Large Language Models (LLMs). Nonetheless, the challenges associated with adapting RL to multimodal data and formats remain largely unaddressed. In this paper, we identify two issues in existing multimodal reasoning models: insufficient global context understanding and shortcut problems. To tackle these issues, we emphasize the necessity for the model to reason with a clear understanding of the global context within multimodal inputs. This global context understanding can effectively prevent the model from overlooking key multimodal cues and ensure a thorough reasoning process. To ensure the accurate interpretation of multimodal context information and improve complex reasoning capability, we implement context reward and logical reward judged by a large language model, alongside format and accuracy rewards. Our proposed method demonstrates advanced performance across multiple omni-modal benchmarks compared to other open-source omni-modal models.
|
6 |
+
|
7 |
+
#### 🌟 Contributions in HumanOmniV2
|
8 |
+
|
9 |
+
1. We propose that models should summarize the context of multimodal inputs before engaging in the reasoning process. This approach aims to mitigate issues such as skipping crucial multimodal information and context understanding on multimodal inputs.
|
10 |
+
|
11 |
+
2. We curate a human-centric benchmark, IntentBench, for omni-modal evaluation, which requires simultaneously understanding video and audio, the global context, complex social relationships, and careful observation.
|
12 |
+
|
13 |
+
3. Our proposed HumanOmniV2 achieves the best performance across multiple omni-modal benchmarks compared to existing open-source omni-modal methods.
|
14 |
+
|
15 |
+
|
16 |
+
|
17 |
+
|
18 |
+
## 📜 License
|
19 |
+
|
20 |
+
- Our models and code are under the Apache License 2.0.
|
21 |
+
- But our self-collected videos are under [**CC BY-NC-SA 4.0**](https://creativecommons.org/licenses/by-nc-nd/4.0/) license.
|
added_tokens.json
ADDED
@@ -0,0 +1,24 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"</tool_call>": 151658,
|
3 |
+
"<tool_call>": 151657,
|
4 |
+
"<|AUDIO|>": 151646,
|
5 |
+
"<|IMAGE|>": 151655,
|
6 |
+
"<|VIDEO|>": 151656,
|
7 |
+
"<|audio_bos|>": 151647,
|
8 |
+
"<|audio_eos|>": 151648,
|
9 |
+
"<|box_end|>": 151649,
|
10 |
+
"<|endoftext|>": 151643,
|
11 |
+
"<|file_sep|>": 151664,
|
12 |
+
"<|fim_middle|>": 151660,
|
13 |
+
"<|fim_pad|>": 151662,
|
14 |
+
"<|fim_prefix|>": 151659,
|
15 |
+
"<|fim_suffix|>": 151661,
|
16 |
+
"<|im_end|>": 151645,
|
17 |
+
"<|im_start|>": 151644,
|
18 |
+
"<|quad_end|>": 151651,
|
19 |
+
"<|quad_start|>": 151650,
|
20 |
+
"<|repo_name|>": 151663,
|
21 |
+
"<|vision_bos|>": 151652,
|
22 |
+
"<|vision_eos|>": 151653,
|
23 |
+
"<|vision_pad|>": 151654
|
24 |
+
}
|
chat_template.jinja
ADDED
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{% set audio_count = namespace(value=0) %}{% set image_count = namespace(value=0) %}{% set video_count = namespace(value=0) %}{% for message in messages %}{% if loop.first and message['role'] != 'system' %}<|im_start|>system
|
2 |
+
You are a helpful assistant.<|im_end|>
|
3 |
+
{% endif %}<|im_start|>{{ message['role'] }}
|
4 |
+
{% if message['content'] is string %}{{ message['content'] }}<|im_end|>
|
5 |
+
{% else %}{% for content in message['content'] %}{% if content['type'] == 'image' or 'image' in content or 'image_url' in content %}{% set image_count.value = image_count.value + 1 %}{% if add_vision_id %}Picture {{ image_count.value }}: {% endif %}<|vision_bos|><|IMAGE|><|vision_eos|>{% elif content['type'] == 'audio' or 'audio' in content or 'audio_url' in content %}{% set audio_count.value = audio_count.value + 1 %}{% if add_audio_id %}Audio {{ audio_count.value }}: {% endif %}<|audio_bos|><|AUDIO|><|audio_eos|>{% elif content['type'] == 'video' or 'video' in content %}{% set video_count.value = video_count.value + 1 %}{% if add_vision_id %}Video {{ video_count.value }}: {% endif %}<|vision_bos|><|VIDEO|><|vision_eos|>{% elif 'text' in content %}{{ content['text'] }}{% endif %}{% endfor %}<|im_end|>
|
6 |
+
{% endif %}{% endfor %}{% if add_generation_prompt %}<|im_start|>assistant
|
7 |
+
{% endif %}
|
config.json
ADDED
@@ -0,0 +1,109 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"architectures": [
|
3 |
+
"Qwen2_5OmniThinkerForConditionalGeneration"
|
4 |
+
],
|
5 |
+
"audio_config": {
|
6 |
+
"_attn_implementation_autoset": true,
|
7 |
+
"activation_dropout": 0.0,
|
8 |
+
"activation_function": "gelu",
|
9 |
+
"attention_dropout": 0.0,
|
10 |
+
"d_model": 1280,
|
11 |
+
"dropout": 0.0,
|
12 |
+
"encoder_attention_heads": 20,
|
13 |
+
"encoder_ffn_dim": 5120,
|
14 |
+
"encoder_layerdrop": 0.0,
|
15 |
+
"encoder_layers": 32,
|
16 |
+
"init_std": 0.02,
|
17 |
+
"initializer_range": 0.02,
|
18 |
+
"max_source_positions": 1500,
|
19 |
+
"model_type": "qwen2_5_omni_audio_encoder",
|
20 |
+
"n_window": 100,
|
21 |
+
"num_hidden_layers": 32,
|
22 |
+
"num_mel_bins": 128,
|
23 |
+
"output_dim": 3584,
|
24 |
+
"scale_embedding": false,
|
25 |
+
"torch_dtype": "bfloat16"
|
26 |
+
},
|
27 |
+
"audio_end_token_id": 151648,
|
28 |
+
"audio_start_token_id": 151647,
|
29 |
+
"audio_token_index": 151646,
|
30 |
+
"bos_token_id": 151644,
|
31 |
+
"eos_token_id": 151645,
|
32 |
+
"ignore_index": -100,
|
33 |
+
"image_token_index": 151655,
|
34 |
+
"init_std": 0.02,
|
35 |
+
"initializer_range": 0.02,
|
36 |
+
"model_type": "qwen2_5_omni_thinker",
|
37 |
+
"pad_token_id": 151643,
|
38 |
+
"position_id_per_seconds": 25,
|
39 |
+
"seconds_per_chunk": 2,
|
40 |
+
"text_config": {
|
41 |
+
"attention_dropout": 0.0,
|
42 |
+
"hidden_act": "silu",
|
43 |
+
"hidden_size": 3584,
|
44 |
+
"init_std": 0.02,
|
45 |
+
"initializer_range": 0.02,
|
46 |
+
"intermediate_size": 18944,
|
47 |
+
"max_position_embeddings": 32768,
|
48 |
+
"max_window_layers": 28,
|
49 |
+
"model_type": "qwen2_5_omni_text",
|
50 |
+
"num_attention_heads": 28,
|
51 |
+
"num_hidden_layers": 28,
|
52 |
+
"num_key_value_heads": 4,
|
53 |
+
"rms_norm_eps": 1e-06,
|
54 |
+
"rope_scaling": {
|
55 |
+
"mrope_section": [
|
56 |
+
16,
|
57 |
+
24,
|
58 |
+
24
|
59 |
+
],
|
60 |
+
"rope_type": "default",
|
61 |
+
"type": "default"
|
62 |
+
},
|
63 |
+
"rope_theta": 1000000.0,
|
64 |
+
"sliding_window": 32768,
|
65 |
+
"torch_dtype": "bfloat16",
|
66 |
+
"use_cache": true,
|
67 |
+
"use_sliding_window": false,
|
68 |
+
"vocab_size": 152064
|
69 |
+
},
|
70 |
+
"tie_word_embeddings": false,
|
71 |
+
"torch_dtype": "bfloat16",
|
72 |
+
"transformers_version": "4.52.0.dev0",
|
73 |
+
"use_cache": false,
|
74 |
+
"use_mrope": false,
|
75 |
+
"user_token_id": 872,
|
76 |
+
"video_token_index": 151656,
|
77 |
+
"vision_config": {
|
78 |
+
"_attn_implementation_autoset": true,
|
79 |
+
"depth": 32,
|
80 |
+
"embed_dim": 1280,
|
81 |
+
"fullatt_block_indexes": [
|
82 |
+
7,
|
83 |
+
15,
|
84 |
+
23,
|
85 |
+
31
|
86 |
+
],
|
87 |
+
"hidden_act": "silu",
|
88 |
+
"hidden_size": 1280,
|
89 |
+
"in_channels": 3,
|
90 |
+
"in_chans": 3,
|
91 |
+
"init_std": 0.02,
|
92 |
+
"initializer_range": 0.02,
|
93 |
+
"intermediate_size": 3420,
|
94 |
+
"model_type": "qwen2_5_omni_vision_encoder",
|
95 |
+
"num_heads": 16,
|
96 |
+
"out_hidden_size": 3584,
|
97 |
+
"patch_size": 14,
|
98 |
+
"spatial_merge_size": 2,
|
99 |
+
"spatial_patch_size": 14,
|
100 |
+
"temporal_patch_size": 2,
|
101 |
+
"tokens_per_second": 25,
|
102 |
+
"torch_dtype": "bfloat16",
|
103 |
+
"window_size": 112
|
104 |
+
},
|
105 |
+
"vision_end_token_id": 151653,
|
106 |
+
"vision_start_token_id": 151652,
|
107 |
+
"vision_token_id": 151654,
|
108 |
+
"vocab_size": 152064
|
109 |
+
}
|
generation_config.json
ADDED
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_from_model_config": true,
|
3 |
+
"bos_token_id": 151644,
|
4 |
+
"eos_token_id": 151645,
|
5 |
+
"pad_token_id": 151643,
|
6 |
+
"transformers_version": "4.52.0.dev0"
|
7 |
+
}
|
merges.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
model-00001-of-00004.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:711522b72018a296790f4c785dd7f70af25eacf3c60ea97bdd06028371f02da1
|
3 |
+
size 4985047208
|
model-00002-of-00004.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:effdbcee51f63be6d7f1d11594b1f00a3628093c6c61f6c0d0bf2588d47719e6
|
3 |
+
size 4991495784
|
model-00003-of-00004.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4f441a712e8da17b759f0fcf3a994ddb1e10b4f9d7c67bb245c9848f3d8f9d8d
|
3 |
+
size 4991495888
|
model-00004-of-00004.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e9081ce987a51516f1cf7516356214091a9f69ce71246e3db5e46755eb9d9914
|
3 |
+
size 2895739720
|
model.safetensors.index.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
preprocessor_config.json
ADDED
@@ -0,0 +1,31 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"chunk_length": 300,
|
3 |
+
"dither": 0.0,
|
4 |
+
"feature_extractor_type": "WhisperFeatureExtractor",
|
5 |
+
"feature_size": 128,
|
6 |
+
"hop_length": 160,
|
7 |
+
"image_mean": [
|
8 |
+
0.48145466,
|
9 |
+
0.4578275,
|
10 |
+
0.40821073
|
11 |
+
],
|
12 |
+
"image_processor_type": "Qwen2VLImageProcessor",
|
13 |
+
"image_std": [
|
14 |
+
0.26862954,
|
15 |
+
0.26130258,
|
16 |
+
0.27577711
|
17 |
+
],
|
18 |
+
"max_pixels": 12845056,
|
19 |
+
"merge_size": 2,
|
20 |
+
"min_pixels": 3136,
|
21 |
+
"n_fft": 400,
|
22 |
+
"n_samples": 4800000,
|
23 |
+
"nb_max_frames": 30000,
|
24 |
+
"padding_side": "right",
|
25 |
+
"padding_value": 0.0,
|
26 |
+
"patch_size": 14,
|
27 |
+
"processor_class": "Qwen2_5OmniProcessor",
|
28 |
+
"return_attention_mask": true,
|
29 |
+
"sampling_rate": 16000,
|
30 |
+
"temporal_patch_size": 2
|
31 |
+
}
|
special_tokens_map.json
ADDED
@@ -0,0 +1,38 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"additional_special_tokens": [
|
3 |
+
"<|im_start|>",
|
4 |
+
"<|im_end|>",
|
5 |
+
"<|AUDIO|>",
|
6 |
+
"<|audio_bos|>",
|
7 |
+
"<|audio_eos|>",
|
8 |
+
"<|box_end|>",
|
9 |
+
"<|quad_start|>",
|
10 |
+
"<|quad_end|>",
|
11 |
+
"<|vision_bos|>",
|
12 |
+
"<|vision_eos|>",
|
13 |
+
"<|vision_pad|>",
|
14 |
+
"<|IMAGE|>",
|
15 |
+
"<|VIDEO|>"
|
16 |
+
],
|
17 |
+
"audio_bos_token": "<|audio_bos|>",
|
18 |
+
"audio_eos_token": "<|audio_eos|>",
|
19 |
+
"audio_token": "<|AUDIO|>",
|
20 |
+
"eos_token": {
|
21 |
+
"content": "<|im_end|>",
|
22 |
+
"lstrip": false,
|
23 |
+
"normalized": false,
|
24 |
+
"rstrip": false,
|
25 |
+
"single_word": false
|
26 |
+
},
|
27 |
+
"image_token": "<|IMAGE|>",
|
28 |
+
"pad_token": {
|
29 |
+
"content": "<|endoftext|>",
|
30 |
+
"lstrip": false,
|
31 |
+
"normalized": false,
|
32 |
+
"rstrip": false,
|
33 |
+
"single_word": false
|
34 |
+
},
|
35 |
+
"video_token": "<|VIDEO|>",
|
36 |
+
"vision_bos_token": "<|vision_bos|>",
|
37 |
+
"vision_eos_token": "<|vision_eos|>"
|
38 |
+
}
|
tokenizer.json
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f9711e245647e88538786834977dc8afb51172e879ee661352c587cf01efd6b0
|
3 |
+
size 11422037
|
tokenizer_config.json
ADDED
@@ -0,0 +1,226 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"add_prefix_space": false,
|
3 |
+
"added_tokens_decoder": {
|
4 |
+
"151643": {
|
5 |
+
"content": "<|endoftext|>",
|
6 |
+
"lstrip": false,
|
7 |
+
"normalized": false,
|
8 |
+
"rstrip": false,
|
9 |
+
"single_word": false,
|
10 |
+
"special": true
|
11 |
+
},
|
12 |
+
"151644": {
|
13 |
+
"content": "<|im_start|>",
|
14 |
+
"lstrip": false,
|
15 |
+
"normalized": false,
|
16 |
+
"rstrip": false,
|
17 |
+
"single_word": false,
|
18 |
+
"special": true
|
19 |
+
},
|
20 |
+
"151645": {
|
21 |
+
"content": "<|im_end|>",
|
22 |
+
"lstrip": false,
|
23 |
+
"normalized": false,
|
24 |
+
"rstrip": false,
|
25 |
+
"single_word": false,
|
26 |
+
"special": true
|
27 |
+
},
|
28 |
+
"151646": {
|
29 |
+
"content": "<|AUDIO|>",
|
30 |
+
"lstrip": false,
|
31 |
+
"normalized": false,
|
32 |
+
"rstrip": false,
|
33 |
+
"single_word": false,
|
34 |
+
"special": true
|
35 |
+
},
|
36 |
+
"151647": {
|
37 |
+
"content": "<|audio_bos|>",
|
38 |
+
"lstrip": false,
|
39 |
+
"normalized": false,
|
40 |
+
"rstrip": false,
|
41 |
+
"single_word": false,
|
42 |
+
"special": true
|
43 |
+
},
|
44 |
+
"151648": {
|
45 |
+
"content": "<|audio_eos|>",
|
46 |
+
"lstrip": false,
|
47 |
+
"normalized": false,
|
48 |
+
"rstrip": false,
|
49 |
+
"single_word": false,
|
50 |
+
"special": true
|
51 |
+
},
|
52 |
+
"151649": {
|
53 |
+
"content": "<|box_end|>",
|
54 |
+
"lstrip": false,
|
55 |
+
"normalized": false,
|
56 |
+
"rstrip": false,
|
57 |
+
"single_word": false,
|
58 |
+
"special": true
|
59 |
+
},
|
60 |
+
"151650": {
|
61 |
+
"content": "<|quad_start|>",
|
62 |
+
"lstrip": false,
|
63 |
+
"normalized": false,
|
64 |
+
"rstrip": false,
|
65 |
+
"single_word": false,
|
66 |
+
"special": true
|
67 |
+
},
|
68 |
+
"151651": {
|
69 |
+
"content": "<|quad_end|>",
|
70 |
+
"lstrip": false,
|
71 |
+
"normalized": false,
|
72 |
+
"rstrip": false,
|
73 |
+
"single_word": false,
|
74 |
+
"special": true
|
75 |
+
},
|
76 |
+
"151652": {
|
77 |
+
"content": "<|vision_bos|>",
|
78 |
+
"lstrip": false,
|
79 |
+
"normalized": false,
|
80 |
+
"rstrip": false,
|
81 |
+
"single_word": false,
|
82 |
+
"special": true
|
83 |
+
},
|
84 |
+
"151653": {
|
85 |
+
"content": "<|vision_eos|>",
|
86 |
+
"lstrip": false,
|
87 |
+
"normalized": false,
|
88 |
+
"rstrip": false,
|
89 |
+
"single_word": false,
|
90 |
+
"special": true
|
91 |
+
},
|
92 |
+
"151654": {
|
93 |
+
"content": "<|vision_pad|>",
|
94 |
+
"lstrip": false,
|
95 |
+
"normalized": false,
|
96 |
+
"rstrip": false,
|
97 |
+
"single_word": false,
|
98 |
+
"special": true
|
99 |
+
},
|
100 |
+
"151655": {
|
101 |
+
"content": "<|IMAGE|>",
|
102 |
+
"lstrip": false,
|
103 |
+
"normalized": false,
|
104 |
+
"rstrip": false,
|
105 |
+
"single_word": false,
|
106 |
+
"special": true
|
107 |
+
},
|
108 |
+
"151656": {
|
109 |
+
"content": "<|VIDEO|>",
|
110 |
+
"lstrip": false,
|
111 |
+
"normalized": false,
|
112 |
+
"rstrip": false,
|
113 |
+
"single_word": false,
|
114 |
+
"special": true
|
115 |
+
},
|
116 |
+
"151657": {
|
117 |
+
"content": "<tool_call>",
|
118 |
+
"lstrip": false,
|
119 |
+
"normalized": false,
|
120 |
+
"rstrip": false,
|
121 |
+
"single_word": false,
|
122 |
+
"special": false
|
123 |
+
},
|
124 |
+
"151658": {
|
125 |
+
"content": "</tool_call>",
|
126 |
+
"lstrip": false,
|
127 |
+
"normalized": false,
|
128 |
+
"rstrip": false,
|
129 |
+
"single_word": false,
|
130 |
+
"special": false
|
131 |
+
},
|
132 |
+
"151659": {
|
133 |
+
"content": "<|fim_prefix|>",
|
134 |
+
"lstrip": false,
|
135 |
+
"normalized": false,
|
136 |
+
"rstrip": false,
|
137 |
+
"single_word": false,
|
138 |
+
"special": false
|
139 |
+
},
|
140 |
+
"151660": {
|
141 |
+
"content": "<|fim_middle|>",
|
142 |
+
"lstrip": false,
|
143 |
+
"normalized": false,
|
144 |
+
"rstrip": false,
|
145 |
+
"single_word": false,
|
146 |
+
"special": false
|
147 |
+
},
|
148 |
+
"151661": {
|
149 |
+
"content": "<|fim_suffix|>",
|
150 |
+
"lstrip": false,
|
151 |
+
"normalized": false,
|
152 |
+
"rstrip": false,
|
153 |
+
"single_word": false,
|
154 |
+
"special": false
|
155 |
+
},
|
156 |
+
"151662": {
|
157 |
+
"content": "<|fim_pad|>",
|
158 |
+
"lstrip": false,
|
159 |
+
"normalized": false,
|
160 |
+
"rstrip": false,
|
161 |
+
"single_word": false,
|
162 |
+
"special": false
|
163 |
+
},
|
164 |
+
"151663": {
|
165 |
+
"content": "<|repo_name|>",
|
166 |
+
"lstrip": false,
|
167 |
+
"normalized": false,
|
168 |
+
"rstrip": false,
|
169 |
+
"single_word": false,
|
170 |
+
"special": false
|
171 |
+
},
|
172 |
+
"151664": {
|
173 |
+
"content": "<|file_sep|>",
|
174 |
+
"lstrip": false,
|
175 |
+
"normalized": false,
|
176 |
+
"rstrip": false,
|
177 |
+
"single_word": false,
|
178 |
+
"special": false
|
179 |
+
}
|
180 |
+
},
|
181 |
+
"additional_special_tokens": [
|
182 |
+
"<|im_start|>",
|
183 |
+
"<|im_end|>",
|
184 |
+
"<|AUDIO|>",
|
185 |
+
"<|audio_bos|>",
|
186 |
+
"<|audio_eos|>",
|
187 |
+
"<|box_end|>",
|
188 |
+
"<|quad_start|>",
|
189 |
+
"<|quad_end|>",
|
190 |
+
"<|vision_bos|>",
|
191 |
+
"<|vision_eos|>",
|
192 |
+
"<|vision_pad|>",
|
193 |
+
"<|IMAGE|>",
|
194 |
+
"<|VIDEO|>"
|
195 |
+
],
|
196 |
+
"audio_bos_token": "<|audio_bos|>",
|
197 |
+
"audio_eos_token": "<|audio_eos|>",
|
198 |
+
"audio_token": "<|AUDIO|>",
|
199 |
+
"bos_token": null,
|
200 |
+
"clean_up_tokenization_spaces": false,
|
201 |
+
"eos_token": "<|im_end|>",
|
202 |
+
"errors": "replace",
|
203 |
+
"extra_special_tokens": {
|
204 |
+
"audio_bos_token": "<|audio_bos|>",
|
205 |
+
"audio_eos_token": "<|audio_eos|>",
|
206 |
+
"audio_token": "<|AUDIO|>",
|
207 |
+
"image_token": "<|IMAGE|>",
|
208 |
+
"video_token": "<|VIDEO|>",
|
209 |
+
"vision_bos_token": "<|vision_bos|>",
|
210 |
+
"vision_eos_token": "<|vision_eos|>"
|
211 |
+
},
|
212 |
+
"image_token": "<|IMAGE|>",
|
213 |
+
"max_length": null,
|
214 |
+
"model_max_length": 32768,
|
215 |
+
"pad_to_multiple_of": null,
|
216 |
+
"pad_token": "<|endoftext|>",
|
217 |
+
"pad_token_type_id": 0,
|
218 |
+
"padding_side": "left",
|
219 |
+
"processor_class": "Qwen2_5OmniProcessor",
|
220 |
+
"split_special_tokens": false,
|
221 |
+
"tokenizer_class": "Qwen2Tokenizer",
|
222 |
+
"unk_token": null,
|
223 |
+
"video_token": "<|VIDEO|>",
|
224 |
+
"vision_bos_token": "<|vision_bos|>",
|
225 |
+
"vision_eos_token": "<|vision_eos|>"
|
226 |
+
}
|
vocab.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|