user
commited on
Commit
·
996aa14
1
Parent(s):
c0c22dc
Add model files
Browse files- .gitattributes +1 -0
- Modelfile +11 -0
- README.md +176 -3
- added_tokens.json +24 -0
- config.json +29 -0
- generation_config.json +14 -0
- merges.txt +0 -0
- model.safetensors +3 -0
- special_tokens_map.json +31 -0
- tokenizer.json +3 -0
- tokenizer_config.json +209 -0
- vocab.json +0 -0
.gitattributes
CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
+
tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
Modelfile
ADDED
@@ -0,0 +1,11 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# ollama modelfile auto-generated by llamafactory
|
2 |
+
|
3 |
+
FROM .
|
4 |
+
|
5 |
+
TEMPLATE """{{ if .System }}System: {{ .System }}<|im_end|>
|
6 |
+
{{ end }}{{ range .Messages }}{{ if eq .Role "user" }}Human: {{ .Content }}<|im_end|>
|
7 |
+
Assistant:{{ else if eq .Role "assistant" }}{{ .Content }}<|im_end|>
|
8 |
+
{{ end }}{{ end }}"""
|
9 |
+
|
10 |
+
PARAMETER stop "<|im_end|>"
|
11 |
+
PARAMETER num_ctx 4096
|
README.md
CHANGED
@@ -1,3 +1,176 @@
|
|
1 |
-
---
|
2 |
-
|
3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
library_name: peft
|
3 |
+
license: apache-2.0
|
4 |
+
base_model: agentlans/Qwen2.5-1.5B-Instruct-drill
|
5 |
+
tags:
|
6 |
+
- llama-factory
|
7 |
+
- generated_from_trainer
|
8 |
+
model-index:
|
9 |
+
- name: Qwen2.5-1.5B-Instruct-Keywords
|
10 |
+
results: []
|
11 |
+
datasets:
|
12 |
+
- agentlans/finewebedu-annotated
|
13 |
+
language:
|
14 |
+
- en
|
15 |
+
---
|
16 |
+
|
17 |
+
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
18 |
+
should probably proofread and complete it, then remove this comment. -->
|
19 |
+
|
20 |
+
# Qwen2.5-1.5B-Instruct-Keywords
|
21 |
+
|
22 |
+
This model is a fine-tuned version of [agentlans/Qwen2.5-1.5B-Instruct-drill](https://huggingface.co/agentlans/Qwen2.5-1.5B-Instruct-drill) on the [agentlans/finewebedu-annotated](https://huggingface.co/datasets/agentlans/finewebedu-annotated) dataset.
|
23 |
+
|
24 |
+
## Model description
|
25 |
+
|
26 |
+
This model generates a list of keywords and key phrases from a piece of text in XML format. The most important keywords are at the front of the list.
|
27 |
+
|
28 |
+
Input format:
|
29 |
+
```text
|
30 |
+
Keywords:
|
31 |
+
{{YOUR_TEXT_HERE}}
|
32 |
+
```
|
33 |
+
|
34 |
+
Output format:
|
35 |
+
```xml
|
36 |
+
<list>
|
37 |
+
<keyword>Keyword 1</keyword>
|
38 |
+
<keyword>Keyword 2</keyword>
|
39 |
+
...
|
40 |
+
</list>
|
41 |
+
```
|
42 |
+
|
43 |
+
### Example
|
44 |
+
|
45 |
+
<details>
|
46 |
+
<summary>Click here for example input prompt</summary>
|
47 |
+
|
48 |
+
[Source](https://www.gartner.com/en/articles/ai-agents)
|
49 |
+
|
50 |
+
```
|
51 |
+
Keywords:
|
52 |
+
|
53 |
+
AI agents can help drive business model innovation
|
54 |
+
|
55 |
+
Though generative AI (GenAI) has captured headlines with its meteoric rise to popularity, AI agents are quickly becoming the “next big thing.” By 2028, Gartner predicts that 33% of enterprise software applications will include agentic AI, up from less than 1% in 2024, with at least 15% of day-to-day work decisions being made autonomously through AI agents.
|
56 |
+
|
57 |
+
But AI agents are also being piloted more strategically to innovate business models, in particular to autonomously address pain points in customer journeys and capitalize on emerging opportunities.
|
58 |
+
Download the Top 10 Strategic Technology Trends for 2025
|
59 |
+
Learn how AI agents and other tech trends align with your digital ambitions — and how to integrate them into your plans.
|
60 |
+
Work Email
|
61 |
+
|
62 |
+
By clicking the "Continue" button, you are agreeing to the Gartner Terms of Use and Privacy Policy.
|
63 |
+
Think critically about integrating AI agents into the business strategy
|
64 |
+
|
65 |
+
Digital leaders cannot afford to overlook the potential business opportunities created by AI agents. However, implementing this technology requires careful consideration of all the potential applications, risks and solutions.
|
66 |
+
Innovate business models with customer journey mapping
|
67 |
+
|
68 |
+
In a market with fierce competition and rapidly emerging technologies — each with their own risks and benefits — digital leaders are under constant pressure to adjust and adapt the business model.
|
69 |
+
|
70 |
+
Business model innovation aims to discover new ways to meet customer needs, differentiate from competitors and increase efficiency and profitability by optimizing or transforming the existing business model.
|
71 |
+
|
72 |
+
Organizations looking to innovate their business models should consider leveraging customer journey mapping. By addressing specific pain points along this journey, they can enhance the experience for customers, streamline operations and uncover new products, services or revenue streams. AI agents can be a key solution for organizations looking to mitigate customer pain points and capitalize on emerging opportunities.
|
73 |
+
Identify business opportunities for AI-agent-powered innovation
|
74 |
+
|
75 |
+
Digital leaders embarking on a journey to capitalize on AI agents should consider six key steps to identify and seize opportunities:
|
76 |
+
|
77 |
+
Define your goals. Figure out what the organization needs to achieve with the business model innovation.
|
78 |
+
|
79 |
+
Map the customer journey and touchpoints. Study how customers interact with the business. Map out their journey and identify key touchpoints.
|
80 |
+
|
81 |
+
Identify the pain points to uncover opportunities. At each touchpoint, flag pain points for improvements or increased efficiency.
|
82 |
+
|
83 |
+
Explore solutions with AI agents. Think about what innovative solutions AI agents could offer to address those pain points and seize opportunities.
|
84 |
+
|
85 |
+
Manage the change. Implementing new AI solutions means change. Strategize on the best way to manage this transition smoothly.
|
86 |
+
|
87 |
+
Evaluate the outcomes. Assess the impact of the AI-agent-powered innovations. Track KPIs and evaluate how well these AI solutions are meeting initial goals. Refine the strategies based on this data.
|
88 |
+
|
89 |
+
Explore all the AI-agent solutions
|
90 |
+
|
91 |
+
AI agents are not a one-size-fits-all solution and come in a variety of shapes and sizes. Organizations should carefully consider each AI-agent solution and select the technology that makes the most sense for their organization and application.
|
92 |
+
|
93 |
+
Types: There are six types of AI agents: reflex agents, goal-based agents, learning-based agents, utility-based agents, hierarchical agents and collaborative agents. Each type is suited to different scenarios and applications.
|
94 |
+
|
95 |
+
Applications: AI agents are ideal for a variety of situations that require automation, decision making and intelligent interaction with the environment.
|
96 |
+
|
97 |
+
Interaction models: AI agents can function with different levels of human involvement, such as human-in-the-loop versus human-out-of-the-loop.
|
98 |
+
|
99 |
+
Multiagent systems: Multiple agents can address complex tasks that individual agents cannot, resulting in more adaptable, scalable and robust solutions.
|
100 |
+
|
101 |
+
Embedded AI techniques: AI agents use AI techniques to sense, navigate and adapt to their environment. Techniques include optimization, natural language processing, and knowledge representation.
|
102 |
+
|
103 |
+
Risks: These agents are designed to act autonomously and proactively within an environment, often learning and adapting as they operate.
|
104 |
+
|
105 |
+
Implement LLM-based AI agents
|
106 |
+
|
107 |
+
Organizations looking for AI agents that are focused on planning, reasoning and processing should consider those that use large language models (LLMs) to drive execution of tasks and processes. LLM-based AI agents can provide a useful and accessible alternative to more established types of AI agents.
|
108 |
+
|
109 |
+
LLM-based AI agents use both programmed and prompted behaviors that require careful design, evaluation and monitoring to ensure the desired outcome. These agents should be built using a modular and composable approach to the software architecture.
|
110 |
+
```
|
111 |
+
</details>
|
112 |
+
|
113 |
+
Output:
|
114 |
+
```xml
|
115 |
+
<list>
|
116 |
+
<keyword>business model innovation</keyword>
|
117 |
+
<keyword>AI agents</keyword>
|
118 |
+
<keyword>genai</keyword>
|
119 |
+
<keyword>customer journey mapping</keyword>
|
120 |
+
<keyword>business model innovation</keyword>
|
121 |
+
<keyword>customer pain points</keyword>
|
122 |
+
<keyword>AI agent solutions</keyword>
|
123 |
+
<keyword>LLM-based AI agents</keyword>
|
124 |
+
<keyword>large language models</keyword>
|
125 |
+
<keyword>business model innovation</keyword>
|
126 |
+
<keyword>AI agent applications</keyword>
|
127 |
+
<keyword>customer experience optimization</keyword>
|
128 |
+
<keyword>business transformation</keyword>
|
129 |
+
<keyword>AI agent interaction models</keyword>
|
130 |
+
<keyword>AI agent types</keyword>
|
131 |
+
<keyword>AI agent risks</keyword>
|
132 |
+
<keyword>AI agent implementation</keyword>
|
133 |
+
<keyword>AI agent evaluation</keyword>
|
134 |
+
<keyword>AI agent optimization</keyword>
|
135 |
+
<keyword>AI agent integration</keyword>
|
136 |
+
</list>
|
137 |
+
```
|
138 |
+
|
139 |
+
## Model description
|
140 |
+
|
141 |
+
More information needed
|
142 |
+
|
143 |
+
## Intended uses & limitations
|
144 |
+
|
145 |
+
More information needed
|
146 |
+
|
147 |
+
## Training and evaluation data
|
148 |
+
|
149 |
+
More information needed
|
150 |
+
|
151 |
+
## Training procedure
|
152 |
+
|
153 |
+
### Training hyperparameters
|
154 |
+
|
155 |
+
The following hyperparameters were used during training:
|
156 |
+
- learning_rate: 5e-05
|
157 |
+
- train_batch_size: 2
|
158 |
+
- eval_batch_size: 8
|
159 |
+
- seed: 42
|
160 |
+
- gradient_accumulation_steps: 8
|
161 |
+
- total_train_batch_size: 16
|
162 |
+
- optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
|
163 |
+
- lr_scheduler_type: cosine
|
164 |
+
- num_epochs: 3.0
|
165 |
+
|
166 |
+
### Training results
|
167 |
+
|
168 |
+
|
169 |
+
|
170 |
+
### Framework versions
|
171 |
+
|
172 |
+
- PEFT 0.15.0
|
173 |
+
- Transformers 4.49.0
|
174 |
+
- Pytorch 2.6.0+cu124
|
175 |
+
- Datasets 3.4.1
|
176 |
+
- Tokenizers 0.21.0
|
added_tokens.json
ADDED
@@ -0,0 +1,24 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"</tool_call>": 151658,
|
3 |
+
"<tool_call>": 151657,
|
4 |
+
"<|box_end|>": 151649,
|
5 |
+
"<|box_start|>": 151648,
|
6 |
+
"<|endoftext|>": 151643,
|
7 |
+
"<|file_sep|>": 151664,
|
8 |
+
"<|fim_middle|>": 151660,
|
9 |
+
"<|fim_pad|>": 151662,
|
10 |
+
"<|fim_prefix|>": 151659,
|
11 |
+
"<|fim_suffix|>": 151661,
|
12 |
+
"<|im_end|>": 151645,
|
13 |
+
"<|im_start|>": 151644,
|
14 |
+
"<|image_pad|>": 151655,
|
15 |
+
"<|object_ref_end|>": 151647,
|
16 |
+
"<|object_ref_start|>": 151646,
|
17 |
+
"<|quad_end|>": 151651,
|
18 |
+
"<|quad_start|>": 151650,
|
19 |
+
"<|repo_name|>": 151663,
|
20 |
+
"<|video_pad|>": 151656,
|
21 |
+
"<|vision_end|>": 151653,
|
22 |
+
"<|vision_pad|>": 151654,
|
23 |
+
"<|vision_start|>": 151652
|
24 |
+
}
|
config.json
ADDED
@@ -0,0 +1,29 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "/drive2/Repo/Qwen2.5-1.5B-Instruct-drill",
|
3 |
+
"architectures": [
|
4 |
+
"Qwen2ForCausalLM"
|
5 |
+
],
|
6 |
+
"attention_dropout": 0.0,
|
7 |
+
"bos_token_id": 151643,
|
8 |
+
"eos_token_id": 151645,
|
9 |
+
"hidden_act": "silu",
|
10 |
+
"hidden_size": 1536,
|
11 |
+
"initializer_range": 0.02,
|
12 |
+
"intermediate_size": 8960,
|
13 |
+
"max_position_embeddings": 32768,
|
14 |
+
"max_window_layers": 21,
|
15 |
+
"model_type": "qwen2",
|
16 |
+
"num_attention_heads": 12,
|
17 |
+
"num_hidden_layers": 28,
|
18 |
+
"num_key_value_heads": 2,
|
19 |
+
"rms_norm_eps": 1e-06,
|
20 |
+
"rope_scaling": null,
|
21 |
+
"rope_theta": 1000000.0,
|
22 |
+
"sliding_window": 32768,
|
23 |
+
"tie_word_embeddings": true,
|
24 |
+
"torch_dtype": "bfloat16",
|
25 |
+
"transformers_version": "4.49.0",
|
26 |
+
"use_cache": true,
|
27 |
+
"use_sliding_window": false,
|
28 |
+
"vocab_size": 151936
|
29 |
+
}
|
generation_config.json
ADDED
@@ -0,0 +1,14 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"bos_token_id": 151643,
|
3 |
+
"do_sample": true,
|
4 |
+
"eos_token_id": [
|
5 |
+
151645,
|
6 |
+
151643
|
7 |
+
],
|
8 |
+
"pad_token_id": 151643,
|
9 |
+
"repetition_penalty": 1.1,
|
10 |
+
"temperature": 0.7,
|
11 |
+
"top_k": 20,
|
12 |
+
"top_p": 0.8,
|
13 |
+
"transformers_version": "4.49.0"
|
14 |
+
}
|
merges.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3178aa4816acc486905db8acbfe53ea80650a28f7cbd57d59425654d9f92069f
|
3 |
+
size 3087467144
|
special_tokens_map.json
ADDED
@@ -0,0 +1,31 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"additional_special_tokens": [
|
3 |
+
"<|im_start|>",
|
4 |
+
"<|im_end|>",
|
5 |
+
"<|object_ref_start|>",
|
6 |
+
"<|object_ref_end|>",
|
7 |
+
"<|box_start|>",
|
8 |
+
"<|box_end|>",
|
9 |
+
"<|quad_start|>",
|
10 |
+
"<|quad_end|>",
|
11 |
+
"<|vision_start|>",
|
12 |
+
"<|vision_end|>",
|
13 |
+
"<|vision_pad|>",
|
14 |
+
"<|image_pad|>",
|
15 |
+
"<|video_pad|>"
|
16 |
+
],
|
17 |
+
"eos_token": {
|
18 |
+
"content": "<|im_end|>",
|
19 |
+
"lstrip": false,
|
20 |
+
"normalized": false,
|
21 |
+
"rstrip": false,
|
22 |
+
"single_word": false
|
23 |
+
},
|
24 |
+
"pad_token": {
|
25 |
+
"content": "<|endoftext|>",
|
26 |
+
"lstrip": false,
|
27 |
+
"normalized": false,
|
28 |
+
"rstrip": false,
|
29 |
+
"single_word": false
|
30 |
+
}
|
31 |
+
}
|
tokenizer.json
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9c5ae00e602b8860cbd784ba82a8aa14e8feecec692e7076590d014d7b7fdafa
|
3 |
+
size 11421896
|
tokenizer_config.json
ADDED
@@ -0,0 +1,209 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"add_bos_token": false,
|
3 |
+
"add_prefix_space": false,
|
4 |
+
"added_tokens_decoder": {
|
5 |
+
"151643": {
|
6 |
+
"content": "<|endoftext|>",
|
7 |
+
"lstrip": false,
|
8 |
+
"normalized": false,
|
9 |
+
"rstrip": false,
|
10 |
+
"single_word": false,
|
11 |
+
"special": true
|
12 |
+
},
|
13 |
+
"151644": {
|
14 |
+
"content": "<|im_start|>",
|
15 |
+
"lstrip": false,
|
16 |
+
"normalized": false,
|
17 |
+
"rstrip": false,
|
18 |
+
"single_word": false,
|
19 |
+
"special": true
|
20 |
+
},
|
21 |
+
"151645": {
|
22 |
+
"content": "<|im_end|>",
|
23 |
+
"lstrip": false,
|
24 |
+
"normalized": false,
|
25 |
+
"rstrip": false,
|
26 |
+
"single_word": false,
|
27 |
+
"special": true
|
28 |
+
},
|
29 |
+
"151646": {
|
30 |
+
"content": "<|object_ref_start|>",
|
31 |
+
"lstrip": false,
|
32 |
+
"normalized": false,
|
33 |
+
"rstrip": false,
|
34 |
+
"single_word": false,
|
35 |
+
"special": true
|
36 |
+
},
|
37 |
+
"151647": {
|
38 |
+
"content": "<|object_ref_end|>",
|
39 |
+
"lstrip": false,
|
40 |
+
"normalized": false,
|
41 |
+
"rstrip": false,
|
42 |
+
"single_word": false,
|
43 |
+
"special": true
|
44 |
+
},
|
45 |
+
"151648": {
|
46 |
+
"content": "<|box_start|>",
|
47 |
+
"lstrip": false,
|
48 |
+
"normalized": false,
|
49 |
+
"rstrip": false,
|
50 |
+
"single_word": false,
|
51 |
+
"special": true
|
52 |
+
},
|
53 |
+
"151649": {
|
54 |
+
"content": "<|box_end|>",
|
55 |
+
"lstrip": false,
|
56 |
+
"normalized": false,
|
57 |
+
"rstrip": false,
|
58 |
+
"single_word": false,
|
59 |
+
"special": true
|
60 |
+
},
|
61 |
+
"151650": {
|
62 |
+
"content": "<|quad_start|>",
|
63 |
+
"lstrip": false,
|
64 |
+
"normalized": false,
|
65 |
+
"rstrip": false,
|
66 |
+
"single_word": false,
|
67 |
+
"special": true
|
68 |
+
},
|
69 |
+
"151651": {
|
70 |
+
"content": "<|quad_end|>",
|
71 |
+
"lstrip": false,
|
72 |
+
"normalized": false,
|
73 |
+
"rstrip": false,
|
74 |
+
"single_word": false,
|
75 |
+
"special": true
|
76 |
+
},
|
77 |
+
"151652": {
|
78 |
+
"content": "<|vision_start|>",
|
79 |
+
"lstrip": false,
|
80 |
+
"normalized": false,
|
81 |
+
"rstrip": false,
|
82 |
+
"single_word": false,
|
83 |
+
"special": true
|
84 |
+
},
|
85 |
+
"151653": {
|
86 |
+
"content": "<|vision_end|>",
|
87 |
+
"lstrip": false,
|
88 |
+
"normalized": false,
|
89 |
+
"rstrip": false,
|
90 |
+
"single_word": false,
|
91 |
+
"special": true
|
92 |
+
},
|
93 |
+
"151654": {
|
94 |
+
"content": "<|vision_pad|>",
|
95 |
+
"lstrip": false,
|
96 |
+
"normalized": false,
|
97 |
+
"rstrip": false,
|
98 |
+
"single_word": false,
|
99 |
+
"special": true
|
100 |
+
},
|
101 |
+
"151655": {
|
102 |
+
"content": "<|image_pad|>",
|
103 |
+
"lstrip": false,
|
104 |
+
"normalized": false,
|
105 |
+
"rstrip": false,
|
106 |
+
"single_word": false,
|
107 |
+
"special": true
|
108 |
+
},
|
109 |
+
"151656": {
|
110 |
+
"content": "<|video_pad|>",
|
111 |
+
"lstrip": false,
|
112 |
+
"normalized": false,
|
113 |
+
"rstrip": false,
|
114 |
+
"single_word": false,
|
115 |
+
"special": true
|
116 |
+
},
|
117 |
+
"151657": {
|
118 |
+
"content": "<tool_call>",
|
119 |
+
"lstrip": false,
|
120 |
+
"normalized": false,
|
121 |
+
"rstrip": false,
|
122 |
+
"single_word": false,
|
123 |
+
"special": false
|
124 |
+
},
|
125 |
+
"151658": {
|
126 |
+
"content": "</tool_call>",
|
127 |
+
"lstrip": false,
|
128 |
+
"normalized": false,
|
129 |
+
"rstrip": false,
|
130 |
+
"single_word": false,
|
131 |
+
"special": false
|
132 |
+
},
|
133 |
+
"151659": {
|
134 |
+
"content": "<|fim_prefix|>",
|
135 |
+
"lstrip": false,
|
136 |
+
"normalized": false,
|
137 |
+
"rstrip": false,
|
138 |
+
"single_word": false,
|
139 |
+
"special": false
|
140 |
+
},
|
141 |
+
"151660": {
|
142 |
+
"content": "<|fim_middle|>",
|
143 |
+
"lstrip": false,
|
144 |
+
"normalized": false,
|
145 |
+
"rstrip": false,
|
146 |
+
"single_word": false,
|
147 |
+
"special": false
|
148 |
+
},
|
149 |
+
"151661": {
|
150 |
+
"content": "<|fim_suffix|>",
|
151 |
+
"lstrip": false,
|
152 |
+
"normalized": false,
|
153 |
+
"rstrip": false,
|
154 |
+
"single_word": false,
|
155 |
+
"special": false
|
156 |
+
},
|
157 |
+
"151662": {
|
158 |
+
"content": "<|fim_pad|>",
|
159 |
+
"lstrip": false,
|
160 |
+
"normalized": false,
|
161 |
+
"rstrip": false,
|
162 |
+
"single_word": false,
|
163 |
+
"special": false
|
164 |
+
},
|
165 |
+
"151663": {
|
166 |
+
"content": "<|repo_name|>",
|
167 |
+
"lstrip": false,
|
168 |
+
"normalized": false,
|
169 |
+
"rstrip": false,
|
170 |
+
"single_word": false,
|
171 |
+
"special": false
|
172 |
+
},
|
173 |
+
"151664": {
|
174 |
+
"content": "<|file_sep|>",
|
175 |
+
"lstrip": false,
|
176 |
+
"normalized": false,
|
177 |
+
"rstrip": false,
|
178 |
+
"single_word": false,
|
179 |
+
"special": false
|
180 |
+
}
|
181 |
+
},
|
182 |
+
"additional_special_tokens": [
|
183 |
+
"<|im_start|>",
|
184 |
+
"<|im_end|>",
|
185 |
+
"<|object_ref_start|>",
|
186 |
+
"<|object_ref_end|>",
|
187 |
+
"<|box_start|>",
|
188 |
+
"<|box_end|>",
|
189 |
+
"<|quad_start|>",
|
190 |
+
"<|quad_end|>",
|
191 |
+
"<|vision_start|>",
|
192 |
+
"<|vision_end|>",
|
193 |
+
"<|vision_pad|>",
|
194 |
+
"<|image_pad|>",
|
195 |
+
"<|video_pad|>"
|
196 |
+
],
|
197 |
+
"bos_token": null,
|
198 |
+
"chat_template": "{% if messages[0]['role'] == 'system' %}{% set loop_messages = messages[1:] %}{% set system_message = messages[0]['content'] %}{% else %}{% set loop_messages = messages %}{% endif %}{% if system_message is defined %}{{ 'System: ' + system_message + '<|im_end|>' + '\n' }}{% endif %}{% for message in loop_messages %}{% set content = message['content'] %}{% if message['role'] == 'user' %}{{ 'Human: ' + content + '<|im_end|>' + '\nAssistant:' }}{% elif message['role'] == 'assistant' %}{{ content + '<|im_end|>' + '\n' }}{% endif %}{% endfor %}",
|
199 |
+
"clean_up_tokenization_spaces": false,
|
200 |
+
"eos_token": "<|im_end|>",
|
201 |
+
"errors": "replace",
|
202 |
+
"extra_special_tokens": {},
|
203 |
+
"model_max_length": 131072,
|
204 |
+
"pad_token": "<|endoftext|>",
|
205 |
+
"padding_side": "left",
|
206 |
+
"split_special_tokens": false,
|
207 |
+
"tokenizer_class": "Qwen2Tokenizer",
|
208 |
+
"unk_token": null
|
209 |
+
}
|
vocab.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|