Files changed (2) hide show
  1. README.md +40 -108
  2. tokenizer_config.json +1 -1
README.md CHANGED
@@ -7,35 +7,30 @@ language:
7
  tags:
8
  - falcon3
9
  base_model: tiiuae/Falcon3-7B-Base
10
- license: other
11
- license_name: falcon-llm-license
12
  license_link: https://falconllm.tii.ae/falcon-terms-and-conditions.html
13
- library_name: transformers
14
  ---
15
 
16
- <div align="center">
17
- <img src="https://huggingface.co/datasets/tiiuae/documentation-images/resolve/main/general/falco3-logo.png" alt="drawing" width="500"/>
18
- </div>
19
-
20
  # Falcon3-7B-Instruct
21
 
22
- **Falcon3** family of Open Foundation Models is a set of pretrained and instruct LLMs ranging from 1B to 10B.
23
 
24
- This repository contains the **Falcon3-7B-Instruct**. It achieves state of art results (at the time of release) on reasoning, language understanding, instruction following, code and mathematics tasks.
25
- Falcon3-7B-Instruct supports 4 languages (english, french, spanish, portuguese) and a context length up to 32K.
26
 
27
  ## Model Details
28
  - Architecture
29
- - Transformer based causal decoder only architecture
30
  - 28 decoder blocks
31
- - Grouped query attention (GQA) for faster inference: 12 query heads and 4 key value heads
32
  - Wider head dimension: 256
33
  - High RoPE value to support long context understanding: 1000042
34
  - Uses SwiGLU and RMSNorm
35
  - 32K context length
36
  - 131K vocab size
37
- - Pretrained on 14 Teratokens of datasets comprising of web, code, STEM, high quality and mutlilingual data using 1024 H100 GPU chips
38
- - Postrained on 1.2 million samples of STEM, conversations, code, safety and function call data
39
  - Supports EN, FR, ES, PT
40
  - Developed by [Technology Innovation Institute](https://www.tii.ae)
41
  - License: TII Falcon-LLM License 2.0
@@ -91,66 +86,7 @@ print(response)
91
  <br>
92
 
93
  ## Benchmarks
94
- We report the official HuggingFace leaderboard normalized evaluations [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard) in the following table.
95
- <table border="1" style="width: 100%; text-align: center; border-collapse: collapse;">
96
- <colgroup>
97
- <col style="width: 10%;">
98
- <col style="width: 7%;">
99
- <col style="width: 7%;">
100
- <col style="background-color: rgba(80, 15, 213, 0.5); width: 7%;">
101
- </colgroup>
102
- <thead>
103
- <tr>
104
- <th>Benchmark</th>
105
- <th>Llama-3.1-8B-Instruct</th>
106
- <th>Qwen2.5-7B-Instruct</th>
107
- <th>Falcon3-7B-Instruct</th>
108
- </tr>
109
- </thead>
110
- <tbody>
111
- <tr>
112
- <td>IFEval</td>
113
- <td><b>78.56</b></td>
114
- <td>75.85</td>
115
- <td>76.12</td>
116
- </tr>
117
- <tr>
118
- <td>BBH (3-shot)</td>
119
- <td>29.89</td>
120
- <td>34.89</td>
121
- <td><b>37.92</b></td>
122
- </tr>
123
- <tr>
124
- <td>MATH Lvl-5 (4-shot)</td>
125
- <td>19.34</td>
126
- <td>0.00</td>
127
- <td><b>31.87</b></td>
128
- </tr>
129
- <tr>
130
- <td>GPQA (0-shot)</td>
131
- <td>2.35</td>
132
- <td>5.48</td>
133
- <td><b>8.05</b></td>
134
- </tr>
135
- <tr>
136
- <td>MUSR (0-shot)</td>
137
- <td>8.41</td>
138
- <td>8.45</td>
139
- <td><b>21.17</b></td>
140
- </tr>
141
- <tr>
142
- <td>MMLU-PRO (5-shot)</td>
143
- <td>30.68</td>
144
- <td><b>36.52</b></td>
145
- <td>34.30</td>
146
- </tr>
147
- </tbody>
148
- </table>
149
-
150
- Also, we report in the following table our internal pipeline benchmarks.
151
- - We use [lm-evaluation harness](https://github.com/EleutherAI/lm-evaluation-harness).
152
- - We report **raw scores** obtained by applying chat template and fewshot_as_multiturn.
153
- - We use same batch-size across all models.
154
 
155
  <table border="1" style="width: 100%; text-align: center; border-collapse: collapse;">
156
  <colgroup>
@@ -173,15 +109,15 @@ Also, we report in the following table our internal pipeline benchmarks.
173
  <tr>
174
  <td rowspan="3">General</td>
175
  <td>MMLU (5-shot)</td>
176
- <td>68.2</td>
177
- <td><b>73.5</b></td>
178
- <td>70.5</td>
179
  </tr>
180
  <tr>
181
  <td>MMLU-PRO (5-shot)</td>
182
- <td>36.4</td>
183
- <td><b>43.1</b></td>
184
- <td>40.7</td>
185
  </tr>
186
  <tr>
187
  <td>IFEval</td>
@@ -192,34 +128,34 @@ Also, we report in the following table our internal pipeline benchmarks.
192
  <tr>
193
  <td rowspan="3">Math</td>
194
  <td>GSM8K (5-shot)</td>
195
- <td><b>82.6</b></td>
196
- <td>72.0</td>
197
- <td>81.4</td>
198
  </tr>
199
  <tr>
200
  <td>GSM8K (8-shot, COT)</td>
201
- <td><b>85.4</b></td>
202
- <td>76.6</td>
203
- <td>79.7</td>
204
  </tr>
205
  <tr>
206
  <td>MATH Lvl-5 (4-shot)</td>
207
- <td>15.4</td>
208
- <td>-</td>
209
- <td><b>29.4</b></td>
210
  </tr>
211
  <tr>
212
  <td rowspan="5">Reasoning</td>
213
  <td>Arc Challenge (25-shot)</td>
214
- <td>58.6</td>
215
- <td>57.8</td>
216
- <td><b>62.6</b></td>
217
  </tr>
218
  <tr>
219
  <td>GPQA (0-shot)</td>
220
- <td><b>33.5</b></td>
221
- <td>32</td>
222
  <td>31.9</td>
 
223
  </tr>
224
  <tr>
225
  <td>GPQA (0-shot, COT)</td>
@@ -230,13 +166,13 @@ Also, we report in the following table our internal pipeline benchmarks.
230
  <tr>
231
  <td>MUSR (0-shot)</td>
232
  <td>38.6</td>
233
- <td>41</td>
234
  <td><b>46.4</b></td>
235
  </tr>
236
  <tr>
237
  <td>BBH (3-shot)</td>
238
- <td>48.6</td>
239
- <td><b>54.1</b></td>
240
  <td>52.4</td>
241
  </tr>
242
  <tr>
@@ -267,30 +203,26 @@ Also, we report in the following table our internal pipeline benchmarks.
267
  <tr>
268
  <td rowspan="2">Instructions following</td>
269
  <td>MT-Bench (avg)</td>
270
- <td>7.9</td>
271
- <td><b>8.5</b></td>
272
- <td>8.4</td>
273
  </tr>
274
  <tr>
275
- <td>Alpaca (WC)</td>
276
- <td>26.6</td>
277
  <td><b>31.5</b></td>
278
- <td>26.1</td>
279
  </tr>
280
  <tr>
281
  <td>Tool use</td>
282
  <td>BFCL AST (avg)</td>
283
  <td>90.6</td>
284
  <td><b>91.4</b></td>
285
- <td>89.5</td>
286
  </tr>
287
  </tbody>
288
  </table>
289
 
290
- ## Useful links
291
- - View our [release blogpost](https://huggingface.co/blog/falcon3).
292
- - Feel free to join [our discord server](https://discord.gg/fwXpMyGc) if you have any questions or to interact with our researchers and developers.
293
-
294
  ## Technical Report
295
  Coming soon....
296
 
 
7
  tags:
8
  - falcon3
9
  base_model: tiiuae/Falcon3-7B-Base
10
+ license: other
11
+ license_name: falcon-llm-license
12
  license_link: https://falconllm.tii.ae/falcon-terms-and-conditions.html
 
13
  ---
14
 
 
 
 
 
15
  # Falcon3-7B-Instruct
16
 
17
+ **Falcon3** family of Open Foundation Models is a set of pretrained and instruct LLMs ranging from 1B to 10B parameters.
18
 
19
+ This repository contains the **Falcon3-7B-Instruct**. It achieves state-of-the-art results (at release's time) on reasoning, language understanding, instruction following, code and mathematics tasks.
20
+ Falcon3-7B-Instruct supports 4 languages (English, French, Spanish, Portuguese) and a context length of up to 32K.
21
 
22
  ## Model Details
23
  - Architecture
24
+ - Transformer-based causal decoder-only architecture
25
  - 28 decoder blocks
26
+ - Grouped Query Attention (GQA) for faster inference: 12 query heads and 4 key-value heads
27
  - Wider head dimension: 256
28
  - High RoPE value to support long context understanding: 1000042
29
  - Uses SwiGLU and RMSNorm
30
  - 32K context length
31
  - 131K vocab size
32
+ - Pretrained on 14 Teratokens of datasets comprising of web, code, STEM, high quality and mutlilingual data using 2048 H100 GPU chips
33
+ - Posttrained on 1.2 million samples of STEM, conversational, code, safety and function call data
34
  - Supports EN, FR, ES, PT
35
  - Developed by [Technology Innovation Institute](https://www.tii.ae)
36
  - License: TII Falcon-LLM License 2.0
 
86
  <br>
87
 
88
  ## Benchmarks
89
+ We report in the following table our internal pipeline benchmarks:
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
90
 
91
  <table border="1" style="width: 100%; text-align: center; border-collapse: collapse;">
92
  <colgroup>
 
109
  <tr>
110
  <td rowspan="3">General</td>
111
  <td>MMLU (5-shot)</td>
112
+ <td>55.9</td>
113
+ <td><b>72.4</b></td>
114
+ <td>68</td>
115
  </tr>
116
  <tr>
117
  <td>MMLU-PRO (5-shot)</td>
118
+ <td>21.8</td>
119
+ <td>35.8</td>
120
+ <td><b>40.7</b></td>
121
  </tr>
122
  <tr>
123
  <td>IFEval</td>
 
128
  <tr>
129
  <td rowspan="3">Math</td>
130
  <td>GSM8K (5-shot)</td>
131
+ <td>78.1</td>
132
+ <td>77.5</td>
133
+ <td><b>79.1</b></td>
134
  </tr>
135
  <tr>
136
  <td>GSM8K (8-shot, COT)</td>
137
+ <td>79.8</td>
138
+ <td>72.7</td>
139
+ <td><b>80.9</b></td>
140
  </tr>
141
  <tr>
142
  <td>MATH Lvl-5 (4-shot)</td>
143
+ <td>10.4</td>
144
+ <td>26</td>
145
+ <td><b>33.1</b></td>
146
  </tr>
147
  <tr>
148
  <td rowspan="5">Reasoning</td>
149
  <td>Arc Challenge (25-shot)</td>
150
+ <td>46.6</td>
151
+ <td>55.7</td>
152
+ <td><b>65.9</b></td>
153
  </tr>
154
  <tr>
155
  <td>GPQA (0-shot)</td>
156
+ <td><b>33.6</b></td>
 
157
  <td>31.9</td>
158
+ <td>32</td>
159
  </tr>
160
  <tr>
161
  <td>GPQA (0-shot, COT)</td>
 
166
  <tr>
167
  <td>MUSR (0-shot)</td>
168
  <td>38.6</td>
169
+ <td>40.7</td>
170
  <td><b>46.4</b></td>
171
  </tr>
172
  <tr>
173
  <td>BBH (3-shot)</td>
174
+ <td>43.7</td>
175
+ <td><b>53.9</b></td>
176
  <td>52.4</td>
177
  </tr>
178
  <tr>
 
203
  <tr>
204
  <td rowspan="2">Instructions following</td>
205
  <td>MT-Bench (avg)</td>
206
+ <td>7.86</td>
207
+ <td><b>8.54</b></td>
208
+ <td>8.36</td>
209
  </tr>
210
  <tr>
211
+ <td>Alapaca (WC)</td>
212
+ <td>26.57</td>
213
  <td><b>31.5</b></td>
214
+ <td>26.13</td>
215
  </tr>
216
  <tr>
217
  <td>Tool use</td>
218
  <td>BFCL AST (avg)</td>
219
  <td>90.6</td>
220
  <td><b>91.4</b></td>
221
+ <td>72.3</td>
222
  </tr>
223
  </tbody>
224
  </table>
225
 
 
 
 
 
226
  ## Technical Report
227
  Coming soon....
228
 
tokenizer_config.json CHANGED
@@ -16219,7 +16219,7 @@
16219
  ">>PASSWORD<<",
16220
  ">>KEY<<"
16221
  ],
16222
- "chat_template": "{%- if tools %}\n{{- '<|system|>\\n' }}\n{%- if messages[0]['role'] == 'system' %}\n{{- messages[0]['content'] }}\n{%- set remaining_messages = messages[1:] %}\n{%- else %}\n{%- set remaining_messages = messages %}\n{%- endif %}\n{{- 'You are a Falcon assistant skilled in function calling. You are helpful, respectful, and concise.\\n\\n# Tools\\n\\nYou have access to the following functions. You MUST use them to answer questions when needed. For each function call, you MUST return a JSON object inside <tool_call></tool_call> tags.\\n\\n<tools>' + tools|tojson(indent=2) + '</tools>\\n\\n# Output Format\\n\\nYour response MUST follow this format when making function calls:\\n<tool_call>\\n[\\n {\"name\": \"function_name\", \"arguments\": {\"arg1\": \"value1\", \"arg2\": \"value2\"}},\\n {\"name\": \"another_function\", \"arguments\": {\"arg\": \"value\"}}\\n]\\n</tool_call>\\nIf no function calls are needed, respond normally without the tool_call tags.\\n' }}\n{%- for message in remaining_messages %}\n{%- if message['role'] == 'user' %}\n{{- '<|user|>\\n' + message['content'] + '\\n' }}\n{%- elif message['role'] == 'assistant' %}\n{%- if message.content %}\n{{- '<|assistant|>\\n' + message['content'] }}\n{%- endif %}\n{%- if message.tool_calls %}\n{{- '\\n<tool_call>\\n' }}\n{{- message.tool_calls|tojson(indent=2) }}\n{{- '\\n</tool_call>' }}\n{%- endif %}\n{{- eos_token + '\\n' }}\n{%- elif message['role'] == 'tool' %}\n{{- '<|assistant|>\\n<tool_response>\\n' + message['content'] + '\\n</tool_response>\\n' }}\n{%- endif %}\n{%- endfor %}\n{{- '<|assistant|>\\n' if add_generation_prompt }}\n{%- else %}\n{%- for message in messages %}\n{%- if message['role'] == 'system' %}\n{{- '<|system|>\\n' + message['content'] + '\\n' }}\n{%- elif message['role'] == 'user' %}\n{{- '<|user|>\\n' + message['content'] + '\\n' }}\n{%- elif message['role'] == 'assistant' %}\n{%- if not loop.last %}\n{{- '<|assistant|>\\n' + message['content'] + eos_token + '\\n' }}\n{%- else %}\n{{- '<|assistant|>\\n' + message['content'] + eos_token }}\n{%- endif %}\n{%- endif %}\n{%- if loop.last and add_generation_prompt %}\n{{- '<|assistant|>\\n' }}\n{%- endif %}\n{%- endfor %}\n{%- endif %}",
16223
  "clean_up_tokenization_spaces": true,
16224
  "eos_token": "<|endoftext|>",
16225
  "extra_special_tokens": {},
 
16219
  ">>PASSWORD<<",
16220
  ">>KEY<<"
16221
  ],
16222
+ "chat_template": "{% for message in messages %}{% if message['role'] == 'system' %}{{ '<|system|>\n' + message['content'] + '\n' }}{% elif message['role'] == 'user' %}{{ '<|user|>\n' + message['content'] + '\n' }}{% elif message['role'] == 'assistant' %}{% if not loop.last %}{{ '<|assistant|>\n' + message['content'] + eos_token + '\n' }}{% else %}{{ '<|assistant|>\n' + message['content'] + eos_token }}{% endif %}{% endif %}{% if loop.last and add_generation_prompt %}{{ '<|assistant|>\n' }}{% endif %}{% endfor %}",
16223
  "clean_up_tokenization_spaces": true,
16224
  "eos_token": "<|endoftext|>",
16225
  "extra_special_tokens": {},