File size: 7,656 Bytes
9fef852 5c76aa4 9fef852 ee41035 9fef852 5c76aa4 1696ee6 7e83a06 9fef852 ee41035 9fef852 ee41035 9fef852 ee41035 5c76aa4 a8b38fa 5c76aa4 ee41035 4cc5760 5c76aa4 a8b38fa 5c76aa4 66c8a70 5c76aa4 a8b38fa 5c76aa4 a8b38fa 9fef852 5c76aa4 a8b38fa 5c76aa4 ee41035 a8b38fa 5c76aa4 a8b38fa 5c76aa4 a8b38fa 5c76aa4 ee41035 a8b38fa 5c76aa4 9fef852 5c76aa4 9fef852 5c76aa4 9fef852 5c76aa4 9fef852 5c76aa4 9fef852 5c76aa4 9fef852 5c76aa4 9fef852 5c76aa4 8c322c7 5c76aa4 8c322c7 5c76aa4 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 |
---
library_name: transformers
license: llama3.2
base_model: scb10x/llama3.2-typhoon2-3b-instruct
model-index:
- name: llama3.2-typhoon2-3b-instruct
results: []
language:
- en
- th
datasets:
- scb10x/typhoon-t1-3b-research-preview-data
pipeline_tag: text-generation
---
# Typhoon T1 3B (Research Preview)
**Typhoon T1 3B (Research Preview)** is the first in a new family of open reasoning model "**Typhoon T**". Reasoning model is a novel type of model that think longer before giving a final answer.
Typhoon T1 3B (Research Preview) is built on top of [Typhoon 2 3B Instruct](https://huggingface.co/scb10x/llama3.2-typhoon2-3b-instruct). It has improved performance on challenging benchmarks like GPQA, MMLU Pro, and AI Mathematics Olympiad validation set.
> [!NOTE]
> 🚀
> **2025-02-01 Update**: We have released a new version of Typhoon T1 3B (Research Preview) with the ability to 🇹🇭 *generate Thai reasoning traces*, improved *Thai performance in general*, and *enhanced instruction following*. This version has the a comparative level of English performance to `v2025-01-23`.
## Key Points
- **Typhoon T1** is a new family of open reasoning models developed by SCB 10X
- **Typhoon T1 3B (Research Preview)**, the *first* in the **Typhoon T** family, shows improved performance across **challenging benchmarks** compared to the original Typhoon 2 3B Instruct
- Typhoon T1 3B (Research Preview) offers a **fast**, **low-compute requirements** model, yet is **capable** in a variety of tasks by scaling test-time compute, enabling the model to think longer before giving a final answer. Typhoon T1 3B (Research Preview) is able to **_reason across domains_**, unlike many open reasoning models limited to mathematics and coding
- We **open** our recipe for data pipeline and training this model without distilling from other reasoning models
- We introduce **a new thinking paradigm** for reasoning models, structured thinking, where we add auxiliary tokens to help structure the thinking process of the model. This approach shows an increase in performance over a common variant of separating only thought and response parts based on our experiments
- Typhoon T1 3B (Research Preview) `v2025-02-01` is the first reasoning model where we intentionally equipped the model with the ability to **generate Thai reasoning traces**, improving *transparency* and *interpretability* of the model.
For more technical details, please visit our [technical blog](https://blog.opentyphoon.ai/introducing-typhoon-t1-a-family-of-open-reasoning-models-research-preview-22daacc88662).
* To acknowledge Meta's effort in creating the foundation model and to comply with the license, we explicitly include `llama-3.2` in the model name.
## Performance
| Model name | GSM8K (↑), 8-shot | HumanEval+ (↑), Pass@10 | GPQA (↑), 0CoT | AIME (↑) |
|-----------------------|-------------------|-------------------------|----------------|----------|
| Typhoon 2 3B Instruct | 56.63 | 66 | 27.01 | 0 |
| Typhoon T1 3B (semi) | 59.59 | 68.99 | 25.89 | 0 |
| **Typhoon T1 3B (Research Preview)** v2025-01-23 | **62.40** | **69.87** | **31.7** | **2.22** |
### MMLU Pro (↑), 5-shot
| Model name | Average | Math | Health | Physics | Business | Biology | Chemistry | Computer Science | Economics | Engineering | Philosophy | Other | History | Psychology | Law |
|-----------------------|-----------|-----------|-----------|----------|-----------|-----------|-----------|------------------|-----------|-------------|------------|-----------|-----------|------------|-----------|
| Typhoon 2 3B Instruct | 26.7 | 26.8 | 33.62 | 23.4 | 25.35 | 43.38 | 19.88 | 28.29 | 35.43 | 18.37 | 28.06 | 27.92 | 25.72 | 37.84 | 13.17 |
| **Typhoon T1 3B (Research Preview)** v2025-01-23 | **30.65** | **30.57** | **36.19** | **27.1** | **31.69** | **50.77** | **22.17** | **31.22** | **38.86** | **21.98** | **30.66** | **32.79** | **26.51** | **43.36** | **17.26** |
## Model description
- **Model type**: A 3B instruct decoder-only model based on Llama architecture.
- **Requirement**: `transformers` 4.46.1 or newer.
- **Primary Language(s)**: English 🇬🇧 and Thai 🇹🇭
- **License**: [Llama 3.2 Community License](https://github.com/meta-llama/llama-models/blob/main/models/llama3_2/LICENSE)
## Usage Examples
⚠️ Please note that `max_new_tokens` should be at least `512`, but is recommended at a minimum of `1,024` to provide space for complete generation.
```python
from transformers import AutoTokenizer, AutoModelForCausalLM
import torch
model_id = "scb10x/llama-3.2-typhoon-t1-3b-research-preview"
revision = "main" # To use the previous version comment this line
# revision = "v2025-01-23" # To use the previous version uncomment this line
tokenizer = AutoTokenizer.from_pretrained(model_id, revision=revision)
model = AutoModelForCausalLM.from_pretrained(
model_id,
torch_dtype=torch.bfloat16,
device_map="auto",
revision=revision
)
messages = [
{"role": "user", "content": "หากแปลคำว่า \"ไต้ฝุ่น\" เป็นภาษาอังกฤษ ในคำที่ถูกแปลแล้วจะมีตัวอักษร \"o\" ทั้งหมดกี่ตัว"},
]
input_ids = tokenizer.apply_chat_template(
messages,
add_generation_prompt=True,
return_tensors="pt"
).to(model.device)
terminators = [
tokenizer.eos_token_id,
tokenizer.convert_tokens_to_ids("<|eot_id|>")
]
outputs = model.generate(
input_ids,
max_new_tokens=1024,
eos_token_id=terminators,
do_sample=False,
temperature=0.0,
top_p=0.9,
)
response = outputs[0][input_ids.shape[-1]:]
print(tokenizer.decode(response, skip_special_tokens=True))
```
### OpenAI API-compatible Server with vLLM
```bash
pip install vllm
vllm serve scb10x/llama-3.2-typhoon-t1-3b-research-preview
# To serve the previous version, add the revision parameter as shown below
# vllm serve scb10x/llama-3.2-typhoon-t1-3b-research-preview --revision v2025-01-23
# see more information at https://docs.vllm.ai/
```
## Intended uses & limitations
While we made an effort to make our model safe, like all generative models, it may generate unsafe content in rare cases. Introducing a reasoning model paradigm may introduce some unforeseen behaviors, as model safety in the reasoning domain is a relatively new and ongoing area of research.
## Follow Us
[https://twitter.com/opentyphoon](https://twitter.com/opentyphoon)
## Support
[https://discord.gg/CqyBscMFpg](https://discord.gg/CqyBscMFpg)
## Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 6
- eval_batch_size: 1
- seed: 42
- distributed_type: multi-GPU
- num_devices: 8
- gradient_accumulation_steps: 6
- total_train_batch_size: 288
- total_eval_batch_size: 8
- optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- num_epochs: 2.0
## Citation
```
@misc{taveekitworachai2025typhoont1openthai,
title={Typhoon T1: An Open Thai Reasoning Model},
author={Pittawat Taveekitworachai and Potsawee Manakul and Kasima Tharnpipitchai and Kunat Pipatanakul},
year={2025},
eprint={2502.09042},
archivePrefix={arXiv},
primaryClass={cs.CL},
url={https://arxiv.org/abs/2502.09042},
}
``` |