pittawat's picture
Add text-generation pipeline tag (#1)
7e83a06 verified
|
raw
history blame
7.66 kB
---
library_name: transformers
license: llama3.2
base_model: scb10x/llama3.2-typhoon2-3b-instruct
model-index:
- name: llama3.2-typhoon2-3b-instruct
results: []
language:
- en
- th
datasets:
- scb10x/typhoon-t1-3b-research-preview-data
pipeline_tag: text-generation
---
# Typhoon T1 3B (Research Preview)
**Typhoon T1 3B (Research Preview)** is the first in a new family of open reasoning model "**Typhoon T**". Reasoning model is a novel type of model that think longer before giving a final answer.
Typhoon T1 3B (Research Preview) is built on top of [Typhoon 2 3B Instruct](https://huggingface.co/scb10x/llama3.2-typhoon2-3b-instruct). It has improved performance on challenging benchmarks like GPQA, MMLU Pro, and AI Mathematics Olympiad validation set.
> [!NOTE]
> 🚀
> **2025-02-01 Update**: We have released a new version of Typhoon T1 3B (Research Preview) with the ability to 🇹🇭 *generate Thai reasoning traces*, improved *Thai performance in general*, and *enhanced instruction following*. This version has the a comparative level of English performance to `v2025-01-23`.
## Key Points
- **Typhoon T1** is a new family of open reasoning models developed by SCB 10X
- **Typhoon T1 3B (Research Preview)**, the *first* in the **Typhoon T** family, shows improved performance across **challenging benchmarks** compared to the original Typhoon 2 3B Instruct
- Typhoon T1 3B (Research Preview) offers a **fast**, **low-compute requirements** model, yet is **capable** in a variety of tasks by scaling test-time compute, enabling the model to think longer before giving a final answer. Typhoon T1 3B (Research Preview) is able to **_reason across domains_**, unlike many open reasoning models limited to mathematics and coding
- We **open** our recipe for data pipeline and training this model without distilling from other reasoning models
- We introduce **a new thinking paradigm** for reasoning models, structured thinking, where we add auxiliary tokens to help structure the thinking process of the model. This approach shows an increase in performance over a common variant of separating only thought and response parts based on our experiments
- Typhoon T1 3B (Research Preview) `v2025-02-01` is the first reasoning model where we intentionally equipped the model with the ability to **generate Thai reasoning traces**, improving *transparency* and *interpretability* of the model.
For more technical details, please visit our [technical blog](https://blog.opentyphoon.ai/introducing-typhoon-t1-a-family-of-open-reasoning-models-research-preview-22daacc88662).
* To acknowledge Meta's effort in creating the foundation model and to comply with the license, we explicitly include `llama-3.2` in the model name.
## Performance
| Model name | GSM8K (↑), 8-shot | HumanEval+ (↑), Pass@10 | GPQA (↑), 0CoT | AIME (↑) |
|-----------------------|-------------------|-------------------------|----------------|----------|
| Typhoon 2 3B Instruct | 56.63 | 66 | 27.01 | 0 |
| Typhoon T1 3B (semi) | 59.59 | 68.99 | 25.89 | 0 |
| **Typhoon T1 3B (Research Preview)** v2025-01-23 | **62.40** | **69.87** | **31.7** | **2.22** |
### MMLU Pro (↑), 5-shot
| Model name | Average | Math | Health | Physics | Business | Biology | Chemistry | Computer Science | Economics | Engineering | Philosophy | Other | History | Psychology | Law |
|-----------------------|-----------|-----------|-----------|----------|-----------|-----------|-----------|------------------|-----------|-------------|------------|-----------|-----------|------------|-----------|
| Typhoon 2 3B Instruct | 26.7 | 26.8 | 33.62 | 23.4 | 25.35 | 43.38 | 19.88 | 28.29 | 35.43 | 18.37 | 28.06 | 27.92 | 25.72 | 37.84 | 13.17 |
| **Typhoon T1 3B (Research Preview)** v2025-01-23 | **30.65** | **30.57** | **36.19** | **27.1** | **31.69** | **50.77** | **22.17** | **31.22** | **38.86** | **21.98** | **30.66** | **32.79** | **26.51** | **43.36** | **17.26** |
## Model description
- **Model type**: A 3B instruct decoder-only model based on Llama architecture.
- **Requirement**: `transformers` 4.46.1 or newer.
- **Primary Language(s)**: English 🇬🇧 and Thai 🇹🇭
- **License**: [Llama 3.2 Community License](https://github.com/meta-llama/llama-models/blob/main/models/llama3_2/LICENSE)
## Usage Examples
⚠️ Please note that `max_new_tokens` should be at least `512`, but is recommended at a minimum of `1,024` to provide space for complete generation.
```python
from transformers import AutoTokenizer, AutoModelForCausalLM
import torch
model_id = "scb10x/llama-3.2-typhoon-t1-3b-research-preview"
revision = "main" # To use the previous version comment this line
# revision = "v2025-01-23" # To use the previous version uncomment this line
tokenizer = AutoTokenizer.from_pretrained(model_id, revision=revision)
model = AutoModelForCausalLM.from_pretrained(
model_id,
torch_dtype=torch.bfloat16,
device_map="auto",
revision=revision
)
messages = [
{"role": "user", "content": "หากแปลคำว่า \"ไต้ฝุ่น\" เป็นภาษาอังกฤษ ในคำที่ถูกแปลแล้วจะมีตัวอักษร \"o\" ทั้งหมดกี่ตัว"},
]
input_ids = tokenizer.apply_chat_template(
messages,
add_generation_prompt=True,
return_tensors="pt"
).to(model.device)
terminators = [
tokenizer.eos_token_id,
tokenizer.convert_tokens_to_ids("<|eot_id|>")
]
outputs = model.generate(
input_ids,
max_new_tokens=1024,
eos_token_id=terminators,
do_sample=False,
temperature=0.0,
top_p=0.9,
)
response = outputs[0][input_ids.shape[-1]:]
print(tokenizer.decode(response, skip_special_tokens=True))
```
### OpenAI API-compatible Server with vLLM
```bash
pip install vllm
vllm serve scb10x/llama-3.2-typhoon-t1-3b-research-preview
# To serve the previous version, add the revision parameter as shown below
# vllm serve scb10x/llama-3.2-typhoon-t1-3b-research-preview --revision v2025-01-23
# see more information at https://docs.vllm.ai/
```
## Intended uses & limitations
While we made an effort to make our model safe, like all generative models, it may generate unsafe content in rare cases. Introducing a reasoning model paradigm may introduce some unforeseen behaviors, as model safety in the reasoning domain is a relatively new and ongoing area of research.
## Follow Us
[https://twitter.com/opentyphoon](https://twitter.com/opentyphoon)
## Support
[https://discord.gg/CqyBscMFpg](https://discord.gg/CqyBscMFpg)
## Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 6
- eval_batch_size: 1
- seed: 42
- distributed_type: multi-GPU
- num_devices: 8
- gradient_accumulation_steps: 6
- total_train_batch_size: 288
- total_eval_batch_size: 8
- optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- num_epochs: 2.0
## Citation
```
@misc{taveekitworachai2025typhoont1openthai,
title={Typhoon T1: An Open Thai Reasoning Model},
author={Pittawat Taveekitworachai and Potsawee Manakul and Kasima Tharnpipitchai and Kunat Pipatanakul},
year={2025},
eprint={2502.09042},
archivePrefix={arXiv},
primaryClass={cs.CL},
url={https://arxiv.org/abs/2502.09042},
}
```