Post
411
Tutorial ๐ฅ Training a non-English reasoning model with GRPO and Unsloth
I wanted to share my experiment with training reasoning models in languages other than English/Chinese.
Using Llama 3.1 8B as base, GRPO trainer from trl, and Unsloth optimizations, I got a working prototype in Bulgarian after ~5 hours on an L40S GPU. The approach should work for any language where the base model has some pre-training coverage.
Full code and tutorial here: https://unfoldai.com/reasoning-in-a-non-english-language/
The model itself: s-emanuilov/LLMBG-Llama-3.1-8B-BG-Reasoning-v0.1
I hope this helps anyone looking to build reasoning models in their language.
I wanted to share my experiment with training reasoning models in languages other than English/Chinese.
Using Llama 3.1 8B as base, GRPO trainer from trl, and Unsloth optimizations, I got a working prototype in Bulgarian after ~5 hours on an L40S GPU. The approach should work for any language where the base model has some pre-training coverage.
Full code and tutorial here: https://unfoldai.com/reasoning-in-a-non-english-language/
The model itself: s-emanuilov/LLMBG-Llama-3.1-8B-BG-Reasoning-v0.1
I hope this helps anyone looking to build reasoning models in their language.