--- license: llama3 language: - ko tags: - korean - llama3 - instruction-tuning - dora datasets: - Acyrl - llm-kr-eval - Counter-MT-bench base_model: - meta-llama/Meta-Llama-3-8B pipeline_tag: text-generation --- # A-LLM: Korean Language Model based on Llama-3 ## Introduction A-LLM is a Korean language model built on Meta's Llama-3-8B architecture, specifically optimized for Korean language understanding and generation. The model was trained using the DoRA (Weight-Decomposed Low-Rank Adaptation) methodology on a comprehensive Korean dataset, achieving state-of-the-art performance among open-source Korean language models. ## Performance Benchmarks ### Horangi Korean LLM Leaderboard The model's performance was evaluated using the Horangi Korean LLM Leaderboard , which combines two major evaluation frameworks normalized to a 1.0 scale and averages their scores. #### 1. LLM-KR-EVAL A comprehensive benchmark that measures fundamental NLP capabilities across 5 core tasks: - Natural Language Inference (NLI) - Question Answering (QA) - Reading Comprehension (RC) - Entity Linking (EL) - Fundamental Analysis (FA) The benchmark comprises 10 different datasets distributed across these tasks, providing a thorough assessment of Korean language understanding and processing capabilities. #### 2. MT-Bench A diverse evaluation framework consisting of 80 questions (10 questions each from 8 categories), evaluated using GPT-4 as the judge. Categories include: - Writing - Roleplay - Extraction - Reasoning - Math - Coding - Knowledge (STEM) - Knowledge (Humanities/social science) ### Performance Results | Model | Total Score | AVG_llm_kr_eval | AVG_mtbench | |-------|-------------|-----------------|-------------| | A-LLM (Ours) | 0.6675 | 0.5937 | 7.413 | | GPT-4 | 0.7363 | 0.6158 | 8.569 | | Mixtral-8x7B | 0.5843 | 0.4304 | 7.381 | | KULLM3 | 0.5764 | 0.5204 | 6.325 | | SOLAR-1-mini | 0.5173 | 0.37 | 6.647 | Our model achieves state-of-the-art performance among open-source Korean language models, demonstrating strong capabilities across both general language understanding (LLM-KR-EVAL) and diverse task-specific applications (MT-Bench). ### Model Components This repository provides: - Tokenizer configuration - Model weights in safetensor format ## Usage Instructions ### Prerequisites - Python 3.8 or higher - PyTorch 2.0 or higher - Transformers library