libray_name: transformers
pipeline_tag: text-generation
license: other
license_name: llama3
license_link: LICENSE
language:
- ko
- en
tags:
- meta
- llama
- llama-3
- akallama
library_name: transformers
AKALLAMA
AkaLlama is a series of Korean language models designed for practical usability across a wide range of tasks. The initial model, AkaLlama-v0.1, is a fine-tuned version of Meta-Llama-3-70b-Instruct. It has been trained on a custom mix of publicly available datasets curated by the MIR Lab. Our goal is to explore cost-effective ways to adapt high-performing LLMs for specific use cases, such as different languages (e.g., Korean) or domains (e.g., organization-specific chatbots).
Model Description
This is the model card of a π€ transformers model that has been pushed on the Hub.
- Developed by: Yonsei MIRLab
- Language(s) (NLP): Korean, English
- License: llama3
- Finetuned from model: meta-llama/Meta-Llama-3-70B-Instruct
How to use
This repo provides full model weight files for AkaLlama-70B-v0.1.
Use with transformers
See the snippet below for usage with Transformers:
import transformers
import torch
model_id = "mirlab/AkaLlama-llama3-70b-v0.1"
pipeline = transformers.pipeline(
"text-generation",
model=model_id,
model_kwargs={"torch_dtype": torch.bfloat16},
device="auto",
)
system_prompt = """
λΉμ μ μ°μΈλνκ΅ λ©ν°λͺ¨λ¬ μ°κ΅¬μ€ (MIR lab) μ΄ λ§λ λκ·λͺ¨ μΈμ΄ λͺ¨λΈμΈ AkaLlama (μμΉ΄λΌλ§) μ
λλ€.\nλ€μ μ§μΉ¨μ λ°λ₯΄μΈμ:\n1. μ¬μ©μκ° λ³λλ‘ μμ²νμ§ μλ ν νμ νκΈλ‘ μν΅νμΈμ.\n2. μ ν΄νκ±°λ λΉμ€λ¦¬μ , μ°¨λ³μ , μννκ±°λ λΆλ²μ μΈ λ΄μ©μ΄ λ΅λ³μ ν¬ν¨λμ΄μλ μ λ©λλ€.\n3. μ§λ¬Έμ΄ λ§μ΄ λμ§ μκ±°λ μ¬μ€μ λΆν©νμ§ μλ κ²½μ° μ λ΅ λμ κ·Έ μ΄μ λ₯Ό μ€λͺ
νμΈμ. μ§λ¬Έμ λν λ΅μ λͺ¨λ₯Έλ€λ©΄ κ±°μ§ μ 보λ₯Ό 곡μ νμ§ λ§μΈμ.\n4. μμ μ΄λ μ€λ¦¬μ μλ°°λμ§ μλ ν μ¬μ©μμ λͺ¨λ μ§λ¬Έμ μμ νκ³ ν¬κ΄μ μΌλ‘ λ΅λ³νμΈμ.
"""
messages = [
{"role": "system", "content": "system_prompt"},
{"role": "user", "content": "λ€ μ΄λ¦μ λμΌ?"},
]
prompt = pipeline.tokenizer.apply_chat_template(
messages,
tokenize=False,
add_generation_prompt=True
)
terminators = [
pipeline.tokenizer.eos_token_id,
pipeline.tokenizer.convert_tokens_to_ids("<|eot_id|>")
]
outputs = pipeline(
prompt,
max_new_tokens=256,
eos_token_id=terminators,
do_sample=True,
temperature=0.6,
top_p=0.9,
)
print(outputs[0]["generated_text"][len(prompt):])
Training Details
Training Procedure
We trained AkaLlama using a preference learning alignment algorithm called Odds Ratio Preference Optimization (ORPO). Our training pipeline is almost identical to that of HuggingFaceH4/zephyr-orpo-141b-A35b-v0.1, aside from minor hyperparameter changes. Please check out Huggingface's alignment handbook for further details, including the chat template.
Training Data
Detailed descriptions regarding training data will be announced later.
Examples
You can find more examples at our project page
Special Thanks
- Data Center of the Department of Artificial Intelligence at Yonsei University for the computation resources