gemma-2-9b-HangulFixer

  • Base Model: unsloth/gemma-2-9b-bnb-4bit
  • Developed by: SeongeonKim
  • Base Model License: Apache 2.0
  • Finetuned Model License: CC BY-NC 4.0 (๋น„์˜๋ฆฌ์  ์‚ฌ์šฉ๋งŒ ํ—ˆ์šฉ)
  • Language: Korean (ko)
  • Tags:
    • text-generation-inference
    • transformers
    • unsloth
    • gemma2
    • trl
    • sft

๋ชจ๋ธ ์„ค๋ช…

gemma-2-9b-HangulFixer๋Š” Unsloth์™€ Hugging Face์˜ TRL ๋ผ์ด๋ธŒ๋Ÿฌ๋ฆฌ๋ฅผ ์‚ฌ์šฉํ•ด fine-tuningํ•œ ํ…์ŠคํŠธ ์ƒ์„ฑ ๋ชจ๋ธ์ž…๋‹ˆ๋‹ค.
์ด ๋ชจ๋ธ์€ ๋‚œ๋…ํ™”๋œ ํ•œ๊ธ€ ํ˜ธํ…” ๋ฆฌ๋ทฐ๋ฅผ ์›๋ž˜์˜ ๋ช…ํ™•ํ•˜๊ณ  ์ž์—ฐ์Šค๋Ÿฌ์šด ๋ฆฌ๋ทฐ๋กœ ๋ณต์›ํ•˜๊ธฐ ์œ„ํ•ด ์„ค๊ณ„๋˜์—ˆ์Šต๋‹ˆ๋‹ค.


๋ฐ์ดํ„ฐ ๋ฐฐ๊ฒฝ

ํ•œ๊ตญ ์ˆ™์†Œ ์˜ˆ์•ฝ ์‚ฌ์ดํŠธ์—์„œ๋Š” ๋ถ€์ •์ ์ธ ๋ฆฌ๋ทฐ๊ฐ€ ์‚ญ์ œ๋  ์ˆ˜ ์žˆ๋Š” ์šฐ๋ ค๊ฐ€ ์žˆ์Šต๋‹ˆ๋‹ค. ์ด๋ฅผ ํ”ผํ•˜๋ฉด์„œ ํ•œ๊ตญ ์‚ฌ์šฉ์ž๋“ค์—๊ฒŒ๋งŒ ์œ ์šฉํ•œ ์ •๋ณด๋ฅผ ์ „๋‹ฌํ•˜๊ธฐ ์œ„ํ•ด ์ผ๋ถ€ ์‚ฌ์šฉ์ž๋“ค์€ ๋ฆฌ๋ทฐ๋ฅผ ๋‚œ๋…ํ™”ํ•˜๋Š” ๋ฐฉ์‹์„ ์‚ฌ์šฉํ•ด ์†”์งํ•œ ํ›„๊ธฐ๋ฅผ ๋‚จ๊ธฐ๊ณ ์ž ํ•ฉ๋‹ˆ๋‹ค.
์ด ๋ฐฉ์‹์€ ํ•œ๊ตญ ์‚ฌ์šฉ์ž๋“ค์—๊ฒŒ๋Š” ์†”์งํ•œ ์ •๋ณด๋ฅผ ์ œ๊ณตํ•˜๋Š” ์žฅ์ ์ด ์žˆ์ง€๋งŒ, ์ˆ™์†Œ ์šด์˜์ž์™€ ๋ฐฉ๋ฌธ๊ฐ ๊ฐ„ ํ”ผ๋“œ๋ฐฑ ์ „๋‹ฌ์„ ์–ด๋ ต๊ฒŒ ๋งŒ๋“ค์–ด ์„œ๋น„์Šค ํ’ˆ์งˆ ํ–ฅ์ƒ์— ์žฅ์• ๊ฐ€ ๋  ์ˆ˜ ์žˆ์Šต๋‹ˆ๋‹ค.


ํ•™์Šต ๋ฐ์ดํ„ฐ

์ด ๋ชจ๋ธ์€ SeongeonKim/ko-scrambled_v0.1 ๋ฐ์ดํ„ฐ์…‹์˜ ํ•™์Šต ๋ฐ์ดํ„ฐ๋ฅผ ์‚ฌ์šฉํ•ด ํ•™์Šต๋˜์—ˆ์Šต๋‹ˆ๋‹ค.
์ด 11,263๊ฐœ์˜ ๋‚œ๋…ํ™”๋œ ํ•œ๊ธ€ ํ˜ธํ…” ๋ฆฌ๋ทฐ์™€ ๊ทธ์— ์ƒ์‘ํ•˜๋Š” ๋ณต์›๋œ ์›๋ณธ ๋ฆฌ๋ทฐ๋ฅผ ์‚ฌ์šฉํ•ด fine-tuning์ด ์ง„ํ–‰๋˜์—ˆ์Šต๋‹ˆ๋‹ค.
์ด ๋ฐ์ดํ„ฐ์…‹์€ ๋ฆฌ๋ทฐ์˜ ๋‚œ๋…ํ™”๋œ ํ˜•ํƒœ(input)์™€ ๋ณต์›๋œ ํ˜•ํƒœ(output)๋กœ ๊ตฌ์„ฑ๋˜์–ด ์žˆ์Šต๋‹ˆ๋‹ค.

์˜ˆ์ œ ๋ฐ์ดํ„ฐ:

  • Input: ์‰ญ๋”ฑ ์œ• ํ„บ๋ ค์œค ๊ตผ๋ฉ•์›จ ๋ท”ํ–… ์™„์ฉ ๋ณ„๋ฃŒ. 2๋นก 3์œŒ์—ฃ 88๋งŽ ์–ธ ์ผค์ณ…ํ–‡๋ˆˆ๋Žจ...
  • Output: ์‹ํƒ ์œ„ ๋”๋Ÿฌ์šด ๊ธˆ์•ก์— ๋น„ํ•ด ์™„์ „ ๋ณ„๋กœ. 2๋ฐ• 3์ผ์— 88๋งŒ ์› ๊ฒฐ์ œํ–ˆ๋Š”๋ฐ...

๋ชจ๋ธ ์„ฑ๋Šฅ

  • Base Model: unsloth/gemma-2-9b-bnb-4bit๋ฅผ ๊ธฐ๋ฐ˜์œผ๋กœ ํ•™์Šต๋˜์—ˆ์Šต๋‹ˆ๋‹ค.
  • Training Efficiency: Unsloth๋ฅผ ์‚ฌ์šฉํ•ด 2๋ฐฐ ๋” ๋น ๋ฅธ ์†๋„๋กœ ํ•™์Šต ์™„๋ฃŒ.
  • Target Task: ํ…์ŠคํŠธ ๋ณต์›
  • Language: ํ•œ๊ธ€

๊ฐœ์„  ์‚ฌํ•ญ (v0.1)

  • ์–ผ๋ฆฌ ์Šคํ† ํ•‘(Early Stopping) ๋„์ž…: EarlyStoppingCallback์„ ์ถ”๊ฐ€ํ•˜์—ฌ ํ‰๊ฐ€ ์†์‹ค(eval_loss) ๊ธฐ์ค€์œผ๋กœ ํ•™์Šต ๊ณผ์ •์„ ์ž๋™ ์ค‘๋‹จ, ๊ณผ์ ํ•ฉ ๋ฐฉ์ง€.
  • ํ•™์Šต ํšจ์œจ์„ฑ ํ–ฅ์ƒ: ๋ฐ์ดํ„ฐ ๋ณ‘๋ ฌ ์ฒ˜๋ฆฌ(dataset_num_proc=2) ๋ฐ ๋ช…์‹œ์  ํŒŒ๋ผ๋ฏธํ„ฐ ์„ค์ •์„ ํ†ตํ•ด ํ•™์Šต ์•ˆ์ •์„ฑ๊ณผ ์†๋„ ๊ฐœ์„ .
  • ํ‰๊ฐ€ ์ง€ํ‘œ ๋ชจ๋‹ˆํ„ฐ๋ง: ํ•™์Šต ํ’ˆ์งˆ์„ ๋†’์ด๊ธฐ ์œ„ํ•ด eval_loss ๋ชจ๋‹ˆํ„ฐ๋ง ์ถ”๊ฐ€.

์‚ฌ์šฉ ๋ฐฉ๋ฒ•

์•„๋ž˜๋Š” ๋ชจ๋ธ์„ ์‚ฌ์šฉํ•˜๋Š” ๊ธฐ๋ณธ์ ์ธ ์˜ˆ์ œ์ž…๋‹ˆ๋‹ค:

from transformers import AutoModelForCausalLM, AutoTokenizer

# ๋ชจ๋ธ ๋กœ๋“œ
model_name = "SeongeonKim/gemma-2-9b-HangulFixer"
model = AutoModelForCausalLM.from_pretrained(model_name).to("cuda")
tokenizer = AutoTokenizer.from_pretrained(model_name)

# ํ…Œ์ŠคํŠธ ์ž…๋ ฅ
input_text = "๋…€๋ฎจ๋„’๋ญ… ๋งŒ์ฃก์ˆญ๋Ÿฌ์œค ํšจํ…œ๋คผ์—์˜ค. ํ‘ธ์‹ธ๋ˆผ ์˜ฐ๋ฉด ์ฝ• ์ธ„์ฉํ•™๊ผฌ ์‹ถ์€ ์ฝง์‰ฐ์›จ์˜ค."
inputs = tokenizer(input_text, return_tensors="pt").to("cuda")
outputs = model.generate(**inputs, max_new_tokens=128)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))

์ƒ์—…์  ์ด์šฉ ๊ด€๋ จ ์ œํ•œ ์‚ฌํ•ญ

์ด ๋ชจ๋ธ์€ ํ•™์Šต ๋ฐ์ดํ„ฐ์˜ ๋ผ์ด์„ ์Šค์— ๋”ฐ๋ผ CC BY-NC 4.0 ๋ผ์ด์„ ์Šค๋ฅผ ๋”ฐ๋ฆ…๋‹ˆ๋‹ค.
์ด ๋ผ์ด์„ ์Šค๋Š” ๋น„์˜๋ฆฌ์  ์‚ฌ์šฉ๋งŒ ํ—ˆ์šฉํ•˜๋ฉฐ, ์ƒ์—…์  ์šฉ๋„๋กœ ์‚ฌ์šฉํ•˜๋ ค๋ฉด ๋ณ„๋„์˜ ํ—ˆ๊ฐ€๊ฐ€ ํ•„์š”ํ•ฉ๋‹ˆ๋‹ค.
๋ฐ์ดํ„ฐ๋ฅผ ์ด์šฉํ•˜๋Š” ๊ฒฝ์šฐ, ์—ฐ๊ตฌ ๋ชฉ์ ์œผ๋กœ ์‚ฌ์šฉํ•˜๊ฑฐ๋‚˜ ํ•™์Šต ๋ฐ์ดํ„ฐ์˜ ์ถœ์ฒ˜์™€ ๋ผ์ด์„ ์Šค๋ฅผ ๋ช…ํ™•ํžˆ ํ‘œ๊ธฐํ•ด์•ผ ํ•ฉ๋‹ˆ๋‹ค.

  • ๋ผ์ด์„ ์Šค ์„ธ๋ถ€์‚ฌํ•ญ: CC BY-NC 4.0

๋ชจ๋ธ ๊ฐœ๋ฐœ ๊ด€๋ จ ์ฐธ๊ณ  ์‚ฌํ•ญ

  • ์‚ฌ์šฉ๋œ ๊ธฐ์ˆ : Unsloth์™€ Hugging Face์˜ TRL(SFT) ์‚ฌ์šฉ
  • ๋ฐ์ดํ„ฐ ์ถœ์ฒ˜: SeongeonKim/ko-scrambled_v0.1 ๋ฐ์ดํ„ฐ์…‹
  • ์‚ฌ์šฉ ์ œํ•œ: ๋ฐ์ดํ„ฐ์˜ ๋ผ์ด์„ ์Šค๋ฅผ ์ค€์ˆ˜ํ•˜๋ฉฐ, ์ƒ์—…์  ์šฉ๋„๋กœ ์‚ฌ์šฉ ๋ถˆ๊ฐ€

This gemma2 model was trained 2x faster with Unsloth and Huggingface's TRL library.

Downloads last month
4
Safetensors
Model size
9.24B params
Tensor type
BF16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for SeongeonKim/gemma-2-9b-HangulFixer_v0.1

Base model

google/gemma-2-9b
Finetuned
(459)
this model