|
--- |
|
language: |
|
- ko |
|
- en |
|
library_name: transformers |
|
license: cc-by-nc-sa-4.0 |
|
pipeline_tag: text-generation |
|
tags: |
|
- pytorch |
|
--- |
|
|
|
|
|
<!-- Provide a quick summary of what the model is/does. --> |
|
|
|
|
|
|
|
## Model Description |
|
|
|
<!-- Provide a longer summary of what this model is. --> |
|
|
|
νκ΅μ΄ LLM νκ° λ°μ΄ν°μ
μΈ kollm μ νμ©νμ¬ Supervised Fine-Tuning(a.k.a SFT) νμ΅ν λͺ¨λΈμ
λλ€. νμ΅ λ°μ΄ν°μ
μ KoAlpaca-v1.1, kollm_kmmlu, korean-parallel-corpora, kobest_sentineg μ κ°μ μ€ν λ°μ΄ν°μ
μΌλ‘ ꡬμ±λμ΄ μμ΅λλ€. λ°μ΄ν°μ λν μμΈν μ€λͺ
μ Train Datasets λ§ν¬λ₯Ό μ°Έκ³ ν΄μ£ΌμΈμ. |
|
|
|
## About the Model |
|
|
|
- **Name:** TwinDoc/RedWhale-tv-10.8B-sft-g |
|
- **Finetuned from model:** [TwinDoc/RedWhale-tv-10.8B-v1.0](https://huggingface.co/TwinDoc/RedWhale-tv-10.8B-v1.0) |
|
- **Train Datasets:** [davidkim205/kollm-converations](https://huggingface.co/datasets/davidkim205/kollm-converations?row=33) |
|
- **Developed by:** μ μμΌμλ€ (AGILESODA) |
|
- **Model type:** llama |
|
- **Language(s) (NLP):** νκ΅μ΄, μμ΄ |
|
- **License:** cc-by-nc-sa-4.0 |
|
|
|
- **train setting** |
|
- Lora r, alpha : 64, 16 |
|
- Dtype : bf16 |
|
- Epoch : 1 |
|
- Learning rate : 2e-4 |
|
- Global batch : 128 |
|
- Context length : 1024 |
|
|
|
|
|
- **inference setting** |
|
- BOS id : 1 |
|
- EOS id : 2 |
|
- Top-p : 0.95 |
|
- Temperature : 0.01 |
|
|
|
|
|
## prompt template |
|
|
|
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> |
|
|
|
``` |
|
A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions. |
|
Human: {input} |
|
Assistant: {output} |
|
``` |
|
|
|
|
|
|
|
## License |
|
|
|
<img src="https://huggingface.co/TwinDoc/agilesoda-model-x/resolve/main/license__icon.png" width="324"> |
|
|
|
The content of this project, created by AGILESODA, is licensed under the [Attribution-NonCommercial-ShareAlike 4.0 International (CC BY-NC-SA 4.0)](https://creativecommons.org/licenses/by-nc-sa/4.0/). |
|
|
|
## Citation |
|
|
|
``` |
|
@misc{vo2024redwhaleadaptedkoreanllm, |
|
title={RedWhale: An Adapted Korean LLM Through Efficient Continual Pretraining}, |
|
author={Anh-Dung Vo and Minseong Jung and Wonbeen Lee and Daewoo Choi}, |
|
year={2024}, |
|
eprint={2408.11294}, |
|
archivePrefix={arXiv}, |
|
primaryClass={cs.CL}, |
|
url={https://arxiv.org/abs/2408.11294}, |
|
} |
|
``` |
|
|
|
|
|
**Built with:** |
|
|
|
<a href="http://www.agilesoda.com/sub/twin_doc.php"> |
|
<img src="https://huggingface.co/TwinDoc/RedWhale-tv-10.8B-v1.0/resolve/main/company_agilesoda_twindoc__icon.png" alt="AgileSoda TwinDoc Icon"> |
|
</a> |