Update README.md
Browse files
README.md
CHANGED
@@ -1,6 +1,8 @@
|
|
1 |
---
|
2 |
library_name: transformers
|
3 |
-
license:
|
|
|
|
|
4 |
base_model: google/gemma-2-27b
|
5 |
tags:
|
6 |
- axolotl
|
@@ -10,6 +12,40 @@ model-index:
|
|
10 |
results: []
|
11 |
---
|
12 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
13 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
14 |
should probably proofread and complete it, then remove this comment. -->
|
15 |
|
|
|
1 |
---
|
2 |
library_name: transformers
|
3 |
+
license:
|
4 |
+
- llama3.1
|
5 |
+
- gemma
|
6 |
base_model: google/gemma-2-27b
|
7 |
tags:
|
8 |
- axolotl
|
|
|
12 |
results: []
|
13 |
---
|
14 |
|
15 |
+
# Llama-Gemma-2-27b-SFT-trial1
|
16 |
+
|
17 |
+
## 概要
|
18 |
+
|
19 |
+
[google/gemma-2-27b](https://huggingface.co/google/gemma-2-27b)を教師あり学習によりInstruction Tuningしたモデルです。
|
20 |
+
|
21 |
+
[松尾研大規模言語モデル講座2024](https://weblab.t.u-tokyo.ac.jp/lecture/course-list/large-language-model/)のコンペ用の提出モデル作成の一環として作成・公開しています。
|
22 |
+
|
23 |
+
This model is built with Llama and Qwen.
|
24 |
+
|
25 |
+
## 使用データセット
|
26 |
+
|
27 |
+
- [Aratako/Magpie-Tanuki-Qwen2.5-72B-Answered](https://huggingface.co/datasets/Aratako/Magpie-Tanuki-Qwen2.5-72B-Answered)
|
28 |
+
- [Aratako/magpie-qwen2.5-32b-reasoning-100k-formatted](https://huggingface.co/datasets/Aratako/magpie-qwen2.5-32b-reasoning-100k-formatted)
|
29 |
+
- [Aratako/magpie-reasoning-llama-nemotron-70b-100k-filtered](https://huggingface.co/datasets/Aratako/magpie-reasoning-llama-nemotron-70b-100k-filtered)
|
30 |
+
- [Aratako/Open-Platypus-Japanese-masked-formatted](https://huggingface.co/datasets/Aratako/Open-Platypus-Japanese-masked-formatted)
|
31 |
+
- [kanhatakeyama/wizardlm8x22b-logical-math-coding-sft_additional-ja](https://huggingface.co/datasets/kanhatakeyama/wizardlm8x22b-logical-math-coding-sft_additional-ja)
|
32 |
+
- [kanhatakeyama/ramdom-to-fixed-multiturn-Calm3](https://huggingface.co/datasets/kanhatakeyama/ramdom-to-fixed-multiturn-Calm3)
|
33 |
+
- [Aratako/magpie-ultra-v0.1-formatted](https://huggingface.co/datasets/Aratako/magpie-ultra-v0.1-formatted)
|
34 |
+
- [Aratako/orca-agentinstruct-1M-v1-selected](https://huggingface.co/datasets/Aratako/orca-agentinstruct-1M-v1-selected)
|
35 |
+
- [Aratako/Synthetic-JP-EN-Coding-Dataset-801k-50k](https://huggingface.co/datasets/Aratako/Synthetic-JP-EN-Coding-Dataset-801k-50k)
|
36 |
+
|
37 |
+
## ライセンス
|
38 |
+
|
39 |
+
本モデルは学習に利用したデータの関係で以下のライセンスの影響を受けます。
|
40 |
+
|
41 |
+
- [META LLAMA 3.1 COMMUNITY LICENSE](https://www.llama.com/llama3_1/license/)を継承します。
|
42 |
+
- [Gemma Terms of Use](https://ai.google.dev/gemma/terms)を継承します。
|
43 |
+
- [Qwen LICENSE AGREEMENT](https://huggingface.co/Qwen/Qwen2.5-72B-Instruct/blob/main/LICENSE)の影響を受けます。ライセンスは継承しませんが、「Built with Qwen」のような文言を記載する必要があります。
|
44 |
+
|
45 |
+
## 学習に関する詳細
|
46 |
+
|
47 |
+
本モデルの学習には[axolotl](https://github.com/axolotl-ai-cloud/axolotl)を使いました。パラメータ等の学習の設定は下記の自動生成された記述をご確認ください。
|
48 |
+
|
49 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
50 |
should probably proofread and complete it, then remove this comment. -->
|
51 |
|