kyujinpy commited on
Commit
2cd47c0
1 Parent(s): fdb3fcb

Upload README.md

Browse files
Files changed (1) hide show
  1. README.md +72 -0
README.md CHANGED
@@ -1,3 +1,75 @@
1
  ---
 
 
 
 
 
 
2
  license: cc-by-nc-sa-4.0
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ language:
3
+ - en
4
+ datasets:
5
+ - garage-bAInd/Open-Platypus
6
+ library_name: transformers
7
+ pipeline_tag: text-generation
8
  license: cc-by-nc-sa-4.0
9
  ---
10
+
11
+ # **PlatYi-34B-Llama-AQ**
12
+ <img src='./PlatYi.png' width=256>
13
+
14
+ ## Model Details
15
+
16
+ **Model Developers** Kyujin Han (kyujinpy)
17
+
18
+ **Input** Models input text only.
19
+
20
+ **Output** Models generate text only.
21
+
22
+ **Model Architecture**
23
+ PlatYi-34B-Llama-AQ is an auto-regressive language model based on the Yi-34B transformer architecture.
24
+
25
+ **Blog Link**
26
+ Blog: [Coming soon...]
27
+ Github: [Coming soon...]
28
+
29
+ **Base Model**
30
+ [chargoddard/Yi-34B-Llama](https://huggingface.co/chargoddard/Yi-34B-Llama)
31
+
32
+ **Training Dataset**
33
+ [garage-bAInd/Open-Platypus](https://huggingface.co/datasets/garage-bAInd/Open-Platypus).
34
+
35
+ ## Fix some bugs
36
+ - Before model, there is some mistakes.
37
+ - I modified the templates and warmup_steps.
38
+
39
+ ## Notice
40
+ While training, I used Q-LoRA.
41
+ The lora_r values is 64.
42
+
43
+
44
+ # **Model Benchmark**
45
+
46
+ ## Open leaderboard
47
+ - Follow up as [link](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).
48
+
49
+ | Model | Average | ARC | HellaSwag | MMLU | TruthfulQA | Winogrande | GSM8K |
50
+ | --- | --- | --- | --- | --- | --- | --- | --- |
51
+ | PlatYi-34B-Llama-AQ | NaN | NaN | NaN | NaN | NaN | NaN | NaN |
52
+ | PlatYi-34B-Llama-Q-v2 | NaN | NaN | NaN | NaN | NaN | NaN | NaN |
53
+ | PlatYi-34B-Llama-Q | 71.13 | 65.70 | 85.22 | 78.78 | 53.64 | 83.03 | 60.42 |
54
+ | PlatYi-34B-Llama | 68.37 | 67.83 | 85.35 | 78.26 | 53.46 | 82.87 | 42.46 |
55
+ | [Yi-34B-Llama](https://huggingface.co/chargoddard/Yi-34B-Llama) | 70.95 | 64.59 | 85.63 | 76.31 | 55.60 | 82.79 | 60.80 |
56
+ | [Yi-34B](https://huggingface.co/01-ai/Yi-34B) | 69.42 | 64.59 | 85.69 | 76.35 | 56.23 | 83.03 | 50.64 |
57
+
58
+
59
+ # Implementation Code
60
+ ```python
61
+ ### KO-Platypus
62
+ from transformers import AutoModelForCausalLM, AutoTokenizer
63
+ import torch
64
+
65
+ repo = "kyujinpy/PlatYi-34B-Llama-AQ"
66
+ OpenOrca = AutoModelForCausalLM.from_pretrained(
67
+ repo,
68
+ return_dict=True,
69
+ torch_dtype=torch.float16,
70
+ device_map='auto'
71
+ )
72
+ OpenOrca_tokenizer = AutoTokenizer.from_pretrained(repo)
73
+ ```
74
+
75
+ ---