JeongwonChoi commited on
Commit
ecfe07c
β€’
1 Parent(s): 4faa809

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +59 -35
README.md CHANGED
@@ -1,77 +1,101 @@
1
  ---
2
  tags:
3
- - text-generation
4
  license: cc-by-nc-sa-4.0
5
  language:
6
- - ko
7
  base_model: yanolja/KoSOLAR-10.7B-v0.1
8
  pipeline_tag: text-generation
 
 
9
  ---
10
 
11
  # **DataVortexS-10.7B-v0.1**
12
- <img src="./DataVortex.png" alt="DataVortex" style="height: 8em;">
13
-
14
- ## **License**
15
 
16
- [cc-by-nc-sa-4.0](https://creativecommons.org/licenses/by-nc-sa/4.0/)
17
 
18
  ## **Model Details**
19
 
20
  ### **Base Model**
21
- [yanolja/KoSOLAR-10.7B-v0.1](https://huggingface.co/yanolja/KoSOLAR-10.7B-v0.1)
 
22
 
23
  ### **Trained On**
24
- H100 80GB 1ea
 
 
 
 
 
 
 
25
 
26
  ### **Instruction format**
27
 
28
- It follows **(No Input) Alpaca** format.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
29
 
30
  ## **Model Benchmark**
31
 
32
  ### **Ko-LLM-Leaderboard**
33
 
34
- On Benchmarking...
 
 
 
 
 
35
 
36
- # **Implementation Code**
37
 
38
- Since, chat_template already contains insturction format above.
39
  You can use the code below.
40
 
41
  ```python
42
  from transformers import AutoModelForCausalLM, AutoTokenizer
43
 
44
- device = "cuda"
45
 
46
- model = AutoModelForCausalLM.from_pretrained("Edentns/DataVortexS-10.7B-v0.1", device_map=device)
47
  tokenizer = AutoTokenizer.from_pretrained("Edentns/DataVortexS-10.7B-v0.1")
48
 
49
  messages = [
50
- { "role": "user", "content": "λŒ€ν•œλ―Όκ΅­μ˜ μˆ˜λ„λŠ” μ–΄λ””μ•Ό?" }
 
 
 
51
  ]
52
 
53
- encoded = tokenizer.apply_chat_template(
54
- messages,
55
- add_generation_prompt=True,
56
- return_tensors="pt",
57
- return_token_type_ids=False
58
- ).to(device)
59
-
60
- decoded = model.generate(
61
- input_ids=encoded,
62
- temperature=0.2,
63
- top_p=0.9,
64
- repetition_penalty=1.2,
65
- do_sample=True,
66
- max_length=4096,
67
- eos_token_id=tokenizer.eos_token_id,
68
- pad_token_id=tokenizer.eos_token_id
69
- )
70
- decoded = decoded[0][encoded.shape[1]:decoded[0].shape[-1]]
71
- decoded_text = tokenizer.decode(decoded, skip_special_tokens=True)
72
- print(decoded_text)
73
  ```
74
 
 
 
 
 
75
  <div align="center">
76
  <a href="https://edentns.com/">
77
  <img src="./Logo.png" alt="Logo" style="height: 3em;">
 
1
  ---
2
  tags:
3
+ - text-generation
4
  license: cc-by-nc-sa-4.0
5
  language:
6
+ - ko
7
  base_model: yanolja/KoSOLAR-10.7B-v0.1
8
  pipeline_tag: text-generation
9
+ datasets:
10
+ - nlpai-lab/kullm-v2
11
  ---
12
 
13
  # **DataVortexS-10.7B-v0.1**
 
 
 
14
 
15
+ <img src="./DataVortex.png" alt="DataVortex" style="height: 8em;">
16
 
17
  ## **Model Details**
18
 
19
  ### **Base Model**
20
+
21
+ [yanolja/KoSOLAR-10.7B-v0.1](https://huggingface.co/yanolja/KoSOLAR-10.7B-v0.1)
22
 
23
  ### **Trained On**
24
+
25
+ - **OS**: Ubuntu 20.04
26
+ - **GPU**: H100 80GB 1ea
27
+ - **transformers**: v4.36.2
28
+
29
+ ### **Dataset**
30
+
31
+ - [nlpai-lab/kullm-v2](https://huggingface.co/datasets/nlpai-lab/kullm-v2) - 152k rows
32
 
33
  ### **Instruction format**
34
 
35
+ It follows **Alpaca** format.
36
+
37
+ E.g.
38
+
39
+ ```python
40
+ text = """\
41
+ 당신은 μ‚¬λžŒλ“€μ΄ 정보λ₯Ό 찾을 수 μžˆλ„λ‘ λ„μ™€μ£ΌλŠ” 인곡지λŠ₯ λΉ„μ„œμž…λ‹ˆλ‹€.
42
+
43
+ ### Instruction:
44
+ λŒ€ν•œλ―Όκ΅­μ˜ μˆ˜λ„λŠ” μ–΄λ””μ•Ό?
45
+
46
+ ### Response:
47
+ λŒ€ν•œλ―Όκ΅­μ˜ μˆ˜λ„λŠ” μ„œμšΈμž…λ‹ˆλ‹€.
48
+
49
+ ### Instruction:
50
+ μ„œμšΈ μΈκ΅¬λŠ” 총 λͺ‡ λͺ…이야?
51
+ """
52
+ ```
53
 
54
  ## **Model Benchmark**
55
 
56
  ### **Ko-LLM-Leaderboard**
57
 
58
+ On Benchmarking ...
59
+
60
+ | Model | Average | Ko-ARC | Ko-HellaSwag | Ko-MMLU | Ko-TruthfulQA | Ko-CommonGen V2 |
61
+ | ------------------------------------ | ------- | ------ | ------------ | ------- | ------------- | --------------- |
62
+ | Edentns/DataVortexM-7B-Instruct-v0.1 | 39.81 | 34.13 | 42.35 | 38.73 | 45.46 | 38.37 |
63
+ | **Edentns/DataVortexS-10.7B-v0.1** | **0** | **0** | **0** | **0** | **0** | **0** |
64
 
65
+ ## **Implementation Code**
66
 
67
+ This model contains the chat_template instruction format.
68
  You can use the code below.
69
 
70
  ```python
71
  from transformers import AutoModelForCausalLM, AutoTokenizer
72
 
73
+ device = "cuda" # the device to load the model onto
74
 
75
+ model = AutoModelForCausalLM.from_pretrained("Edentns/DataVortexS-10.7B-v0.1")
76
  tokenizer = AutoTokenizer.from_pretrained("Edentns/DataVortexS-10.7B-v0.1")
77
 
78
  messages = [
79
+ {"role": "system", "content": "당신은 μ‚¬λžŒλ“€μ΄ 정보λ₯Ό 찾을 수 μžˆλ„λ‘ λ„μ™€μ£ΌλŠ” 인곡지λŠ₯ λΉ„μ„œμž…λ‹ˆλ‹€."},
80
+ {"role": "user", "content": "λŒ€ν•œλ―Όκ΅­μ˜ μˆ˜λ„λŠ” μ–΄λ””μ•Ό?"},
81
+ {"role": "assistant", "content": "λŒ€ν•œλ―Όκ΅­μ˜ μˆ˜λ„λŠ” μ„œμšΈμž…λ‹ˆλ‹€."},
82
+ {"role": "user", "content": "μ„œμšΈ μΈκ΅¬λŠ” 총 λͺ‡ λͺ…이야?"}
83
  ]
84
 
85
+ encodeds = tokenizer.apply_chat_template(messages, return_tensors="pt")
86
+
87
+ model_inputs = encodeds.to(device)
88
+ model.to(device)
89
+
90
+ generated_ids = model.generate(model_inputs, max_new_tokens=1000, do_sample=True)
91
+ decoded = tokenizer.batch_decode(generated_ids)
92
+ print(decoded[0])
 
 
 
 
 
 
 
 
 
 
 
 
93
  ```
94
 
95
+ ## **License**
96
+
97
+ The model is licensed under the [cc-by-nc-sa-4.0](https://creativecommons.org/licenses/by-nc-sa/4.0/) license, which allows others to copy, modify, and share the work non-commercially, as long as they give appropriate credit and distribute any derivative works under the same license.
98
+
99
  <div align="center">
100
  <a href="https://edentns.com/">
101
  <img src="./Logo.png" alt="Logo" style="height: 3em;">