fujiki commited on
Commit
45086ce
1 Parent(s): 5778d0d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -5
README.md CHANGED
@@ -15,23 +15,23 @@ language:
15
  - ja
16
  ---
17
 
18
- # Japanese StableLM Base Gamma
19
 
20
  ## Model Description
21
 
22
  This is a 7B-parameter decoder-only language model with a focus on maximizing Japanese language modeling performance and Japanese downstream task performance.
23
  We conducted continued pretraining using Japanese data on the English language model, [Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1), to transfer the model's knowledge and capabilities to Japanese.
24
 
25
- *If you are looking for an instruction-following model, check [Japanese StableLM Instruct Gamma](https://huggingface.co/stabilityai/japanese-stablelm-instruct-gamma)*.
26
 
27
 
28
  ## Usage
29
 
30
  ```python
31
  from transformers import AutoModelForCausalLM, AutoTokenizer
32
- tokenizer = AutoTokenizer.from_pretrained("stabilityai/japanese-stablelm-base-gamma")
33
  model = AutoModelForCausalLM.from_pretrained(
34
- "stabilityai/japanese-stablelm-base-gamma",
35
  trust_remote_code=True,
36
  torch_dtype="auto",
37
  )
@@ -50,7 +50,7 @@ print(tokenizer.decode(tokens[0], skip_special_tokens=True))
50
  ## Model Details
51
 
52
  * **Developed by**: [Stability AI](https://stability.ai/)
53
- * **Model type**: `Japanese StableLM Base Gamma` model is an auto-regressive language model based on the transformer decoder architecture.
54
  * **Language(s)**: Japanese
55
  * **License**: This model is licensed under [Apache License, Version 2.0](https://www.apache.org/licenses/LICENSE-2.0).
56
  * **Contact**: For questions and comments about the model, please email `[email protected]`
 
15
  - ja
16
  ---
17
 
18
+ # Japanese StableLM Base Gamma 7B
19
 
20
  ## Model Description
21
 
22
  This is a 7B-parameter decoder-only language model with a focus on maximizing Japanese language modeling performance and Japanese downstream task performance.
23
  We conducted continued pretraining using Japanese data on the English language model, [Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1), to transfer the model's knowledge and capabilities to Japanese.
24
 
25
+ *If you are looking for an instruction-following model, check [Japanese StableLM Instruct Gamma 7B](https://huggingface.co/stabilityai/japanese-stablelm-instruct-gamma7b)*.
26
 
27
 
28
  ## Usage
29
 
30
  ```python
31
  from transformers import AutoModelForCausalLM, AutoTokenizer
32
+ tokenizer = AutoTokenizer.from_pretrained("stabilityai/japanese-stablelm-base-gamma-7b")
33
  model = AutoModelForCausalLM.from_pretrained(
34
+ "stabilityai/japanese-stablelm-base-gamma-7b",
35
  trust_remote_code=True,
36
  torch_dtype="auto",
37
  )
 
50
  ## Model Details
51
 
52
  * **Developed by**: [Stability AI](https://stability.ai/)
53
+ * **Model type**: `Japanese StableLM Base Gamma 7B` model is an auto-regressive language model based on the transformer decoder architecture.
54
  * **Language(s)**: Japanese
55
  * **License**: This model is licensed under [Apache License, Version 2.0](https://www.apache.org/licenses/LICENSE-2.0).
56
  * **Contact**: For questions and comments about the model, please email `[email protected]`