hungeni commited on
Commit
0f2ee3d
1 Parent(s): bebdb8e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -16,14 +16,14 @@ tags:
16
 
17
  InfCapital LLama2-7b clone of [Meta's Llama 2 7B Chat](https://huggingface.co/meta-llama/Llama-2-7b-chat-hf).
18
 
19
- Adopt for Vietnamese continued pretraining or fine-tuning by extend vocab size from 32,000 to 48,281
20
 
21
 
22
  ## Model Architecture
23
  ```
24
  LlamaForCausalLM(
25
  (model): LlamaModel(
26
- (embed_tokens): Embedding(32000, 4096, padding_idx=0)
27
  (layers): ModuleList(
28
  (0-31): 32 x LlamaDecoderLayer(
29
  (self_attn): LlamaAttention(
@@ -45,6 +45,6 @@ LlamaForCausalLM(
45
  )
46
  (norm): LlamaRMSNorm()
47
  )
48
- (lm_head): Linear(in_features=4096, out_features=32000, bias=False)
49
  )
50
  ```
 
16
 
17
  InfCapital LLama2-7b clone of [Meta's Llama 2 7B Chat](https://huggingface.co/meta-llama/Llama-2-7b-chat-hf).
18
 
19
+ Adopt for Vietnamese continued pretraining or fine-tuning by extend vocab size from 32,000 to 44,800. Vocabs added by training sentencepiece method from dataset vnnews-corpus.
20
 
21
 
22
  ## Model Architecture
23
  ```
24
  LlamaForCausalLM(
25
  (model): LlamaModel(
26
+ (embed_tokens): Embedding(44800, 4096)
27
  (layers): ModuleList(
28
  (0-31): 32 x LlamaDecoderLayer(
29
  (self_attn): LlamaAttention(
 
45
  )
46
  (norm): LlamaRMSNorm()
47
  )
48
+ (lm_head): Linear(in_features=4096, out_features=44800, bias=False)
49
  )
50
  ```