csabakecskemeti commited on
Commit
95c0f7e
·
verified ·
1 Parent(s): e57cc7c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -1
README.md CHANGED
@@ -15,7 +15,8 @@ contains less than 100 tokens). The maximum token size for Orca2 is 4096 so a si
15
  (considering propt instructions) has been used. Chunking did not consider context (text data might split within the context).
16
  Evaluation set has been generated similar method on 1% of the raw data with LLama2 chat (https://huggingface.co/TheBloke/Llama-2-13B-chat-GGUF).
17
 
18
- Trained locally on 2x3090 GPU with vanila DDP with HuggingFace Accelerate for 50 Epoch.
 
19
 
20
 
21
  Chat with model sample code:
 
15
  (considering propt instructions) has been used. Chunking did not consider context (text data might split within the context).
16
  Evaluation set has been generated similar method on 1% of the raw data with LLama2 chat (https://huggingface.co/TheBloke/Llama-2-13B-chat-GGUF).
17
 
18
+ Trained locally on 2x3090 GPU with vanila DDP with HuggingFace Accelerate for 50 Epoch.
19
+ As I wanted to add new knowledge to the base model r=128 and lora_alpha=128 has been used -> LoRA weights are 3.5% of the base model.
20
 
21
 
22
  Chat with model sample code: