antebe1 commited on
Commit
4862e6b
·
verified ·
1 Parent(s): 7d35468

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +31 -0
README.md CHANGED
@@ -11,6 +11,37 @@ tags: []
11
 
12
  ## Model Details
13
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
14
  ### Model Description
15
 
16
  <!-- Provide a longer summary of what this model is. -->
 
11
 
12
  ## Model Details
13
 
14
+
15
+ peft_config = LoraConfig(
16
+ r=5, # TODO: play with this number
17
+ lora_alpha= 8, # TODO: play with this number
18
+ target_modules=['q_proj', 'v_proj', 'k_proj'],
19
+ lora_dropout=0.05,
20
+ bias="none",
21
+ task_type="CAUSAL_LM" # TODO: you need to figure this out. HINT https://github.com/huggingface/peft/blob/3d2bf9a8b261ed2960f26e61246cf0aa624a6115/src/peft/utils/peft_types.py#L67
22
+ )
23
+
24
+ training_args = TrainingArguments(
25
+ per_device_train_batch_size=2,
26
+ gradient_accumulation_steps=2,
27
+ gradient_checkpointing =False,
28
+ max_grad_norm= 0.3,
29
+ num_train_epochs=2, # TODO: play with this number
30
+ save_steps= 100,
31
+ learning_rate=0.0002, # TODO: play with this number
32
+ bf16=True,
33
+ save_total_limit=3,
34
+ logging_steps=10,
35
+ output_dir='./sft_models',
36
+ optim="adamw_torch",
37
+ lr_scheduler_type="cosine",
38
+ warmup_ratio=0.05,
39
+ remove_unused_columns=False,
40
+ report_to="none",
41
+ )
42
+
43
+ generate_max_length: int = 64
44
+
45
  ### Model Description
46
 
47
  <!-- Provide a longer summary of what this model is. -->