plaguss HF staff commited on
Commit
b867579
1 Parent(s): f285af2

Upload train-config.yaml with huggingface_hub

Browse files
Files changed (1) hide show
  1. train-config.yaml +66 -0
train-config.yaml ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ dataset_args:
2
+ path: argilla/10k_prompts_dpo
3
+
4
+ format_args:
5
+ prompt_format: zephyr
6
+
7
+ model_args:
8
+ pretrained_model_name_or_path: alignment-handbook/zephyr-7b-sft-full
9
+ torch_dtype: float16
10
+
11
+ peft_config:
12
+ r: 16
13
+ lora_alpha: 16
14
+ lora_dropout: 0.05
15
+ bias: none
16
+ task_type: CAUSAL_LM
17
+ target_modules:
18
+ - k_proj
19
+ - gate_proj
20
+ - v_proj
21
+ - up_proj
22
+ - q_proj
23
+ - o_proj
24
+ - down_proj
25
+
26
+ wandb_args:
27
+ entity: argilla-io
28
+ project: dibt-dpo
29
+ name: zephyr-7b-lora-dpo-dibt-v0
30
+
31
+ training_args:
32
+ bf16: true
33
+ beta: 0.1
34
+ loss_type: sigmoid
35
+ do_eval: true
36
+ do_train: true
37
+ evaluation_strategy: steps
38
+ eval_steps: 15
39
+ gradient_accumulation_steps: 2
40
+ gradient_checkpointing: true
41
+ gradient_checkpointing_kwargs:
42
+ use_reentrant: False
43
+ hub_model_id: plaguss/zephyr-7b-lora-dpo-dibt-v0
44
+ hub_model_revision: v0
45
+ hub_strategy: every_save
46
+ hub_private_repo: true
47
+ push_to_hub: true
48
+ learning_rate: 5.0e-7
49
+ logging_steps: 10
50
+ lr_scheduler_type: cosine
51
+ max_length: 1024
52
+ max_prompt_length: 512
53
+ num_train_epochs: 2
54
+ optim: paged_adamw_32bit
55
+ output_dir: data/zephyr-7b-sft-lora-dpo-v0
56
+ per_device_train_batch_size: 8
57
+ per_device_eval_batch_size: 8
58
+ save_strategy: epoch
59
+ save_total_limit: null
60
+ seed: 42
61
+ warmup_ratio: 0.1
62
+ report_to:
63
+ - wandb
64
+
65
+ use_accelerate: true
66
+ use_unsloth: false