RichardErkhov commited on
Commit
8bb4e2b
·
verified ·
1 Parent(s): aaaf351

uploaded readme

Browse files
Files changed (1) hide show
  1. README.md +179 -0
README.md ADDED
@@ -0,0 +1,179 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Quantization made by Richard Erkhov.
2
+
3
+ [Github](https://github.com/RichardErkhov)
4
+
5
+ [Discord](https://discord.gg/pvy7H8DZMG)
6
+
7
+ [Request more models](https://github.com/RichardErkhov/quant_request)
8
+
9
+
10
+ WizardCoder-1B-V1.0-dpo-beta-0.01 - AWQ
11
+ - Model creator: https://huggingface.co/AlekseyKorshuk/
12
+ - Original model: https://huggingface.co/AlekseyKorshuk/WizardCoder-1B-V1.0-dpo-beta-0.01/
13
+
14
+
15
+
16
+
17
+ Original model description:
18
+ ---
19
+ license: bigcode-openrail-m
20
+ base_model: WizardLM/WizardCoder-1B-V1.0
21
+ tags:
22
+ - axolotl
23
+ - dpo
24
+ - trl
25
+ - dpo
26
+ - generated_from_trainer
27
+ model-index:
28
+ - name: WizardCoder-1B-V1.0-dpo-beta-0.01
29
+ results: []
30
+ ---
31
+
32
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
33
+ should probably proofread and complete it, then remove this comment. -->
34
+
35
+ [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
36
+ <details><summary>See axolotl config</summary>
37
+
38
+ axolotl version: `0.4.0`
39
+ ```yaml
40
+ base_model: WizardLM/WizardCoder-1B-V1.0
41
+ model_type: AutoModelForCausalLM
42
+ tokenizer_type: AutoTokenizer
43
+ trust_remote_code: true
44
+
45
+ hub_model_id: AlekseyKorshuk/WizardCoder-1B-V1.0-dpo-beta-0.01
46
+ hub_strategy: every_save
47
+
48
+ load_in_8bit: false
49
+ load_in_4bit: false
50
+ strict: false
51
+
52
+ rl: dpo
53
+ datasets:
54
+ - path: AlekseyKorshuk/evol-codealpaca-v1-dpo
55
+ split: train
56
+ type: wizardcoder.intel
57
+
58
+
59
+ dataset_prepared_path:
60
+ #val_set_size: 0.001
61
+ output_dir: ./output
62
+
63
+ sequence_len: 2048
64
+ #sample_packing: false # currently unsupported
65
+ pad_to_sequence_len:
66
+
67
+ lora_r:
68
+ lora_alpha:
69
+ lora_dropout:
70
+ lora_target_modules:
71
+ lora_target_linear:
72
+ lora_fan_in_fan_out:
73
+
74
+ wandb_project: ui-thesis
75
+ wandb_entity:
76
+ wandb_watch:
77
+ wandb_name: ultrachat-stable-code-3b-dpo-chatml-beta-0.01
78
+ wandb_log_model:
79
+
80
+ gradient_accumulation_steps: 8
81
+ micro_batch_size: 4
82
+ num_epochs: 1
83
+ optimizer: paged_adamw_8bit
84
+ adam_beta1: 0.9
85
+ adam_beta2: 0.95
86
+ max_grad_norm: 1.0
87
+ adam_epsilon: 0.00001
88
+ lr_scheduler: cosine
89
+ cosine_min_lr_ratio: 0.1
90
+ learning_rate: 8.0e-7
91
+ warmup_steps: 32
92
+ #warmup_ratio: 0.1
93
+ weight_decay: 0.01
94
+ dpo_beta: 0.01
95
+
96
+ train_on_inputs: false
97
+ group_by_length: false
98
+ bf16: true
99
+ fp16: false
100
+ tf32: true
101
+ #float16: true
102
+
103
+ gradient_checkpointing: true
104
+ early_stopping_patience:
105
+ resume_from_checkpoint:
106
+ local_rank:
107
+ logging_steps: 1
108
+ xformers_attention:
109
+ flash_attention: false
110
+
111
+
112
+ #evals_per_epoch: 5
113
+ #eval_table_size: 8 # Approximate number of predictions sent to wandb depending on batch size. Enabled above 0. Default is 0
114
+ #eval_table_max_new_tokens: 768 # Total number of tokens generated for predictions sent to wandb. Default is 128
115
+
116
+ #chat_template: chatml
117
+ #saves_per_epoch: 1
118
+ save_steps: 500
119
+ save_total_limit: 1
120
+ seed: 42
121
+ debug:
122
+ deepspeed:
123
+
124
+
125
+ fsdp:
126
+ fsdp_config:
127
+ resize_token_embeddings_to_32x: true
128
+
129
+ ```
130
+
131
+ </details><br>
132
+
133
+ # WizardCoder-1B-V1.0-dpo-beta-0.01
134
+
135
+ This model is a fine-tuned version of [WizardLM/WizardCoder-1B-V1.0](https://huggingface.co/WizardLM/WizardCoder-1B-V1.0) on the None dataset.
136
+
137
+ ## Model description
138
+
139
+ More information needed
140
+
141
+ ## Intended uses & limitations
142
+
143
+ More information needed
144
+
145
+ ## Training and evaluation data
146
+
147
+ More information needed
148
+
149
+ ## Training procedure
150
+
151
+ ### Training hyperparameters
152
+
153
+ The following hyperparameters were used during training:
154
+ - learning_rate: 8e-07
155
+ - train_batch_size: 4
156
+ - eval_batch_size: 8
157
+ - seed: 42
158
+ - distributed_type: multi-GPU
159
+ - num_devices: 4
160
+ - gradient_accumulation_steps: 8
161
+ - total_train_batch_size: 128
162
+ - total_eval_batch_size: 32
163
+ - optimizer: Adam with betas=(0.9,0.95) and epsilon=1e-05
164
+ - lr_scheduler_type: cosine
165
+ - lr_scheduler_warmup_steps: 32
166
+ - training_steps: 312
167
+
168
+ ### Training results
169
+
170
+
171
+
172
+ ### Framework versions
173
+
174
+ - Transformers 4.38.0.dev0
175
+ - Pytorch 2.1.2+cu118
176
+ - Datasets 2.16.1
177
+ - Tokenizers 0.15.0
178
+
179
+