Merge pull request #350 from tmm1/group-len-false-examples
Browse files- README.md +3 -1
- examples/cerebras/qlora.yml +1 -1
- examples/gptj/qlora.yml +1 -1
- examples/llama-2/lora.yml +1 -1
- examples/llama-2/qlora.yml +1 -1
- examples/openllama-3b/qlora.yml +1 -1
README.md
CHANGED
@@ -426,7 +426,9 @@ save_safetensors:
|
|
426 |
|
427 |
# whether to mask out or include the human's prompt from the training labels
|
428 |
train_on_inputs: false
|
429 |
-
#
|
|
|
|
|
430 |
group_by_length: false
|
431 |
|
432 |
# Whether to use gradient checkpointing https://huggingface.co/docs/transformers/v4.18.0/en/performance#gradient-checkpointing
|
|
|
426 |
|
427 |
# whether to mask out or include the human's prompt from the training labels
|
428 |
train_on_inputs: false
|
429 |
+
# group similarly sized data to minimize padding
|
430 |
+
# may be slower to start, as it must download and sort the entire dataset
|
431 |
+
# note that training loss may have an oscillating pattern with this enabled
|
432 |
group_by_length: false
|
433 |
|
434 |
# Whether to use gradient checkpointing https://huggingface.co/docs/transformers/v4.18.0/en/performance#gradient-checkpointing
|
examples/cerebras/qlora.yml
CHANGED
@@ -35,7 +35,7 @@ torchdistx_path:
|
|
35 |
lr_scheduler: cosine
|
36 |
learning_rate: 0.0002
|
37 |
train_on_inputs: false
|
38 |
-
group_by_length:
|
39 |
bf16: true
|
40 |
fp16: false
|
41 |
tf32: true
|
|
|
35 |
lr_scheduler: cosine
|
36 |
learning_rate: 0.0002
|
37 |
train_on_inputs: false
|
38 |
+
group_by_length: false
|
39 |
bf16: true
|
40 |
fp16: false
|
41 |
tf32: true
|
examples/gptj/qlora.yml
CHANGED
@@ -32,7 +32,7 @@ torchdistx_path:
|
|
32 |
lr_scheduler: cosine
|
33 |
learning_rate: 0.0001
|
34 |
train_on_inputs: false
|
35 |
-
group_by_length:
|
36 |
bf16: true
|
37 |
fp16: false
|
38 |
tf32: true
|
|
|
32 |
lr_scheduler: cosine
|
33 |
learning_rate: 0.0001
|
34 |
train_on_inputs: false
|
35 |
+
group_by_length: false
|
36 |
bf16: true
|
37 |
fp16: false
|
38 |
tf32: true
|
examples/llama-2/lora.yml
CHANGED
@@ -38,7 +38,7 @@ lr_scheduler: cosine
|
|
38 |
learning_rate: 0.0002
|
39 |
|
40 |
train_on_inputs: false
|
41 |
-
group_by_length:
|
42 |
bf16: true
|
43 |
fp16: false
|
44 |
tf32: false
|
|
|
38 |
learning_rate: 0.0002
|
39 |
|
40 |
train_on_inputs: false
|
41 |
+
group_by_length: false
|
42 |
bf16: true
|
43 |
fp16: false
|
44 |
tf32: false
|
examples/llama-2/qlora.yml
CHANGED
@@ -39,7 +39,7 @@ lr_scheduler: cosine
|
|
39 |
learning_rate: 0.0002
|
40 |
|
41 |
train_on_inputs: false
|
42 |
-
group_by_length:
|
43 |
bf16: true
|
44 |
fp16: false
|
45 |
tf32: false
|
|
|
39 |
learning_rate: 0.0002
|
40 |
|
41 |
train_on_inputs: false
|
42 |
+
group_by_length: false
|
43 |
bf16: true
|
44 |
fp16: false
|
45 |
tf32: false
|
examples/openllama-3b/qlora.yml
CHANGED
@@ -34,7 +34,7 @@ torchdistx_path:
|
|
34 |
lr_scheduler: cosine
|
35 |
learning_rate: 0.0002
|
36 |
train_on_inputs: false
|
37 |
-
group_by_length:
|
38 |
bf16: true
|
39 |
fp16: false
|
40 |
tf32: true
|
|
|
34 |
lr_scheduler: cosine
|
35 |
learning_rate: 0.0002
|
36 |
train_on_inputs: false
|
37 |
+
group_by_length: false
|
38 |
bf16: true
|
39 |
fp16: false
|
40 |
tf32: true
|