Spaces:
Runtime error
Runtime error
Added adapter merging to the fine-tuning script generation
Browse files
src/training/prep_finetuning.py
CHANGED
@@ -232,7 +232,7 @@ def generate_training_scripts(out_model: str, hf_user: str, hf_token: str) -> No
|
|
232 |
f.write("# DO NOT COMMIT THIS FILE TO GIT AS IT CONTAINS THE HUGGING FACE WRITE TOKEN FOR THE REPO\n\n")
|
233 |
f.write('echo "STARTING TRAINING AND PUSH TO HUB"\n')
|
234 |
f.write("start=$(date +%s)\n")
|
235 |
-
f.write(f"autotrain llm --train --project-name {out_model} --model meta-llama/Llama-2-7b-chat-hf --data-path . --peft --lr 2e-4 --batch-size 12 --epochs 3 --trainer sft --push-to-hub --username {hf_user} --token {hf_token}\n")
|
236 |
f.write("end=$(date +%s)\n")
|
237 |
f.write('echo "TRAINING AND PUSH TOOK $(($end-$start)) seconds"')
|
238 |
stats = os.stat(shell_file)
|
|
|
232 |
f.write("# DO NOT COMMIT THIS FILE TO GIT AS IT CONTAINS THE HUGGING FACE WRITE TOKEN FOR THE REPO\n\n")
|
233 |
f.write('echo "STARTING TRAINING AND PUSH TO HUB"\n')
|
234 |
f.write("start=$(date +%s)\n")
|
235 |
+
f.write(f"autotrain llm --train --project-name {out_model} --model meta-llama/Llama-2-7b-chat-hf --data-path . --peft --lr 2e-4 --batch-size 12 --epochs 3 --trainer sft --merge-adapters --push-to-hub --username {hf_user} --token {hf_token}\n")
|
236 |
f.write("end=$(date +%s)\n")
|
237 |
f.write('echo "TRAINING AND PUSH TOOK $(($end-$start)) seconds"')
|
238 |
stats = os.stat(shell_file)
|