|
import sys |
|
import logging |
|
|
|
import datasets |
|
|
|
|
|
import torch |
|
import transformers |
|
from trl import SFTTrainer |
|
from transformers import AutoModelForCausalLM, AutoTokenizer, TrainingArguments, BitsAndBytesConfig |
|
|
|
""" |
|
A simple example on using SFTTrainer and Accelerate to finetune Phi-3 models. For |
|
a more advanced example, please follow HF alignment-handbook/scripts/run_sft.py. |
|
This example has utilized DeepSpeed ZeRO3 offload to reduce the memory usage. The |
|
script can be run on V100 or later generation GPUs. Here are some suggestions on |
|
futher reducing memory consumption: |
|
- reduce batch size |
|
- decrease lora dimension |
|
- restrict lora target modules |
|
Please follow these steps to run the script: |
|
1. Install dependencies: |
|
conda install -c conda-forge accelerate |
|
pip3 install -i https://pypi.org/simple/ bitsandbytes |
|
pip3 install peft transformers trl datasets |
|
pip3 install deepspeed |
|
2. Setup accelerate and deepspeed config based on the machine used: |
|
accelerate config |
|
Here is a sample config for deepspeed zero3: |
|
compute_environment: LOCAL_MACHINE |
|
debug: false |
|
deepspeed_config: |
|
gradient_accumulation_steps: 1 |
|
offload_optimizer_device: none |
|
offload_param_device: none |
|
zero3_init_flag: true |
|
zero3_save_16bit_model: true |
|
zero_stage: 3 |
|
distributed_type: DEEPSPEED |
|
downcast_bf16: 'no' |
|
enable_cpu_affinity: false |
|
machine_rank: 0 |
|
main_training_function: main |
|
mixed_precision: bf16 |
|
num_machines: 1 |
|
num_processes: 4 |
|
rdzv_backend: static |
|
same_network: true |
|
tpu_env: [] |
|
tpu_use_cluster: false |
|
tpu_use_sudo: false |
|
use_cpu: false |
|
3. check accelerate config: |
|
accelerate env |
|
4. Run the code: |
|
accelerate launch sample_finetune.py |
|
""" |
|
|
|
logger = logging.getLogger(__name__) |
|
|
|
|
|
|
|
|
|
|
|
training_config = { |
|
"bf16": True, |
|
"do_eval": False, |
|
"learning_rate": 5.0e-06, |
|
"log_level": "info", |
|
"logging_steps": 20, |
|
"logging_strategy": "steps", |
|
"lr_scheduler_type": "cosine", |
|
"num_train_epochs": 1, |
|
"max_steps": -1, |
|
"output_dir": "./checkpoint_dir", |
|
"overwrite_output_dir": True, |
|
"per_device_eval_batch_size": 4, |
|
"per_device_train_batch_size": 4, |
|
"remove_unused_columns": True, |
|
"save_steps": 100, |
|
"save_total_limit": 1, |
|
"seed": 0, |
|
"gradient_checkpointing": True, |
|
"gradient_checkpointing_kwargs":{"use_reentrant": False}, |
|
"gradient_accumulation_steps": 1, |
|
"warmup_ratio": 0.2, |
|
} |
|
|
|
peft_config = { |
|
"r": 16, |
|
"lora_alpha": 32, |
|
"lora_dropout": 0.05, |
|
"bias": "none", |
|
"task_type": "CAUSAL_LM", |
|
"target_modules": "all-linear", |
|
"modules_to_save": None, |
|
} |
|
train_conf = TrainingArguments(**training_config) |
|
peft_conf = LoraConfig(**peft_config) |
|
|
|
|
|
|
|
|
|
|
|
logging.basicConfig( |
|
format="%(asctime)s - %(levelname)s - %(name)s - %(message)s", |
|
datefmt="%Y-%m-%d %H:%M:%S", |
|
handlers=[logging.StreamHandler(sys.stdout)], |
|
) |
|
log_level = train_conf.get_process_log_level() |
|
logger.setLevel(log_level) |
|
datasets.utils.logging.set_verbosity(log_level) |
|
transformers.utils.logging.set_verbosity(log_level) |
|
transformers.utils.logging.enable_default_handler() |
|
transformers.utils.logging.enable_explicit_format() |
|
|
|
|
|
logger.warning( |
|
f"Process rank: {train_conf.local_rank}, device: {train_conf.device}, n_gpu: {train_conf.n_gpu}" |
|
+ f" distributed training: {bool(train_conf.local_rank != -1)}, 16-bits training: {train_conf.fp16}" |
|
) |
|
logger.info(f"Training/evaluation parameters {train_conf}") |
|
logger.info(f"PEFT parameters {peft_conf}") |
|
|
|
|
|
|
|
|
|
|
|
checkpoint_path = "microsoft/Phi-3-mini-4k-instruct" |
|
|
|
model_kwargs = dict( |
|
use_cache=False, |
|
trust_remote_code=True, |
|
attn_implementation="flash_attention_2", |
|
torch_dtype=torch.bfloat16, |
|
device_map=None |
|
) |
|
model = AutoModelForCausalLM.from_pretrained(checkpoint_path, **model_kwargs) |
|
tokenizer = AutoTokenizer.from_pretrained(checkpoint_path) |
|
tokenizer.model_max_length = 2048 |
|
tokenizer.pad_token = tokenizer.unk_token |
|
tokenizer.pad_token_id = tokenizer.convert_tokens_to_ids(tokenizer.pad_token) |
|
tokenizer.padding_side = 'right' |
|
|
|
print(model) |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|