--- library_name: peft base_model: meta-llama/Llama-2-7b-hf --- # Model Details - Supervised fine-tuning (sft) based on [meta-llama/Llama-2-7b-hf](https://huggingface.co/meta-llama/Llama-2-7b-hf) on [yahma/alpaca-cleaned](https://huggingface.co/datasets/yahma/alpaca-cleaned) - Trained with Deepspeed ZeRO-1 + TRL + QLoRA + Flash-Attntion 2 within 1h with 3090x4 - The LoRa adapter is uploaded ## Model and Training Details - **Finetuned from model:** [meta-llama/Llama-2-7b-hf](https://huggingface.co/meta-llama/Llama-2-7b-hf) - **Dataset:** [yahma/alpaca-cleaned](https://huggingface.co/datasets/yahma/alpaca-cleaned) #### Preprocessing - preprocessed and packed the sft dataset with trl.trainer.ConstantLengthDataset ### Results ![image/png](https://cdn-uploads.huggingface.co/production/uploads/65b1dd2a855f6b5fe621bc0e/QayTbFCmbt4N5jxnj3Sna.png) ### Compute Infrastructure The model is trained using 4 * RTX 3090 - 24GB ## Model Card Authors Yiyu (Michael) Ren ## Model Card Contact Email: renyiyuap@gmail.com ### Framework versions - PEFT 0.8.2