llama-2-7b-sft-lora / README.md
renyiyu's picture
Update README.md
748db81 verified
---
library_name: peft
base_model: meta-llama/Llama-2-7b-hf
---
# Model Details
- Supervised fine-tuning (sft) based on [meta-llama/Llama-2-7b-hf](https://huggingface.co/meta-llama/Llama-2-7b-hf) on [yahma/alpaca-cleaned](https://huggingface.co/datasets/yahma/alpaca-cleaned)
- Trained with Deepspeed ZeRO-1 + TRL + QLoRA + Flash-Attntion 2 within 1h with 3090x4
- The LoRa adapter is uploaded
## Model and Training Details
- **Finetuned from model:** [meta-llama/Llama-2-7b-hf](https://huggingface.co/meta-llama/Llama-2-7b-hf)
- **Dataset:** [yahma/alpaca-cleaned](https://huggingface.co/datasets/yahma/alpaca-cleaned)
#### Preprocessing
- preprocessed and packed the sft dataset with trl.trainer.ConstantLengthDataset
### Results
![image/png](https://cdn-uploads.huggingface.co/production/uploads/65b1dd2a855f6b5fe621bc0e/QayTbFCmbt4N5jxnj3Sna.png)
### Compute Infrastructure
The model is trained using 4 * RTX 3090 - 24GB
## Model Card Authors
Yiyu (Michael) Ren
## Model Card Contact
Email: [email protected]
### Framework versions
- PEFT 0.8.2