|
--- |
|
library_name: peft |
|
base_model: meta-llama/Llama-2-7b-hf |
|
--- |
|
|
|
# Model Details |
|
|
|
- Supervised fine-tuning (sft) based on [meta-llama/Llama-2-7b-hf](https://huggingface.co/meta-llama/Llama-2-7b-hf) on [yahma/alpaca-cleaned](https://huggingface.co/datasets/yahma/alpaca-cleaned) |
|
- Trained with Deepspeed ZeRO-1 + TRL + QLoRA + Flash-Attntion 2 within 1h with 3090x4 |
|
- The LoRa adapter is uploaded |
|
|
|
## Model and Training Details |
|
|
|
- **Finetuned from model:** [meta-llama/Llama-2-7b-hf](https://huggingface.co/meta-llama/Llama-2-7b-hf) |
|
|
|
- **Dataset:** [yahma/alpaca-cleaned](https://huggingface.co/datasets/yahma/alpaca-cleaned) |
|
|
|
#### Preprocessing |
|
|
|
- preprocessed and packed the sft dataset with trl.trainer.ConstantLengthDataset |
|
|
|
### Results |
|
|
|
![image/png](https://cdn-uploads.huggingface.co/production/uploads/65b1dd2a855f6b5fe621bc0e/QayTbFCmbt4N5jxnj3Sna.png) |
|
|
|
### Compute Infrastructure |
|
|
|
The model is trained using 4 * RTX 3090 - 24GB |
|
|
|
## Model Card Authors |
|
|
|
Yiyu (Michael) Ren |
|
|
|
## Model Card Contact |
|
|
|
Email: [email protected] |
|
|
|
### Framework versions |
|
|
|
- PEFT 0.8.2 |