Adishah31's picture
Update README.md
123b9e6
---
library_name: peft
base_model: unsloth/mistral-7b-bnb-4bit
datasets:
- yahma/alpaca-cleaned
---
# Model Card for Model ID
A 4bit Mistral 7B model finetuned using unsloth on T4 GPU
## Model Details
### Model Description
- **Finetuned from model:** unsloth/mistral-7b-bnb-4bit
- **Repository:** https://github.com/unslothai/unsloth
## Training Details
### Training Data
https://huggingface.co/datasets/yahma/alpaca-cleaned
### Training Procedure
#### Preprocessing
Alpaca prompt template is used:
```
alpaca_prompt = """Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request.
### Instruction:
{}
### Input:
{}
### Response:
{}"""
```
#### Training Hyperparameters
```
per_device_train_batch_size = 2,
gradient_accumulation_steps = 4,
warmup_steps = 5,
max_steps = 60,
learning_rate = 2e-4,
fp16 = not torch.cuda.is_bf16_supported(),
bf16 = torch.cuda.is_bf16_supported(),
logging_steps = 1,
optim = "adamw_8bit",
weight_decay = 0.01,
lr_scheduler_type = "linear",
seed = 3407
```
- **Hardware Type:** T4 GPU
- **Cloud Provider:** Google Colab
### Framework versions
- PEFT 0.7.1