Uploaded model
- Developed by: ruslandev
- License: apache-2.0
- Finetuned from model : unsloth/llama-3-8b-bnb-4bit
This model is finetuned on the data of Samantha.
Prompt format is Alpaca. I used the same system prompt as the original Samantha.
"""Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request.
### Instruction:
{SYSTEM_PROMPT}
### Input:
{QUESTION}
### Response:
"""
Training
gptchain framework has been used for training.
Training hyperparameters
- learning_rate: 2e-4
- seed: 3407
- gradient_accumulation_steps: 4
- per_device_train_batch_size: 2
- optimizer: adamw_8bit
- lr_scheduler_type: linear
- warmup_steps: 5
- num_epochs: 2
- weight_decay: 0.01
Training results
Training Loss | Epoch | Step |
---|---|---|
2.0778 | 0.0 | 1 |
0.6255 | 0.18 | 120 |
0.6208 | 0.94 | 620 |
0.6244 | 2.0 | 1306 |
2 epoch finetuning from llama-3-8b took 1 hour on a single A100 with Unsloth and Huggingface's TRL library.
- Downloads last month
- 47
Inference Providers
NEW
This model is not currently available via any of the supported third-party Inference Providers, and
the model is not deployed on the HF Inference API.
Model tree for ruslandev/llama-3-8b-samantha
Base model
meta-llama/Meta-Llama-3-8B
Quantized
unsloth/llama-3-8b-bnb-4bit