Model Description

This model is a fine-tuned version of unsloth/Meta-Llama-3.2-3B optimized for Prompt Generation tasks when given a act. The fine-tuning was done using the Unsloth library with LoRA (Low-Rank Adaptation) for parameter-efficient fine-tuning. The training was done on fka/awesome-chatgpt-prompts dataset.

  • Developed by: Vedant Rajpurohit
  • Model type: Causal Language Model
  • Language(s): English
  • Fine-tuned from model: unsloth/Meta-Llama-3.2-3B
  • Precision: F32

Direct Use


# !pip install bitsandbytes peft

from transformers import AutoModelForCausalLM, AutoTokenizer
from peft import PeftModel

# Load the tokenizer for the base model
tokenizer = AutoTokenizer.from_pretrained("Vedant3907/Prompt-Generator-Lora-model", use_fast=False)

# Load the base model in 4-bit quantization mode
base_model = AutoModelForCausalLM.from_pretrained(
    "Vedant3907/Prompt-Generator-Lora-model",
    # load_in_4bit=True,
    trust_remote_code=True
)

gpt_prompt = """
### Instruction:
{}

### Response:
{}"""

inputs = tokenizer(
[
    gpt_prompt.format(
        "Rapper", # instruction
        "", # output - leave this blank for generation!
    )
], return_tensors = "pt").to("cuda")

outputs = base_model.generate(**inputs, max_new_tokens = 200, use_cache = True)
tokenizer.batch_decode(outputs)


"""
'<|begin_of_text|>

### Instruction:
  Rapper

### Response:
  I want you to act as a rapper. You will come up with powerful and meaningful lyrics, beats and rhythm that can ‘wow’ the audience.
  Your lyrics should have an intriguing meaning and message that people can relate too. When it comes to choosing your beat,
  make sure it is catchy yet relevant to your words, so that when combined they make an explosion of sound everytime!
  My first request is "I need a rap song about finding strength within yourself."
<|end_of_text|>'
"""

Training Details

Training Procedure

The model was fine-tuned using the Unsloth library with LoRA adapters, enabling efficient training. Below are the hyperparameters used:

args = TrainingArguments(
        per_device_train_batch_size = 2,
        gradient_accumulation_steps = 4,
        warmup_steps = 5,
        num_train_epochs = 8, 
        # max_steps = 60,
        learning_rate = 2e-4,
        fp16 = not is_bfloat16_supported(),
        bf16 = is_bfloat16_supported(),
        logging_steps = 1,
        optim = "adamw_8bit",
        weight_decay = 0.01,
        lr_scheduler_type = "linear",
        seed = 3407,
        output_dir = "outputs",
        report_to = "none",
    )

Hardware

  • Trained on google colab with its T4 GPU
Downloads last month

-

Downloads are not tracked for this model. How to track
Safetensors
Model size
195M params
Tensor type
F32
·
Inference Examples
Unable to determine this model's library. Check the docs .

Model tree for Vedant3907/Prompt-Generator

Finetuned
(34)
this model

Dataset used to train Vedant3907/Prompt-Generator