Uploaded model

This Model is trained with alpaca dataset max_seq_length = 8192 dtype = None load_in_4bit = False

warmup_steps = 10, max_steps = 70,

    learning_rate = 2e-5,
    fp16 = not is_bfloat16_supported(),
    bf16 = is_bfloat16_supported(),
    logging_steps = 1,
    optim = "adamw_8bit",
    weight_decay = 0.01,
    lr_scheduler_type = "linear",
    seed = 3407,
    output_dir = "outputs",
  • Developed by: Kaan35
  • License: apache-2.0
  • Finetuned from model : QWEN-2

Downloads last month
10
GGUF
Model size
7.62B params
Architecture
qwen2
Hardware compatibility
Log In to view the estimation

8-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Dataset used to train Kaan35/Qwen2_Alpaca_Lora