Training procedure

The following bitsandbytes quantization config was used during training:

  • quant_method: bitsandbytes
  • load_in_8bit: False
  • load_in_4bit: True
  • llm_int8_threshold: 6.0
  • llm_int8_skip_modules: None
  • llm_int8_enable_fp32_cpu_offload: False
  • llm_int8_has_fp16_weight: False
  • bnb_4bit_quant_type: nf4
  • bnb_4bit_use_double_quant: False
  • bnb_4bit_compute_dtype: float16

Framework versions

Fine-tuning Llama-2-7b using QLoRA in French on Google Colab

Goal

The goal of this project is to adapt the Llama-2-7b model, which initially might not have proficiency in French, to understand and respond accurately to queries in the French language. This adaptation involves fine-tuning the model on a dataset of French novels, allowing it to comprehend the nuances, syntax, and semantics of the French language. By leveraging the PEFT library from the Hugging Face ecosystem and QLoRA for more memory-efficient fine-tuning on a single T4 GPU provided by Google Colab, we aim to create a chatbot that can effectively answer questions posed in French.

Overview

This project involves several steps including setting up the environment, loading the dataset and model, configuring QLoRA and training parameters, training the model, and finally testing and pushing the fine-tuned model to Hugging Face.

Features

  • Dataset Loading: Load and process a French novels dataset using Hugging Face datasets library.
  • Model Quantization: Quantize the base Llama-2-7b model into 4-bit using bitsandbytes.
  • Configuration for QLoRA: Apply the QLoRA configuration for more memory-efficient fine-tuning using the PEFT library.
  • Training: Use the SFTTrainer from the TRL library for instruction-based fine-tuning.
  • Testing and Pushing to Hugging Face: Test the fine-tuned model and push it to Hugging Face.

Prerequisites

  • Google Colab with T4 GPU

  • Python libraries: trl, transformers, accelerate, peft, datasets, bitsandbytes, einops

Downloads last month
2
Inference API
Unable to determine this model’s pipeline type. Check the docs .

Model tree for Omid-sar/fine-tuning-llama2-7b-qlora-french

Adapter
(1761)
this model