aashish1904's picture
Upload README.md with huggingface_hub
4fda491 verified
|
raw
history blame
2.12 kB
metadata
license: gpl-3.0
datasets:
  - Orion-zhen/kto-gutenberg
language:
  - zh
  - en
base_model:
  - Orion-zhen/Qwen2.5-7B-Instruct-Uncensored
pipeline_tag: text-generation

QuantFactory Banner

QuantFactory/Qwen2.5-7B-Gutenberg-KTO-GGUF

This is quantized version of Orion-zhen/Qwen2.5-7B-Gutenberg-KTO created using llama.cpp

Original Model Card

Qwen2.5-7B-Gutenberg-KTO

This model is fine tuned over gutenberg datasets using kto strategy. It's my first time to use kto strategy, and I'm not sure how the model actually performs.

Compared to those large companies which remove accessories such as charger and cables from packages, I have achieved real environment protection by truly reducing energy consumption, rather than shifting costs to consumers.

Checkout GGUF here: Orion-zhen/Qwen2.5-7B-Gutenberg-KTO-Q6_K-GGUF

Details

Platform

I randomly grabbed some rubbish from a second-hand market and built a PC

I carefully selected various dedicated hardwares and constructed an incomparable home server, which I entitled the Great Server:

  • CPU: Intel Core i3-4160
  • Memory: 8G DDR3, single channel
  • GPU: Tesla P4, TDP 75W, boasting its Eco friendly energy consumption
  • Disk: 1TB M.2 NVME, PCIe 4.0

Training

To practice the eco-friendly training, I utilized various methods, including adam-mini, qlora and unsloth, to minimize VRAM and energy usage, as well as accelerating training speed.

Train log

training_loss

training_eval_loss