|
|
|
--- |
|
|
|
license: gpl-3.0 |
|
datasets: |
|
- Orion-zhen/kto-gutenberg |
|
language: |
|
- zh |
|
- en |
|
base_model: |
|
- Orion-zhen/Qwen2.5-7B-Instruct-Uncensored |
|
pipeline_tag: text-generation |
|
|
|
--- |
|
|
|
[![QuantFactory Banner](https://lh7-rt.googleusercontent.com/docsz/AD_4nXeiuCm7c8lEwEJuRey9kiVZsRn2W-b4pWlu3-X534V3YmVuVc2ZL-NXg2RkzSOOS2JXGHutDuyyNAUtdJI65jGTo8jT9Y99tMi4H4MqL44Uc5QKG77B0d6-JfIkZHFaUA71-RtjyYZWVIhqsNZcx8-OMaA?key=xt3VSDoCbmTY7o-cwwOFwQ)](https://hf.co/QuantFactory) |
|
|
|
|
|
# QuantFactory/Qwen2.5-7B-Gutenberg-KTO-GGUF |
|
This is quantized version of [Orion-zhen/Qwen2.5-7B-Gutenberg-KTO](https://huggingface.co/Orion-zhen/Qwen2.5-7B-Gutenberg-KTO) created using llama.cpp |
|
|
|
# Original Model Card |
|
|
|
|
|
# Qwen2.5-7B-Gutenberg-KTO |
|
|
|
This model is fine tuned over gutenberg datasets using kto strategy. It's my first time to use kto strategy, and I'm not sure how the model actually performs. |
|
|
|
Compared to those large companies which remove accessories such as charger and cables from packages, I have achieved **real** environment protection by **truly** reducing energy consumption, rather than shifting costs to consumers. |
|
|
|
Checkout GGUF here: [Orion-zhen/Qwen2.5-7B-Gutenberg-KTO-Q6_K-GGUF](https://huggingface.co/Orion-zhen/Qwen2.5-7B-Gutenberg-KTO-Q6_K-GGUF) |
|
|
|
## Details |
|
|
|
### Platform |
|
|
|
~~I randomly grabbed some rubbish from a second-hand market and built a PC~~ |
|
|
|
I carefully selected various dedicated hardwares and constructed an incomparable home server, which I entitled the **Great Server**: |
|
|
|
- CPU: Intel Core i3-4160 |
|
- Memory: 8G DDR3, single channel |
|
- GPU: Tesla P4, TDP 75W, boasting its **Eco friendly energy consumption** |
|
- Disk: 1TB M.2 NVME, PCIe 4.0 |
|
|
|
### Training |
|
|
|
To practice the **eco-friendly training**, I utilized various methods, including adam-mini, qlora and unsloth, to minimize VRAM and energy usage, as well as accelerating training speed. |
|
|
|
- dataset: [Orion-zhen/kto-gutenberg](https://huggingface.co/datasets/Orion-zhen/kto-gutenberg) |
|
- epoch: 2 |
|
- gradient accumulation: 8 |
|
- batch size: 1 |
|
- KTO perf beta: 0.1 |
|
|
|
### Train log |
|
|
|
![training_loss](./training_loss.png) |
|
|
|
![training_eval_loss](./training_eval_loss.png) |
|
|