Edit model card

drawing

Trendyol LLM 7b base v0.1

Description

This repo contains GGUF format model files for Trendyol's Trendyol LLM 7b base v0.1

Quantization methods

quantization method bits size use case recommended
Q2_K 2 2.59 GB smallest, significant quality loss - not recommended for most purposes
Q3_K_S 3 3.01 GB very small, high quality loss
Q3_K_M 3 3.36 GB very small, high quality loss
Q3_K_L 3 3.66 GB small, substantial quality loss
Q4_0 4 3.9 GB legacy; small, very high quality loss - prefer using Q3_K_M
Q4_K_M 4 4.15 GB medium, balanced quality - recommended
Q5_0 5 4.73 GB legacy; medium, balanced quality - prefer using Q4_K_M
Q5_K_S 5 4.73 GB large, low quality loss - recommended
Q5_K_M 5 4.86 GB large, very low quality loss - recommended
Q6_K 6 5.61 GB very large, extremely low quality loss
Q8_0 8 13.7 GB very large, extremely low quality loss - not recommended
Downloads last month
115
GGUF
Model size
6.84B params
Architecture
llama

2-bit

3-bit

4-bit

5-bit

6-bit

Inference Examples
Inference API (serverless) has been turned off for this model.

Model tree for sayhan/Trendyol-LLM-7b-base-v0.1-GGUF

Quantized
(2)
this model

Collection including sayhan/Trendyol-LLM-7b-base-v0.1-GGUF