Is it worth hosting a quantized DeepSeek V3? Cost & performance insights?
#15 opened 2 days ago
by
Techw
Can you add 1.58bit?
#14 opened 4 days ago
by
PSM24
Dynamic quants
2
#13 opened 13 days ago
by
XelotX
Core dumped when merging DeepSeek-V3-Q2_K_XS
1
#12 opened 21 days ago
by
MyJerry1996
First review, Q5-K-M require 502Gb RAM, better than Meta 405billions
6
#11 opened 29 days ago
by
krustik
Issue with --n-gpu-layers 5 Parameter: Model Only Running on CPU
12
#10 opened 29 days ago
by
vuk123
I’m new to GGUF quants
1
#9 opened 30 days ago
by
fsaudm
I loaded DeepSeek-V3-Q5_K_M up on my 10yrs old old Tesla M40 (Dell C4130)
3
#8 opened about 1 month ago
by
gng2info
why use q5 for key cache?
1
#7 opened about 1 month ago
by
CHNtentes
What is the required GPU size to run Is a 4090 possible and does it support ollama
10
#5 opened about 1 month ago
by
sminbb
I'm a newbie. How to use?
1
#4 opened about 1 month ago
by
huangkk
Getting error with Q3-K-M
7
#2 opened about 1 month ago
by
alain401
Are these imatrix GGUF quants?
4
#1 opened about 1 month ago
by
Kearm
![](https://cdn-avatars.huggingface.co/v1/production/uploads/655dc641accde1bbc8b41aec/9sR2Mm7mMsyh_SpSH7ilq.jpeg)