-
-
-
-
-
-
Inference status
Active filters:
gptq
TheBloke/Mistral-7B-Claude-Chat-GPTQ
Text Generation
•
Updated
•
34
•
11
TheBloke/Nous-Capybara-34B-GPTQ
Text Generation
•
Updated
•
40
•
20
TheBloke/TimeCrystal-L2-13B-GPTQ
Text Generation
•
Updated
•
18
•
6
Qwen/Qwen-72B-Chat-Int4
Text Generation
•
Updated
•
249
•
46
TheBloke/dolphin-2.5-mixtral-8x7b-GPTQ
Text Generation
•
Updated
•
97
•
107
TheBloke/DaringMaid-13B-GPTQ
Text Generation
•
Updated
•
44
•
7
TheBloke/Noromaid-v0.1-mixtral-8x7b-v3-GPTQ
Text Generation
•
Updated
•
118
•
10
TheBloke/TinyLlama-1.1B-Chat-v1.0-GPTQ
Text Generation
•
Updated
•
38.7k
•
13
TheBloke/toxicqa-Llama2-13B-GPTQ
TheBloke/dolphin-2.7-mixtral-8x7b-GPTQ
Text Generation
•
Updated
•
82
•
19
TheBloke/Etheria-55b-v0.1-GPTQ
Text Generation
•
Updated
•
46
•
4
TheBloke/CapybaraHermes-2.5-Mistral-7B-GPTQ
Updated
•
4.61k
•
56
Qwen/Qwen1.5-14B-Chat-GPTQ-Int4
Text Generation
•
Updated
•
208
•
20
Qwen/Qwen1.5-0.5B-Chat-GPTQ-Int8
Text Generation
•
Updated
•
69
•
3
Intel/gemma-2b-int4-inc
Text Generation
•
Updated
•
25
•
1
explodinggradients/Ragas-critic-llm-Qwen1.5-GPTQ
Text Generation
•
Updated
•
15
•
11
neuralmagic/Mistral-7B-Instruct-v0.3-GPTQ-4bit
Text Generation
•
Updated
•
2.18k
•
17
Intel/Qwen2-1.5B-Instuct-int4-inc
Text Generation
•
Updated
•
21
•
1
neuralmagic/Meta-Llama-3-8B-Instruct-quantized.w8a16
Text Generation
•
Updated
•
7.97k
•
3
openerotica/WizardLM-2-8x22B-GPTQ-ERQ
Text Generation
•
Updated
•
26
•
1
pentagoniac/SEMIKONG-8b-GPTQ
Text Generation
•
Updated
•
661
•
23
shuyuej/gemma-2-27b-it-GPTQ
Text Generation
•
Updated
•
623
•
3
shuyuej/Meta-Llama-3.1-8B-Instruct-GPTQ
Text Generation
•
Updated
•
126
•
1
hugging-quants/Meta-Llama-3.1-70B-Instruct-GPTQ-INT4
Text Generation
•
Updated
•
25.2k
•
22
hugging-quants/Meta-Llama-3.1-8B-Instruct-GPTQ-INT4
Text Generation
•
Updated
•
76.8k
•
21
TechxGenus/Mistral-Large-Instruct-2407-GPTQ
Text Generation
•
Updated
•
49
•
3
shuyuej/bge-en-icl-GPTQ
Feature Extraction
•
Updated
•
21
•
2
shuyuej/gte-Qwen2-7B-instruct-GPTQ
Text Generation
•
Updated
•
65
•
2
nejumi/calm3-22b-chat-GPTQ-Int4-calib-ja-1k
Text Generation
•
Updated
•
420
•
2
team-hatakeyama-phase2/Tanuki-8x8B-dpo-v1.0-GPTQ-4bit
Text Generation
•
Updated
•
14.4k
•
3