-
-
-
-
-
-
Inference status
Active filters:
vllm
FlorianJc/Mistral-Nemo-Instruct-2407-vllm-fp8
Text Generation
•
Updated
•
35.2k
•
8
neuralmagic/DeepSeek-Coder-V2-Base-FP8
Text Generation
•
Updated
•
12
neuralmagic/DeepSeek-Coder-V2-Instruct-FP8
Text Generation
•
Updated
•
1.63k
•
7
mgoin/Minitron-4B-Base-FP8
Text Generation
•
Updated
•
748
•
3
mgoin/Minitron-8B-Base-FP8
Text Generation
•
Updated
•
22
•
3
mgoin/nemotron-3-8b-chat-4k-sft-hf
Text Generation
•
Updated
•
70
neuralmagic/Meta-Llama-3.1-8B-Instruct-FP8
Text Generation
•
Updated
•
370k
•
37
neuralmagic/Meta-Llama-3.1-8B-Instruct-FP8-dynamic
Text Generation
•
Updated
•
1.79k
•
5
neuralmagic/Meta-Llama-3.1-405B-Instruct-FP8
Text Generation
•
Updated
•
1.14k
•
31
neuralmagic/Meta-Llama-3.1-405B-Instruct-FP8-dynamic
Text Generation
•
Updated
•
201
•
14
neuralmagic/Meta-Llama-3.1-8B-Instruct-quantized.w8a16
Text Generation
•
Updated
•
11.3k
•
9
mgoin/Nemotron-4-340B-Base-hf
Text Generation
•
Updated
•
15
•
1
mgoin/Nemotron-4-340B-Base-hf-FP8
Text Generation
•
Updated
•
29
•
2
neuralmagic/Meta-Llama-3.1-70B-Instruct-quantized.w8a16
Text Generation
•
Updated
•
233
•
4
mgoin/Nemotron-4-340B-Instruct-hf
Text Generation
•
Updated
•
56
•
4
mgoin/Nemotron-4-340B-Instruct-hf-FP8
Text Generation
•
Updated
•
170
•
3
FlorianJc/ghost-8b-beta-vllm-fp8
Text Generation
•
Updated
•
12
FlorianJc/Meta-Llama-3.1-8B-Instruct-vllm-fp8
Text Generation
•
Updated
•
216
neuralmagic/Meta-Llama-3.1-8B-Instruct-quantized.w4a16
Text Generation
•
Updated
•
14.9k
•
23
neuralmagic/Meta-Llama-3.1-8B-FP8
Text Generation
•
Updated
•
2.34k
•
5
neuralmagic/Meta-Llama-3.1-70B-FP8
Text Generation
•
Updated
•
345
•
1
neuralmagic/Meta-Llama-3.1-8B-quantized.w8a16
Text Generation
•
Updated
•
4.89k
•
1
neuralmagic/Meta-Llama-3.1-70B-Instruct-quantized.w4a16
Text Generation
•
Updated
•
17.5k
•
28
neuralmagic/starcoder2-15b-FP8
Text Generation
•
Updated
•
730
neuralmagic/starcoder2-7b-FP8
Text Generation
•
Updated
•
13
neuralmagic/starcoder2-3b-FP8
Text Generation
•
Updated
•
16
neuralmagic/Meta-Llama-3.1-405B-FP8
Text Generation
•
Updated
•
13
bprice9/Palmyra-Medical-70B-FP8
Text Generation
•
Updated
•
13
neuralmagic/gemma-2-2b-it-FP8
Updated
•
2.8k
•
1
neuralmagic/Meta-Llama-3.1-405B-Instruct-quantized.w4a16
Text Generation
•
Updated
•
976
•
12