Zheng Han
traphix
·
AI & ML interests
None yet
Recent Activity
new activity
5 days ago
Qwen/QwQ-32B:This model beats Qwen Max!
new activity
6 days ago
Qwen/Qwen2.5-14B-Instruct-1M:Does vllm 0.7.3 support this model?
Organizations
None yet
traphix's activity
This model beats Qwen Max!
4
#33 opened 7 days ago
by
MrDevolver

Does vllm 0.7.3 support this model?
#10 opened 6 days ago
by
traphix
Is there any accuracy results comparing to original DeepSeek-V3?
#6 opened 20 days ago
by
traphix
why "MLA is not supported with awq_marlin quantization. Disabling MLA." with 4090 * 32 (4 node / vllm 0.7.2)
3
#14 opened 21 days ago
by
FightLLM
Is there any accuracy results comparing to original DeepSeek-R1?
2
#15 opened 20 days ago
by
traphix
Has anyone evaluated the performance of the AWQ version of the model on benchmarks?
4
#8 opened 28 days ago
by
liuqianchao
skips the thinking process
11
#5 opened about 1 month ago
by
muzizon
Deployment framework
27
#2 opened about 2 months ago
by
xro7
vllm support a100
17
#2 opened about 2 months ago
by
HuggingLianWang
Any plans to quantize Qwen/Qwen2.5-72B-Instruct to w8a8?
#1 opened about 1 month ago
by
traphix
Can it run on A100/A800 with VLLM?
3
#1 opened 8 months ago
by
Parkerlambert123
Quantize DeepSeek-Coder-V2-Instruct to W8A8(INT8)?
#2 opened 7 months ago
by
traphix