ElliotGao
tclf90
AI & ML interests
None yet
Recent Activity
new activity 5 days ago
QuantTrio/GLM-5-AWQ:[Request] Great work! Do you have plans to also create GLM-5.1-AWQ? new activity 5 days ago
QuantTrio/Qwen3.5-122B-A10B-AWQ:CUDA version 13? new activity 5 days ago
QuantTrio/gemma-4-31B-it-AWQ:Request for awq of the gemma 4 26B A4B MoEOrganizations
[Request] Great work! Do you have plans to also create GLM-5.1-AWQ?
🤗 1
4
#6 opened 5 days ago
by
ag1988
CUDA version 13?
1
#1 opened 7 days ago
by
pathosethoslogos
Request for awq of the gemma 4 26B A4B MoE
5
#1 opened 8 days ago
by
rks2302
AWQ 4/5/6-bit request for Qwopus3.5-27B-v3
🚀❤️ 3
3
#2 opened 11 days ago
by
0xburakcelik
AWQ 4-bit version of this Opus-Distilled-v2 model?
9
#5 opened 15 days ago
by
0xburakcelik
--max-model-len 32768 seems a bit too small for agent use cases ?
3
#3 opened about 1 month ago
by
edwarddukewu
Install & run QuantTrio/MiniMax-M2-AWQ easily using llmpm
👍 1
1
#8 opened about 1 month ago
by
sarthak-saxena
My personal vLLM launch cmd on my old personal 2x3090 workstation
7
#1 opened about 1 month ago
by
tclf90
Can't get vLLM running on 1xRTX 4090
3
#1 opened about 2 months ago
by
slyfox1186
Easy to fall into infinite loop
👍 1
7
#2 opened about 1 month ago
by
dwaynedu
GLM-5-AWQ vLLM 部署指南
👍 1
2
#2 opened about 1 month ago
by
CharlesChen2023
Great work
5
#1 opened about 1 month ago
by
JoeyHwong
How run this model on Sglang?
1
#2 opened about 1 month ago
by
Salvadori
Anyone else getting only exclamation marks?
15
#3 opened about 2 months ago
by
Halbin
QuantTrio/Qwen3.5-397B-A17B-AWQ reponse is !
8
#5 opened about 1 month ago
by
duyuting
GPTQ int4-int8 mixed
2
#4 opened about 2 months ago
by
darkstar3537
--max-model-len 32768 ?
1
#1 opened 3 months ago
by
pathosethoslogos
The model startup using vllm failed.
10
#5 opened 3 months ago
by
beausoft
Accessing LLM, response without<think>start tag
5
#2 opened 4 months ago
by
sudage
Minimax-M2.1 AWQ Please
2
#6 opened 4 months ago
by
mtcl