Michael Goin
mgoin
AI & ML interests
LLM inference optimization, compression, quantization, pruning, distillation
Recent Activity
new activity
2 days ago
RedHatAI/Qwen2.5-VL-72B-Instruct-quantized.w8a8:用vllm serve启动不了
updated
a Space
6 days ago
RedHatAI/README
updated
a model
10 days ago
nm-testing/gemma-3-27b-it-FP8-dynamic
Organizations
mgoin's activity
用vllm serve启动不了
1
#2 opened about 1 month ago
by
VenomEY
Fix processor_class to match upstream
#4 opened 17 days ago
by
zifeitong
Remove image_processor_type
#1 opened about 1 month ago
by
pooya-davoodi-parasail
how to deploy this model without internet connection
1
#1 opened 22 days ago
by
superahn
Why not FP8 with static and per-tensor quantization?
1
1
#2 opened 28 days ago
by
wanzhenchn
Address discrepancies in the languages supported by the Mistral Small 3.1 2503
1
3
#54 opened about 1 month ago
by
fpaupier

Please update the chat template
1
#1 opened about 1 month ago
by
stelterlab

FP8 Dynamic/W8A16 Quants Please
4
#44 opened about 1 month ago
by
rjmehta
Problem hosting the model using vllm
3
4
#45 opened about 1 month ago
by
ShaoServient
Remove image_processor_type
#1 opened 2 months ago
by
pooya-davoodi-parasail
Remove image_processor_type
1
#1 opened 2 months ago
by
pooya-davoodi-parasail
Remove image_processor_type
#2 opened 2 months ago
by
pooya-davoodi-parasail
Use Qwen2VLImageProcessor for image_processor_type
5
#2 opened 3 months ago
by
pooya-davoodi-parasail
Use Qwen2VLImageProcessor for image_processor_type
#3 opened 2 months ago
by
pooya-davoodi-parasail
when i use vllm v0.7.2 to deploy r1 awq, i got empty content
13
#10 opened 3 months ago
by
bupalinyu
MLA is not supported with moe_wna16 quantization. Disabling MLA.
5
#7 opened 3 months ago
by
AMOSE
compressed-tensors MLA support requires fp8 activations and weights in group 'group_0',
2
#1 opened 3 months ago
by
samos123
How to load this model?
2
#1 opened 10 months ago
by
Frz614