LagPixelLOL
v2ray
AI & ML interests
Looking for compute sponsors, please contact me through my email [email protected]!
Recent Activity
updated
a model
1 day ago
x2ray/stuffs
published
a model
1 day ago
x2ray/stuffs
new activity
1 day ago
cognitivecomputations/DeepSeek-R1-AWQ:update vllm to 0.8.x and meet some trouble
Organizations
v2ray's activity
update vllm to 0.8.x and meet some trouble
3
#34 opened 2 days ago
by
HuggingLianWang
Big L.O.V.E. to YOU <3
2
1
#1 opened 11 days ago
by
v2ray

Stuck when run on 8xH100
1
#8 opened 17 days ago
by
Thai
why tokenizer_config.json changed for AWQ model.
2
#7 opened 25 days ago
by
rockcat-miao
Does FlashMLA support kv cache fp8 dtype and how to enable FlashMLA ?
9
#6 opened about 1 month ago
by
CharlesLincoln
How to Resolve "GLIBC_2.32 Not Found" Error When Deploying vLLM Environment?
8
#32 opened about 1 month ago
by
lastsummerLi
Can the 4090 device run this model?
3
#3 opened about 1 month ago
by
jinzhongwei
vllm crach with a slightly longer prompt
1
#4 opened about 1 month ago
by
rockcat-miao
可以添加一下LICENSE文件吗?
2
#2 opened about 1 month ago
by
adol-ch
Are there any updates to the recommended commands?
5
#27 opened about 2 months ago
by
NaiveYan
Why hasn't the MTP layer of the 61st layer been quantized?
1
#30 opened about 1 month ago
by
yang001002
Is there any testing on the support for running on other memory capacities
1
#29 opened about 1 month ago
by
HRan2004
Any one can run this model with SGlang framework?
5
#13 opened 3 months ago
by
muziyongshixin
DeepSeek-R1-AWQ quantized model missing one layer of experts
4
#28 opened about 2 months ago
by
virilo
About the group size
1
#26 opened about 2 months ago
by
Skyeaee

The awq quantization model may encounter garbled characters when performing inference on long texts.
9
#24 opened 2 months ago
by
wx111
How can I quantify my BF16 format model into AWQ?
1
#25 opened 2 months ago
by
AlipaySimon

Support for inference with MTP module?
1
#23 opened 2 months ago
by
yhh001