Template bug fixed in llama.cpp
3
#11 opened about 13 hours ago
by
matteogeniaccio
vllm depolyment error
#10 opened 1 day ago
by
Saicy
Higher than usual refusal rate with Q6_K_L quant GGUF
3
#9 opened 2 days ago
by
smcleod

Tool use?
2
#8 opened 3 days ago
by
johnpyp
llama.cpp fixes have just been merged
2
21
#5 opened 5 days ago
by
Mushoz
LM Studio: unknown model architecture: 'glm4'?
5
#4 opened 9 days ago
by
DrNicefellow

please regenerate ggufs
3
1
#3 opened 13 days ago
by
jacek2024
Broken results
1
8
#2 opened 14 days ago
by
RamoreRemora
Yarn quantization for long context
1
#1 opened 14 days ago
by
sovetboga