exl3 quants

#4
by homeworkace - opened

In the off-chance that you take requests from the Community tab, may I humbly ask for ExLlamav3 quants of c4ai-command-r-08-2024 and c4ai-command-a-03-2025? Whichever you think can fit into 24GB of VRAM!

Command-A is a huge model, so that's going to be tough in 24 GB. Not impossible, perhaps. But I converted Command-R now. It's here. 4.0bpw should fit in 24 GB with about a 40k context.

Your need to confirm your account before you can post a new comment.

Sign up or log in to comment