GGUF quantized versions of very large models - over 100B parameters
-
DevQuasar/nvidia.Llama-3_1-Nemotron-Ultra-253B-CPT-v1-GGUF
Text Generation • Updated • 260 -
DevQuasar/meta-llama.Llama-4-Maverick-17B-128E-Instruct-GGUF
Text Generation • Updated • 2.76k -
DevQuasar/meta-llama.Llama-4-Scout-17B-16E-Instruct-GGUF
Text Generation • Updated • 3.79k • 2 -
DevQuasar/nvidia.Llama-3_1-Nemotron-Ultra-253B-v1-GGUF
Text Generation • Updated • 3.67k • 7