totob-1.5B

Overview

DeepSeek-R1 has garnered attention for matching OpenAI’s O1 reasoning model while being fully open-source, making it an attractive option for users who value local deployment for data privacy, reduced latency, and offline access. Traditionally, running such large models on personal devices involves quantization (e.g., Q4_K_M), which can compromise accuracy by as much as ~22% and diminish the benefits of local inference. With our new totob-1.5B model, we’ve overcome this trade-off by quantizing the DeepSeek-R1 Distilled model to just a quarter of its original size — without any loss in accuracy.

Benchmarks

Coming soon!!

Downloads last month
21
GGUF
Model size
1.78B params
Architecture
qwen2
Hardware compatibility
Log In to view the estimation
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 2 Ask for provider support

Model tree for TotoB12/totob-1.5B

Quantized
(195)
this model

Space using TotoB12/totob-1.5B 1