distilltest / README.md
genesisclay's picture
Update README.md
e6fc759 verified
---
title: Distilltest
emoji: 💬
colorFrom: yellow
colorTo: purple
sdk: gradio
sdk_version: 5.0.1
app_file: app.py
pinned: false
license: mit
short_description: Uncensored R1 Distill
---
# DeepSeek-R1-Distill-Qwen-1.5B Demo
This is a demonstration of the DeepSeek-R1-Distill-Qwen-1.5B-uncensored model, a lightweight 1.5B parameter language model from ThirdEyeAI. Despite its small size, this model can generate coherent text across various topics.
## Features
- Efficient inference with a small memory footprint
- Adjustable generation parameters (length, temperature, top-p)
- Simple interface for text generation
## Usage
1. Enter your prompt in the text box
2. Adjust generation parameters if desired
3. Click "Submit" to generate text
## Model Information
This space runs the DeepSeek-R1-Distill-Qwen-1.5B-uncensored model from ThirdEyeAI. It's a distilled version of the Qwen architecture, optimized for efficiency while maintaining good performance. The model is available on Hugging Face at [thirdeyeai/DeepSeek-R1-Distill-Qwen-1.5B-uncensored](https://huggingface.co/thirdeyeai/DeepSeek-R1-Distill-Qwen-1.5B-uncensored).
## Limitations
As a small 1.5B parameter model, this LLM has certain limitations:
- Less knowledge than larger models
- More limited reasoning capabilities
- May produce less coherent outputs on complex topics
## License
[Include license information for your model]
An example chatbot using [Gradio](https://gradio.app), [`huggingface_hub`](https://huggingface.co/docs/huggingface_hub/v0.22.2/en/index), and the [Hugging Face Inference API](https://huggingface.co/docs/api-inference/index).