
QWQ-32B-Dawnwhisper
Use Qwen2.5-32B tokenizer. It should be better with normal roleplay instead of reasoning roleplay, I guess.
Like many people said: Tiny Deepseek R1 at home if you don't have too good specs. 16GB Vram card could run IQ3 variants very well.
After quick test, this merge perform very good result and strong capability in roleplay. Nothing more. Thank you for using my merge model.
GGUF (Thank mradermacher and his team, especially nicoboss)
Setting
Please use ChatML template
Reasoning is not necessary to turn on, but a nice feature to enable if you want to 'boost' your experience when roleplaying and multitasking (but more time consuming :/).
Reasoning token is <thinking> </thinking>
. You could search how to enable thinking mode in internet. Note that in silly tarven, you should turn off Always add character's name to prompt in Context Formatting and Include names Never in Instruct Template.
Configuration
The following YAML configuration was used to produce this model:
models:
- model: trashpanda-org/QwQ-32B-Snowdrop-v0
parameters:
density: 0.9
weight: 1
- model: ArliAI/QwQ-32B-ArliAI-RpR-v3
parameters:
density: 0.8
weight: 0.8
merge_method: dare_ties
base_model: Qwen/QwQ-32B
parameters:
normalize: true
rescale: true
tokenizer_source: Qwen/Qwen2.5-32B-Instruct
dtype: bfloat16
- Downloads last month
- 65