cover

QWQ-32B-Dawnwhisper

Use Qwen2.5-32B tokenizer. It should be better with normal roleplay instead of reasoning roleplay, I guess.

Like many people said: Tiny Deepseek R1 at home if you don't have too good specs. 16GB Vram card could run IQ3 variants very well.

After quick test, this merge perform very good result and strong capability in roleplay. Nothing more. Thank you for using my merge model.

GGUF (Thank mradermacher and his team, especially nicoboss)

Static - Imatrix

Setting

Please use ChatML template

Reasoning is not necessary to turn on, but a nice feature to enable if you want to 'boost' your experience when roleplaying and multitasking (but more time consuming :/).

Reasoning token is <thinking> </thinking>. You could search how to enable thinking mode in internet. Note that in silly tarven, you should turn off Always add character's name to prompt in Context Formatting and Include names Never in Instruct Template.

Configuration

The following YAML configuration was used to produce this model:

models:
  - model: trashpanda-org/QwQ-32B-Snowdrop-v0
    parameters:
      density: 0.9
      weight: 1
  - model: ArliAI/QwQ-32B-ArliAI-RpR-v3
    parameters:
      density: 0.8
      weight: 0.8
merge_method: dare_ties
base_model: Qwen/QwQ-32B
parameters:
  normalize: true
  rescale: true
tokenizer_source: Qwen/Qwen2.5-32B-Instruct
dtype: bfloat16
Downloads last month
65
Safetensors
Model size
32.8B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for DoppelReflEx/QWQ-32B-Dawnwhisper

Collection including DoppelReflEx/QWQ-32B-Dawnwhisper