Model Card for radm/Qwen2.5-32B-simpo-FP8
Model Details
Improved quality on hard tasks by 25 percent relative to the base model Qwen2.5-32B-Instruct. Improved multilingual support.
Fine-tuning on A100 in 4-bit with unsloth using SIMPO and custom dataset
LoRA adapter: radm/Qwen2.5-32B-simpo-LoRA
Eval results
Eval results on ZebraLogic
- Downloads last month
- 6
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
🙋
Ask for provider support