Create README.md
Browse files
README.md
ADDED
@@ -0,0 +1,62 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: mit
|
3 |
+
datasets:
|
4 |
+
- nvidia/Llama-Nemotron-Post-Training-Dataset
|
5 |
+
language:
|
6 |
+
- en
|
7 |
+
base_model:
|
8 |
+
- microsoft/phi-4
|
9 |
+
- unsloth/phi-4
|
10 |
+
pipeline_tag: text-generation
|
11 |
+
library_name: transformers
|
12 |
+
tags:
|
13 |
+
- phi-4
|
14 |
+
- deepseek
|
15 |
+
- r1
|
16 |
+
- reasoning
|
17 |
+
- code
|
18 |
+
- math
|
19 |
+
- science
|
20 |
+
- unsloth
|
21 |
+
---
|
22 |
+
# Model Card for Model ID
|
23 |
+
|
24 |
+
Phi-4 unsloth model trained to generate deepseek-r1 styled reasoning based on a system prompt "detailed thinking on"
|
25 |
+
|
26 |
+
|
27 |
+
## Model Details
|
28 |
+
|
29 |
+
### Model Description
|
30 |
+
|
31 |
+
This fine-tuned model generated ehanced chain-of-thoughts, reasoning and produce "Aha-moments" akin to deepseek whenever the system prompt is set to "detailed thinking on".
|
32 |
+
Test any questions from trending datasets about code, math and science, with the system prompt set and unset and you can clearly see the difference in the generated output.
|
33 |
+
A Modelfile is included with the gguf files that can be used to load the model into Ollama. You have to set the system prompt manually after loading the model in Ollama since by
|
34 |
+
default there is no system prompt. You can use /set SYSTEM "detailed thinking on" and then input your prompt. The Modelfile includes optimal parameters but you can experimment
|
35 |
+
with different set of parameters based on your desired goal/output.
|
36 |
+
|
37 |
+
|
38 |
+
## Uses
|
39 |
+
|
40 |
+
Tasks requiring reasoning, chain-of-thoughts, several approaches etc...
|
41 |
+
|
42 |
+
|
43 |
+
### Recommendations
|
44 |
+
|
45 |
+
Use the parameters in the Modefile and set system prompt to "detailed thinking on" whenever you require long reasoning outputs. Set system parameter off when you want a direct
|
46 |
+
to the point quick answer without reasoning chains.
|
47 |
+
It is important to utilize the chat-template embedded in the Modelfile to ensure optimal generations and avoid endless generations or loops.
|
48 |
+
|
49 |
+
## How to Get Started with the Model
|
50 |
+
|
51 |
+
Download the gguf file and Modelfile into the same folder then use ollama create phi4-deepseek -f Modelfile. Then run the model using ollama run. Set the system parameter
|
52 |
+
Finally, start prompting.
|
53 |
+
|
54 |
+
## Training Details
|
55 |
+
|
56 |
+
### Training Data
|
57 |
+
|
58 |
+
Nvidia datasets containing reasoning context produced by DeeepSeek-R1.
|
59 |
+
|
60 |
+
### Training Procedure
|
61 |
+
|
62 |
+
Unsloth SFT Trainer
|