ykarout commited on
Commit
376a9eb
·
verified ·
1 Parent(s): 655bbfc

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +62 -0
README.md ADDED
@@ -0,0 +1,62 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ datasets:
4
+ - nvidia/Llama-Nemotron-Post-Training-Dataset
5
+ language:
6
+ - en
7
+ base_model:
8
+ - microsoft/phi-4
9
+ - unsloth/phi-4
10
+ pipeline_tag: text-generation
11
+ library_name: transformers
12
+ tags:
13
+ - phi-4
14
+ - deepseek
15
+ - r1
16
+ - reasoning
17
+ - code
18
+ - math
19
+ - science
20
+ - unsloth
21
+ ---
22
+ # Model Card for Model ID
23
+
24
+ Phi-4 unsloth model trained to generate deepseek-r1 styled reasoning based on a system prompt "detailed thinking on"
25
+
26
+
27
+ ## Model Details
28
+
29
+ ### Model Description
30
+
31
+ This fine-tuned model generated ehanced chain-of-thoughts, reasoning and produce "Aha-moments" akin to deepseek whenever the system prompt is set to "detailed thinking on".
32
+ Test any questions from trending datasets about code, math and science, with the system prompt set and unset and you can clearly see the difference in the generated output.
33
+ A Modelfile is included with the gguf files that can be used to load the model into Ollama. You have to set the system prompt manually after loading the model in Ollama since by
34
+ default there is no system prompt. You can use /set SYSTEM "detailed thinking on" and then input your prompt. The Modelfile includes optimal parameters but you can experimment
35
+ with different set of parameters based on your desired goal/output.
36
+
37
+
38
+ ## Uses
39
+
40
+ Tasks requiring reasoning, chain-of-thoughts, several approaches etc...
41
+
42
+
43
+ ### Recommendations
44
+
45
+ Use the parameters in the Modefile and set system prompt to "detailed thinking on" whenever you require long reasoning outputs. Set system parameter off when you want a direct
46
+ to the point quick answer without reasoning chains.
47
+ It is important to utilize the chat-template embedded in the Modelfile to ensure optimal generations and avoid endless generations or loops.
48
+
49
+ ## How to Get Started with the Model
50
+
51
+ Download the gguf file and Modelfile into the same folder then use ollama create phi4-deepseek -f Modelfile. Then run the model using ollama run. Set the system parameter
52
+ Finally, start prompting.
53
+
54
+ ## Training Details
55
+
56
+ ### Training Data
57
+
58
+ Nvidia datasets containing reasoning context produced by DeeepSeek-R1.
59
+
60
+ ### Training Procedure
61
+
62
+ Unsloth SFT Trainer