reign12 commited on
Commit
b39ef45
·
verified ·
1 Parent(s): cf3ba6d

Upload folder using huggingface_hub

Browse files
model-00001-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5859b6e59df3d005f665005693be85b20e77db30af5eb2b9aeb49b8a6f164b94
3
  size 4877660776
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fcf2748b8b4713be58edc34b4bb4440b4d5eefcb97f95129eddd5b5dad7741e5
3
  size 4877660776
model-00002-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0618f94ea53aa37601e5bca7b4a1511ba27283cb809acf1c20df8fb6d752ca03
3
  size 4932751008
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fa94fdbc2059722aab66366f9c2187bbfb1ae4d195717b4c5466eec95fd138b8
3
  size 4932751008
model-00003-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2daee7068c0828447ed67e8a0a40002815d8d30fbe74512353cf248150b09778
3
  size 4330865200
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bf6f4531c4263b490b0015c7f8d36b392b8e109769a3f328a9167f699714d472
3
  size 4330865200
model-00004-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cef74e0e0020a4f5df992c348271671959127d8354a180a32e120fe1ec5b98f6
3
  size 1089994880
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e6310f9343efeabb7a748f9caf803c8ab40ebbee4ab6ead5b80ca33c48dc8d97
3
  size 1089994880
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9c5ae00e602b8860cbd784ba82a8aa14e8feecec692e7076590d014d7b7fdafa
3
- size 11421896
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a8506e7111b80c6d8635951a02eab0f4e1a8e4e5772da83846579e97b16f61bf
3
+ size 7031673
tokenizer_config.json CHANGED
@@ -195,7 +195,7 @@
195
  "<|video_pad|>"
196
  ],
197
  "bos_token": null,
198
- "chat_template": "{{ bos_token }}A conversation between User and Assistant. The User asks a question, and the Assistant solves it. The Assistant first thinks about the reasoning process in the mind and then provides the User with the answer. The reasoning process is enclosed within <think> </think> and answer is enclosed within <answer> </answer> tags, respectively, i.e., <think> reasoning process here </think> <answer> answer here </answer>. {% for message in messages %}{% if message['role'] == 'user' %}User: You must put your answer inside <answer> </answer> tags, i.e., <answer> answer here </answer>. And your final answer will be extracted automatically by the \\boxed{} tag.\nThis is the problem:\n{{ message['content'] }}\n{% elif message['role'] == 'assistant' %}Assistant: <think>{{ message['content'] }}</answer>\n{% endif %}{% endfor %}{% if add_generation_prompt %}Assistant: <think>{% endif %}",
199
  "clean_up_tokenization_spaces": false,
200
  "eos_token": "<|endoftext|>",
201
  "errors": "replace",
 
195
  "<|video_pad|>"
196
  ],
197
  "bos_token": null,
198
+ "chat_template": "{%- if messages[0]['role'] == 'system' %}\n {{- messages[0]['content'] }}\n{%- else %}\n {{- 'A conversation between User and Assistant. The User asks a question, and the Assistant solves it. The Assistant first thinks about the reasoning process in the mind and then provides the User with the answer. The reasoning process is enclosed within <think> </think> and answer is enclosed within <answer> </answer> tags, respectively, i.e., <think> reasoning process here </think> <answer> answer here </answer>. ' }}\n{%- endif %}\n{%- for message in messages %}\n {%- if (message.role == \"user\") %}\n {{- 'User: ' + message.content + '\\n' }}\n {%- elif message.role == \"assistant\" %}\n {{- 'Assistant: <think>' + message.content + '\\n' }}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- 'Assistant: <think>' }}\n{%- endif %}\n",
199
  "clean_up_tokenization_spaces": false,
200
  "eos_token": "<|endoftext|>",
201
  "errors": "replace",