user commited on
Commit
b2630e5
·
1 Parent(s): d089450

Add model files

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
Modelfile ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # ollama modelfile auto-generated by llamafactory
2
+
3
+ FROM .
4
+
5
+ TEMPLATE """{{ if .System }}System: {{ .System }}<|im_end|>
6
+ {{ end }}{{ range .Messages }}{{ if eq .Role "user" }}Human: {{ .Content }}<|im_end|>
7
+ Assistant:{{ else if eq .Role "assistant" }}{{ .Content }}<|im_end|>
8
+ {{ end }}{{ end }}"""
9
+
10
+ PARAMETER stop "<|im_end|>"
11
+ PARAMETER num_ctx 4096
README.md CHANGED
@@ -1,3 +1,150 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ license: other
4
+ base_model: agentlans/Qwen2.5-1.5B-Instruct-drill
5
+ tags:
6
+ - llama-factory
7
+ - generated_from_trainer
8
+ model-index:
9
+ - name: Qwen2.5-1.5B-Instruct-Titler
10
+ results: []
11
+ ---
12
+
13
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
+ should probably proofread and complete it, then remove this comment. -->
15
+
16
+ # Qwen2.5-1.5B-Instruct-Titler
17
+
18
+ This model is a fine-tuned version of [agentlans/Qwen2.5-1.5B-Instruct-drill](https://huggingface.co/agentlans/Qwen2.5-1.5B-Instruct-drill) on the [agentlans/finewebedu-annotated](https://huggingface.co/datasets/agentlans/finewebedu-annotated) dataset.
19
+
20
+ ## Model description
21
+
22
+ This model generates a summary from a piece of text in XML format.
23
+
24
+ Input format:
25
+ ```text
26
+ Title:
27
+ {{YOUR_TEXT_HERE}}
28
+ ```
29
+
30
+ Output format:
31
+ ```xml
32
+ <titles>
33
+ <title>Title</title>
34
+ <subtitle>Subtitle</subtitle>
35
+ </titles>
36
+ ```
37
+
38
+ ### Example
39
+
40
+ <details>
41
+ <summary>Click here for example input prompt</summary>
42
+
43
+ From Wikisource contributors. ["Sophocles' King Oedipus."](https://en.wikisource.org/wiki/Sophocles%27_King_Oedipus) Wikisource . Wikisource , 13 Mar. 2025. Web. 19 Apr. 2025.
44
+
45
+ ```
46
+ Summarize:
47
+ Oedipus. Children, descendants of old Cad­mus, why do you come before me, why do you carry the branches of suppliants, while the city smokes with incense and murmurs with prayer and lamentation? I would not learn from any mouth but yours, old man, therefore I question you myself. Do you know of anything that I can do and have not done? How can I, being the man I am, being King Oedipus, do other than all I know? I were indeed hard of heart did I not pity such suppliants.
48
+
49
+ Priest. Oedipus, King of my country, you can see our ages who are before your door; some it may be too young for such a journey, and some too old, Priests of Zeus such as I, and these chosen young men; while the rest of the people crowd the market-places with their suppliant branches, for the city stumbles towards death, hardly able to raise up its head. A blight has fallen upon the fruitful blossoms of the land, a blight upon flock and field and upon the bed of marriage—plague ravages the city. Oedipus, King, not god but foremost of living men, seeing that when you first came to this town of Thebes you freed us from that harsh singer, the riddling sphinx, we beseech you, all we suppliants, to find some help. Whether you find it by your power as a man, or because, being near the gods, a god has whispered you. Uplift our State; think upon your fame; your coming brought us luck, be lucky to us still, remember that it is better to rule over men than over a waste place, since neither walled town nor ship is anything if it be empty and no man within it.
50
+
51
+ Oedipus. My unhappy children! I know well what need has brought you, what suffering you endure, yet sufferers though you be, there is not a single one whose suffering is as mine—each mourns himself, but my soul mourns the city, myself, and you. It is not therefore as if you came to arouse a sleeping man. No! Be certain that I have wept many tears and searched hither and thither for some remedy. I have already done the only thing that came in to my head for all my search. I have sent the son of Menoeceus, Creon, my own wife’s brother, to the Pythian House of Phoebus, to hear if deed or word of mine may yet deliver this town. I am troubled, for he is a long time away—a longer time than should be—but when he comes I shall not be an honest man unless I do whatever the god commands.
52
+
53
+ Priest. You have spoken at the right time. They have just signalled to us that Creon has arrived.
54
+
55
+ Oedipus. O King Apollo, may he bring brighter fortune, for his face is shining.
56
+
57
+ Priest. He brings good news, for he is crowned with bay.
58
+
59
+ Oedipus. We shall know soon. Brother-in-law, Menoeceus’ son, what news from the god?
60
+
61
+ Creon. Good news; for pain turns to pleasure when we have set the crooked straight.
62
+
63
+ Oedipus. But what is the oracle—so far the news is neither good nor bad.
64
+
65
+ Creon. If you would hear it with all these about you, I am ready to speak. Or do we go within?
66
+
67
+ Oedipus. Speak before all. The sorrow I endure is less for my own life than these.
68
+
69
+ Creon. Then, with your leave, I speak. Our lord Phoebus bids us drive out a defiling thing that has been cherished in this land.
70
+
71
+ Oedipus. By what purification?
72
+
73
+ Creon. King Laius was our king before you came to pilot us.
74
+
75
+ Oedipus. I know—but not of my own knowledge, for I never saw him.
76
+
77
+ Creon. He was killed; and the god now bids us revenge it on his murderers, whoever they be.
78
+
79
+ Oedipus. Where shall we come upon their track after all these years? Did he meet his death in house or field, at home or in some foreign land?
80
+
81
+ Creon. In a foreign land: he was journey­ing to Delphi.
82
+
83
+ Oedipus. Did no fellow-traveller see the deed? Was there none there who could be questioned?
84
+
85
+ Creon. All perished but one man who fled in terror and could tell for certain but one thing of all he had seen.
86
+
87
+ Oedipus. One thing might be a clue to many things.
88
+
89
+ Creon. He said that they were fallen upon by a great troop of robbers.
90
+
91
+ Oedipus. What robbers would be so daring unless bribed from here?
92
+
93
+ Creon. Such things were indeed guessed at, but Laius once dead no avenger arose. We were amid our troubles.
94
+
95
+ Oedipus. But when royalty had fallen what troubles could have hindered search?
96
+
97
+ Creon. The riddling sphinx put those dark things out of our thoughts—we thought of what had come to our own doors.
98
+
99
+ Oedipus. But I will start afresh and make the dark things plain. In doing right by Laius I protect myself, for whoever slew Laius might turn a hand against me. Come, my children, rise up from the altar steps; lift up these suppliant boughs and let all the children of Cadmus be called hither that I may search out everything and find for all happiness or misery as god wills.
100
+
101
+ Priest. May Phoebus, sender of the oracle, come with it and be our saviour and deliverer.
102
+ ```
103
+ </details>
104
+
105
+ Output:
106
+ ```xml
107
+ <titles>
108
+ <title>Oedipus Rex: Unraveling the Mysteries of Ancient Greek Tragedy</title>
109
+ <subtitle>A Study of Fate, Family, and Redemption through the Eyes of Sophocles' Masterpiece</subtitle>
110
+ </titles>
111
+ ```
112
+
113
+ ## Model description
114
+
115
+ More information needed
116
+
117
+ ## Intended uses & limitations
118
+
119
+ More information needed
120
+
121
+ ## Training and evaluation data
122
+
123
+ More information needed
124
+
125
+ ## Training procedure
126
+
127
+ ### Training hyperparameters
128
+
129
+ The following hyperparameters were used during training:
130
+ - learning_rate: 5e-05
131
+ - train_batch_size: 2
132
+ - eval_batch_size: 8
133
+ - seed: 42
134
+ - gradient_accumulation_steps: 8
135
+ - total_train_batch_size: 16
136
+ - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
137
+ - lr_scheduler_type: cosine
138
+ - num_epochs: 3.0
139
+
140
+ ### Training results
141
+
142
+
143
+
144
+ ### Framework versions
145
+
146
+ - PEFT 0.15.0
147
+ - Transformers 4.49.0
148
+ - Pytorch 2.6.0+cu124
149
+ - Datasets 3.4.1
150
+ - Tokenizers 0.21.0
added_tokens.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</tool_call>": 151658,
3
+ "<tool_call>": 151657,
4
+ "<|box_end|>": 151649,
5
+ "<|box_start|>": 151648,
6
+ "<|endoftext|>": 151643,
7
+ "<|file_sep|>": 151664,
8
+ "<|fim_middle|>": 151660,
9
+ "<|fim_pad|>": 151662,
10
+ "<|fim_prefix|>": 151659,
11
+ "<|fim_suffix|>": 151661,
12
+ "<|im_end|>": 151645,
13
+ "<|im_start|>": 151644,
14
+ "<|image_pad|>": 151655,
15
+ "<|object_ref_end|>": 151647,
16
+ "<|object_ref_start|>": 151646,
17
+ "<|quad_end|>": 151651,
18
+ "<|quad_start|>": 151650,
19
+ "<|repo_name|>": 151663,
20
+ "<|video_pad|>": 151656,
21
+ "<|vision_end|>": 151653,
22
+ "<|vision_pad|>": 151654,
23
+ "<|vision_start|>": 151652
24
+ }
config.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "/drive2/Repo/Qwen2.5-1.5B-Instruct-drill",
3
+ "architectures": [
4
+ "Qwen2ForCausalLM"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 151643,
8
+ "eos_token_id": 151645,
9
+ "hidden_act": "silu",
10
+ "hidden_size": 1536,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 8960,
13
+ "max_position_embeddings": 32768,
14
+ "max_window_layers": 21,
15
+ "model_type": "qwen2",
16
+ "num_attention_heads": 12,
17
+ "num_hidden_layers": 28,
18
+ "num_key_value_heads": 2,
19
+ "rms_norm_eps": 1e-06,
20
+ "rope_scaling": null,
21
+ "rope_theta": 1000000.0,
22
+ "sliding_window": 32768,
23
+ "tie_word_embeddings": true,
24
+ "torch_dtype": "bfloat16",
25
+ "transformers_version": "4.49.0",
26
+ "use_cache": true,
27
+ "use_sliding_window": false,
28
+ "vocab_size": 151936
29
+ }
generation_config.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 151643,
3
+ "do_sample": true,
4
+ "eos_token_id": [
5
+ 151645,
6
+ 151643
7
+ ],
8
+ "pad_token_id": 151643,
9
+ "repetition_penalty": 1.1,
10
+ "temperature": 0.7,
11
+ "top_k": 20,
12
+ "top_p": 0.8,
13
+ "transformers_version": "4.49.0"
14
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5511f22f16a4f6807d314c87cf22b285f700d78b7eae1bc8d2366e5af63b03ca
3
+ size 3087467144
special_tokens_map.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>",
5
+ "<|object_ref_start|>",
6
+ "<|object_ref_end|>",
7
+ "<|box_start|>",
8
+ "<|box_end|>",
9
+ "<|quad_start|>",
10
+ "<|quad_end|>",
11
+ "<|vision_start|>",
12
+ "<|vision_end|>",
13
+ "<|vision_pad|>",
14
+ "<|image_pad|>",
15
+ "<|video_pad|>"
16
+ ],
17
+ "eos_token": {
18
+ "content": "<|im_end|>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
+ "pad_token": {
25
+ "content": "<|endoftext|>",
26
+ "lstrip": false,
27
+ "normalized": false,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ }
31
+ }
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c5ae00e602b8860cbd784ba82a8aa14e8feecec692e7076590d014d7b7fdafa
3
+ size 11421896
tokenizer_config.json ADDED
@@ -0,0 +1,209 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "151643": {
6
+ "content": "<|endoftext|>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "151644": {
14
+ "content": "<|im_start|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "151645": {
22
+ "content": "<|im_end|>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "151646": {
30
+ "content": "<|object_ref_start|>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "151647": {
38
+ "content": "<|object_ref_end|>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ },
45
+ "151648": {
46
+ "content": "<|box_start|>",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": true
52
+ },
53
+ "151649": {
54
+ "content": "<|box_end|>",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false,
59
+ "special": true
60
+ },
61
+ "151650": {
62
+ "content": "<|quad_start|>",
63
+ "lstrip": false,
64
+ "normalized": false,
65
+ "rstrip": false,
66
+ "single_word": false,
67
+ "special": true
68
+ },
69
+ "151651": {
70
+ "content": "<|quad_end|>",
71
+ "lstrip": false,
72
+ "normalized": false,
73
+ "rstrip": false,
74
+ "single_word": false,
75
+ "special": true
76
+ },
77
+ "151652": {
78
+ "content": "<|vision_start|>",
79
+ "lstrip": false,
80
+ "normalized": false,
81
+ "rstrip": false,
82
+ "single_word": false,
83
+ "special": true
84
+ },
85
+ "151653": {
86
+ "content": "<|vision_end|>",
87
+ "lstrip": false,
88
+ "normalized": false,
89
+ "rstrip": false,
90
+ "single_word": false,
91
+ "special": true
92
+ },
93
+ "151654": {
94
+ "content": "<|vision_pad|>",
95
+ "lstrip": false,
96
+ "normalized": false,
97
+ "rstrip": false,
98
+ "single_word": false,
99
+ "special": true
100
+ },
101
+ "151655": {
102
+ "content": "<|image_pad|>",
103
+ "lstrip": false,
104
+ "normalized": false,
105
+ "rstrip": false,
106
+ "single_word": false,
107
+ "special": true
108
+ },
109
+ "151656": {
110
+ "content": "<|video_pad|>",
111
+ "lstrip": false,
112
+ "normalized": false,
113
+ "rstrip": false,
114
+ "single_word": false,
115
+ "special": true
116
+ },
117
+ "151657": {
118
+ "content": "<tool_call>",
119
+ "lstrip": false,
120
+ "normalized": false,
121
+ "rstrip": false,
122
+ "single_word": false,
123
+ "special": false
124
+ },
125
+ "151658": {
126
+ "content": "</tool_call>",
127
+ "lstrip": false,
128
+ "normalized": false,
129
+ "rstrip": false,
130
+ "single_word": false,
131
+ "special": false
132
+ },
133
+ "151659": {
134
+ "content": "<|fim_prefix|>",
135
+ "lstrip": false,
136
+ "normalized": false,
137
+ "rstrip": false,
138
+ "single_word": false,
139
+ "special": false
140
+ },
141
+ "151660": {
142
+ "content": "<|fim_middle|>",
143
+ "lstrip": false,
144
+ "normalized": false,
145
+ "rstrip": false,
146
+ "single_word": false,
147
+ "special": false
148
+ },
149
+ "151661": {
150
+ "content": "<|fim_suffix|>",
151
+ "lstrip": false,
152
+ "normalized": false,
153
+ "rstrip": false,
154
+ "single_word": false,
155
+ "special": false
156
+ },
157
+ "151662": {
158
+ "content": "<|fim_pad|>",
159
+ "lstrip": false,
160
+ "normalized": false,
161
+ "rstrip": false,
162
+ "single_word": false,
163
+ "special": false
164
+ },
165
+ "151663": {
166
+ "content": "<|repo_name|>",
167
+ "lstrip": false,
168
+ "normalized": false,
169
+ "rstrip": false,
170
+ "single_word": false,
171
+ "special": false
172
+ },
173
+ "151664": {
174
+ "content": "<|file_sep|>",
175
+ "lstrip": false,
176
+ "normalized": false,
177
+ "rstrip": false,
178
+ "single_word": false,
179
+ "special": false
180
+ }
181
+ },
182
+ "additional_special_tokens": [
183
+ "<|im_start|>",
184
+ "<|im_end|>",
185
+ "<|object_ref_start|>",
186
+ "<|object_ref_end|>",
187
+ "<|box_start|>",
188
+ "<|box_end|>",
189
+ "<|quad_start|>",
190
+ "<|quad_end|>",
191
+ "<|vision_start|>",
192
+ "<|vision_end|>",
193
+ "<|vision_pad|>",
194
+ "<|image_pad|>",
195
+ "<|video_pad|>"
196
+ ],
197
+ "bos_token": null,
198
+ "chat_template": "{% if messages[0]['role'] == 'system' %}{% set loop_messages = messages[1:] %}{% set system_message = messages[0]['content'] %}{% else %}{% set loop_messages = messages %}{% endif %}{% if system_message is defined %}{{ 'System: ' + system_message + '<|im_end|>' + '\n' }}{% endif %}{% for message in loop_messages %}{% set content = message['content'] %}{% if message['role'] == 'user' %}{{ 'Human: ' + content + '<|im_end|>' + '\nAssistant:' }}{% elif message['role'] == 'assistant' %}{{ content + '<|im_end|>' + '\n' }}{% endif %}{% endfor %}",
199
+ "clean_up_tokenization_spaces": false,
200
+ "eos_token": "<|im_end|>",
201
+ "errors": "replace",
202
+ "extra_special_tokens": {},
203
+ "model_max_length": 131072,
204
+ "pad_token": "<|endoftext|>",
205
+ "padding_side": "left",
206
+ "split_special_tokens": false,
207
+ "tokenizer_class": "Qwen2Tokenizer",
208
+ "unk_token": null
209
+ }
vocab.json ADDED
The diff for this file is too large to render. See raw diff