diff --git "a/notebooks/07_Qwen2.5_models.ipynb" "b/notebooks/07_Qwen2.5_models.ipynb" --- "a/notebooks/07_Qwen2.5_models.ipynb" +++ "b/notebooks/07_Qwen2.5_models.ipynb" @@ -19954,7 +19954,7 @@ }, { "cell_type": "code", - "execution_count": 8, + "execution_count": 11, "metadata": {}, "outputs": [ { @@ -19963,7 +19963,7 @@ "text": [ "Current Directory:\n", "/home/inflaton/code/logical-reasoning\n", - "Sat Sep 21 23:55:43 2024 \n", + "Sun Sep 22 08:58:48 2024 \n", "+-----------------------------------------------------------------------------------------+\n", "| NVIDIA-SMI 560.35.02 Driver Version: 560.94 CUDA Version: 12.6 |\n", "|-----------------------------------------+------------------------+----------------------+\n", @@ -19972,7 +19972,7 @@ "| | | MIG M. |\n", "|=========================================+========================+======================|\n", "| 0 NVIDIA GeForce RTX 4090 On | 00000000:01:00.0 Off | Off |\n", - "| 54% 59C P3 41W / 450W | 471MiB / 24564MiB | 2% Default |\n", + "| 30% 46C P8 17W / 450W | 432MiB / 24564MiB | 4% Default |\n", "| | | N/A |\n", "+-----------------------------------------+------------------------+----------------------+\n", " \n", @@ -20052,21 +20052,16 @@ "MemTotal: 49330024 kB\n", "Current Directory:\n", "/home/inflaton/code/logical-reasoning\n", - "Evaluating Qwen/Qwen2.5-7B-Instruct with few-shot learning\n", + "Eval Qwen/Qwen2.5-3B-Instruct with llama-factory/saves/Qwen2.5-3B-Instruct\n", "loading env vars from: /home/inflaton/code/logical-reasoning/.env\n", - "Adding /home/inflaton/code/logical-reasoning to sys.path\n", - "loading /home/inflaton/code/logical-reasoning/llm_toolkit/logical_reasoning_utils.py\n", - "CUDA is available, we have found 1 GPU(s)\n", - "NVIDIA GeForce RTX 4090\n", - "CUDA version: 12.1\n", - "Qwen/Qwen2.5-7B-Instruct None False datasets/mgtv data/Qwen2.5-7B-Instruct_results.csv 2048 1\n", - "(0) GPU = NVIDIA GeForce RTX 4090. Max memory = 23.988 GB.\n", - "0.0 GB of memory reserved.\n", - "loading model: Qwen/Qwen2.5-7B-Instruct with adapter: None\n", - "config.json: 100%|█████████████████████████████| 663/663 [00:00<00:00, 10.9MB/s]\n", - "[INFO|configuration_utils.py:733] 2024-09-21 23:55:48,860 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-7B-Instruct/snapshots/acbd96531cda22292a3ceaa67e984955d3965282/config.json\n", - "[INFO|configuration_utils.py:800] 2024-09-21 23:55:48,861 >> Model config Qwen2Config {\n", - " \"_name_or_path\": \"Qwen/Qwen2.5-7B-Instruct\",\n", + "workding dir: /home/inflaton/code/logical-reasoning\n", + "adding /home/inflaton/code/logical-reasoning to sys.path\n", + "found 10 checkpoints: ['checkpoint-35', 'checkpoint-70', 'checkpoint-105', 'checkpoint-140', 'checkpoint-175', 'checkpoint-210', 'checkpoint-245', 'checkpoint-280', 'checkpoint-315', 'checkpoint-350']\n", + "Epoch 0\n", + "adapter path: None\n", + "[INFO|configuration_utils.py:733] 2024-09-22 08:59:31,511 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 08:59:31,512 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-3B-Instruct\",\n", " \"architectures\": [\n", " \"Qwen2ForCausalLM\"\n", " ],\n", @@ -20074,40 +20069,36 @@ " \"bos_token_id\": 151643,\n", " \"eos_token_id\": 151645,\n", " \"hidden_act\": \"silu\",\n", - " \"hidden_size\": 3584,\n", + " \"hidden_size\": 2048,\n", " \"initializer_range\": 0.02,\n", - " \"intermediate_size\": 18944,\n", + " \"intermediate_size\": 11008,\n", " \"max_position_embeddings\": 32768,\n", - " \"max_window_layers\": 28,\n", + " \"max_window_layers\": 70,\n", " \"model_type\": \"qwen2\",\n", - " \"num_attention_heads\": 28,\n", - " \"num_hidden_layers\": 28,\n", - " \"num_key_value_heads\": 4,\n", + " \"num_attention_heads\": 16,\n", + " \"num_hidden_layers\": 36,\n", + " \"num_key_value_heads\": 2,\n", " \"rms_norm_eps\": 1e-06,\n", " \"rope_theta\": 1000000.0,\n", " \"sliding_window\": null,\n", - " \"tie_word_embeddings\": false,\n", + " \"tie_word_embeddings\": true,\n", " \"torch_dtype\": \"bfloat16\",\n", " \"transformers_version\": \"4.43.3\",\n", " \"use_cache\": true,\n", " \"use_sliding_window\": false,\n", - " \"vocab_size\": 152064\n", + " \"vocab_size\": 151936\n", "}\n", "\n", - "tokenizer_config.json: 100%|███████████████| 7.30k/7.30k [00:00<00:00, 53.0MB/s]\n", - "vocab.json: 100%|██████████████████████████| 2.78M/2.78M [00:01<00:00, 1.76MB/s]\n", - "merges.txt: 100%|██████████████████████████| 1.67M/1.67M [00:00<00:00, 23.9MB/s]\n", - "tokenizer.json: 100%|██████████████████████| 7.03M/7.03M [00:00<00:00, 10.8MB/s]\n", - "[INFO|tokenization_utils_base.py:2289] 2024-09-21 23:55:53,949 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-7B-Instruct/snapshots/acbd96531cda22292a3ceaa67e984955d3965282/vocab.json\n", - "[INFO|tokenization_utils_base.py:2289] 2024-09-21 23:55:53,950 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-7B-Instruct/snapshots/acbd96531cda22292a3ceaa67e984955d3965282/merges.txt\n", - "[INFO|tokenization_utils_base.py:2289] 2024-09-21 23:55:53,950 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-7B-Instruct/snapshots/acbd96531cda22292a3ceaa67e984955d3965282/tokenizer.json\n", - "[INFO|tokenization_utils_base.py:2289] 2024-09-21 23:55:53,950 >> loading file added_tokens.json from cache at None\n", - "[INFO|tokenization_utils_base.py:2289] 2024-09-21 23:55:53,950 >> loading file special_tokens_map.json from cache at None\n", - "[INFO|tokenization_utils_base.py:2289] 2024-09-21 23:55:53,950 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-7B-Instruct/snapshots/acbd96531cda22292a3ceaa67e984955d3965282/tokenizer_config.json\n", - "[INFO|tokenization_utils_base.py:2533] 2024-09-21 23:55:54,041 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", - "[INFO|configuration_utils.py:733] 2024-09-21 23:55:55,157 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-7B-Instruct/snapshots/acbd96531cda22292a3ceaa67e984955d3965282/config.json\n", - "[INFO|configuration_utils.py:800] 2024-09-21 23:55:55,157 >> Model config Qwen2Config {\n", - " \"_name_or_path\": \"Qwen/Qwen2.5-7B-Instruct\",\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 08:59:31,765 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 08:59:31,765 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 08:59:31,765 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 08:59:31,765 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 08:59:31,765 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 08:59:31,765 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 08:59:31,952 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 08:59:32,998 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 08:59:32,998 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-3B-Instruct\",\n", " \"architectures\": [\n", " \"Qwen2ForCausalLM\"\n", " ],\n", @@ -20115,37 +20106,36 @@ " \"bos_token_id\": 151643,\n", " \"eos_token_id\": 151645,\n", " \"hidden_act\": \"silu\",\n", - " \"hidden_size\": 3584,\n", + " \"hidden_size\": 2048,\n", " \"initializer_range\": 0.02,\n", - " \"intermediate_size\": 18944,\n", + " \"intermediate_size\": 11008,\n", " \"max_position_embeddings\": 32768,\n", - " \"max_window_layers\": 28,\n", + " \"max_window_layers\": 70,\n", " \"model_type\": \"qwen2\",\n", - " \"num_attention_heads\": 28,\n", - " \"num_hidden_layers\": 28,\n", - " \"num_key_value_heads\": 4,\n", + " \"num_attention_heads\": 16,\n", + " \"num_hidden_layers\": 36,\n", + " \"num_key_value_heads\": 2,\n", " \"rms_norm_eps\": 1e-06,\n", " \"rope_theta\": 1000000.0,\n", " \"sliding_window\": null,\n", - " \"tie_word_embeddings\": false,\n", + " \"tie_word_embeddings\": true,\n", " \"torch_dtype\": \"bfloat16\",\n", " \"transformers_version\": \"4.43.3\",\n", " \"use_cache\": true,\n", " \"use_sliding_window\": false,\n", - " \"vocab_size\": 152064\n", + " \"vocab_size\": 151936\n", "}\n", "\n", - "[INFO|tokenization_utils_base.py:2289] 2024-09-21 23:55:55,421 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-7B-Instruct/snapshots/acbd96531cda22292a3ceaa67e984955d3965282/vocab.json\n", - "[INFO|tokenization_utils_base.py:2289] 2024-09-21 23:55:55,421 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-7B-Instruct/snapshots/acbd96531cda22292a3ceaa67e984955d3965282/merges.txt\n", - "[INFO|tokenization_utils_base.py:2289] 2024-09-21 23:55:55,421 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-7B-Instruct/snapshots/acbd96531cda22292a3ceaa67e984955d3965282/tokenizer.json\n", - "[INFO|tokenization_utils_base.py:2289] 2024-09-21 23:55:55,422 >> loading file added_tokens.json from cache at None\n", - "[INFO|tokenization_utils_base.py:2289] 2024-09-21 23:55:55,422 >> loading file special_tokens_map.json from cache at None\n", - "[INFO|tokenization_utils_base.py:2289] 2024-09-21 23:55:55,422 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-7B-Instruct/snapshots/acbd96531cda22292a3ceaa67e984955d3965282/tokenizer_config.json\n", - "[INFO|tokenization_utils_base.py:2533] 2024-09-21 23:55:55,509 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", - "09/21/2024 23:55:55 - INFO - llamafactory.data.template - Replace eos token: <|im_end|>\n", - "[INFO|configuration_utils.py:733] 2024-09-21 23:55:55,814 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-7B-Instruct/snapshots/acbd96531cda22292a3ceaa67e984955d3965282/config.json\n", - "[INFO|configuration_utils.py:800] 2024-09-21 23:55:55,814 >> Model config Qwen2Config {\n", - " \"_name_or_path\": \"Qwen/Qwen2.5-7B-Instruct\",\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 08:59:33,269 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 08:59:33,269 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 08:59:33,269 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 08:59:33,269 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 08:59:33,269 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 08:59:33,269 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 08:59:33,360 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 08:59:33,621 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 08:59:33,621 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-3B-Instruct\",\n", " \"architectures\": [\n", " \"Qwen2ForCausalLM\"\n", " ],\n", @@ -20153,131 +20143,5546 @@ " \"bos_token_id\": 151643,\n", " \"eos_token_id\": 151645,\n", " \"hidden_act\": \"silu\",\n", - " \"hidden_size\": 3584,\n", + " \"hidden_size\": 2048,\n", " \"initializer_range\": 0.02,\n", - " \"intermediate_size\": 18944,\n", + " \"intermediate_size\": 11008,\n", " \"max_position_embeddings\": 32768,\n", - " \"max_window_layers\": 28,\n", + " \"max_window_layers\": 70,\n", " \"model_type\": \"qwen2\",\n", - " \"num_attention_heads\": 28,\n", - " \"num_hidden_layers\": 28,\n", - " \"num_key_value_heads\": 4,\n", + " \"num_attention_heads\": 16,\n", + " \"num_hidden_layers\": 36,\n", + " \"num_key_value_heads\": 2,\n", " \"rms_norm_eps\": 1e-06,\n", " \"rope_theta\": 1000000.0,\n", " \"sliding_window\": null,\n", - " \"tie_word_embeddings\": false,\n", + " \"tie_word_embeddings\": true,\n", " \"torch_dtype\": \"bfloat16\",\n", " \"transformers_version\": \"4.43.3\",\n", " \"use_cache\": true,\n", " \"use_sliding_window\": false,\n", - " \"vocab_size\": 152064\n", + " \"vocab_size\": 151936\n", "}\n", "\n", - "09/21/2024 23:55:55 - INFO - llamafactory.model.patcher - Using KV cache for faster generation.\n", - "model.safetensors.index.json: 100%|████████| 27.8k/27.8k [00:00<00:00, 24.5MB/s]\n", - "[INFO|modeling_utils.py:3634] 2024-09-21 23:55:56,890 >> loading weights file model.safetensors from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-7B-Instruct/snapshots/acbd96531cda22292a3ceaa67e984955d3965282/model.safetensors.index.json\n", - "Downloading shards: 0%| | 0/4 [00:00\n", - " model, tokenizer = load_model(\n", - " ^^^^^^^^^^^\n", - " File \"/home/inflaton/code/logical-reasoning/llm_toolkit/llm_utils.py\", line 52, in load_model\n", - " chat_model = ChatModel(args)\n", - " ^^^^^^^^^^^^^^^\n", - " File \"/home/inflaton/code/LLaMA-Factory/src/llamafactory/chat/chat_model.py\", line 52, in __init__\n", - " self.engine: \"BaseEngine\" = HuggingfaceEngine(model_args, data_args, finetuning_args, generating_args)\n", - " ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n", - " File \"/home/inflaton/code/LLaMA-Factory/src/llamafactory/chat/hf_engine.py\", line 59, in __init__\n", - " self.model = load_model(\n", - " ^^^^^^^^^^^\n", - " File \"/home/inflaton/code/LLaMA-Factory/src/llamafactory/model/loader.py\", line 162, in load_model\n", - " model = load_class.from_pretrained(**init_kwargs)\n", - " ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n", - " File \"/home/inflaton/miniconda3/envs/llm-finetuning/lib/python3.11/site-packages/transformers/models/auto/auto_factory.py\", line 564, in from_pretrained\n", - " return model_class.from_pretrained(\n", - " ^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n", - " File \"/home/inflaton/miniconda3/envs/llm-finetuning/lib/python3.11/site-packages/transformers/modeling_utils.py\", line 3671, in from_pretrained\n", - " resolved_archive_file, sharded_metadata = get_checkpoint_shard_files(\n", - " ^^^^^^^^^^^^^^^^^^^^^^^^^^^\n", - " File \"/home/inflaton/miniconda3/envs/llm-finetuning/lib/python3.11/site-packages/transformers/utils/hub.py\", line 1079, in get_checkpoint_shard_files\n", - " cached_filename = cached_file(\n", - " ^^^^^^^^^^^^\n", - " File \"/home/inflaton/miniconda3/envs/llm-finetuning/lib/python3.11/site-packages/transformers/utils/hub.py\", line 402, in cached_file\n", - " resolved_file = hf_hub_download(\n", - " ^^^^^^^^^^^^^^^^\n", - " File \"/home/inflaton/miniconda3/envs/llm-finetuning/lib/python3.11/site-packages/huggingface_hub/utils/_validators.py\", line 114, in _inner_fn\n", - " return fn(*args, **kwargs)\n", - " ^^^^^^^^^^^^^^^^^^^\n", - " File \"/home/inflaton/miniconda3/envs/llm-finetuning/lib/python3.11/site-packages/huggingface_hub/file_download.py\", line 1221, in hf_hub_download\n", - " return _hf_hub_download_to_cache_dir(\n", - " ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n", - " File \"/home/inflaton/miniconda3/envs/llm-finetuning/lib/python3.11/site-packages/huggingface_hub/file_download.py\", line 1367, in _hf_hub_download_to_cache_dir\n", - " _download_to_tmp_and_move(\n", - " File \"/home/inflaton/miniconda3/envs/llm-finetuning/lib/python3.11/site-packages/huggingface_hub/file_download.py\", line 1884, in _download_to_tmp_and_move\n", - " http_get(\n", - " File \"/home/inflaton/miniconda3/envs/llm-finetuning/lib/python3.11/site-packages/huggingface_hub/file_download.py\", line 539, in http_get\n", - " for chunk in r.iter_content(chunk_size=DOWNLOAD_CHUNK_SIZE):\n", - " File \"/home/inflaton/miniconda3/envs/llm-finetuning/lib/python3.11/site-packages/requests/models.py\", line 820, in generate\n", - " yield from self.raw.stream(chunk_size, decode_content=True)\n", - " File \"/home/inflaton/miniconda3/envs/llm-finetuning/lib/python3.11/site-packages/urllib3/response.py\", line 1060, in stream\n", - " data = self.read(amt=amt, decode_content=decode_content)\n", - " ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n", - " File \"/home/inflaton/miniconda3/envs/llm-finetuning/lib/python3.11/site-packages/urllib3/response.py\", line 949, in read\n", - " data = self._raw_read(amt)\n", - " ^^^^^^^^^^^^^^^^^^^\n", - " File \"/home/inflaton/miniconda3/envs/llm-finetuning/lib/python3.11/site-packages/urllib3/response.py\", line 873, in _raw_read\n", - " data = self._fp_read(amt, read1=read1) if not fp_closed else b\"\"\n", - " ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n", - " File \"/home/inflaton/miniconda3/envs/llm-finetuning/lib/python3.11/site-packages/urllib3/response.py\", line 856, in _fp_read\n", - " return self._fp.read(amt) if amt is not None else self._fp.read()\n", - " ^^^^^^^^^^^^^^^^^^\n", - " File \"/home/inflaton/miniconda3/envs/llm-finetuning/lib/python3.11/http/client.py\", line 473, in read\n", - " s = self.fp.read(amt)\n", - " ^^^^^^^^^^^^^^^^^\n", - " File \"/home/inflaton/miniconda3/envs/llm-finetuning/lib/python3.11/socket.py\", line 706, in readinto\n", - " return self._sock.recv_into(b)\n", - " ^^^^^^^^^^^^^^^^^^^^^^^\n", - " File \"/home/inflaton/miniconda3/envs/llm-finetuning/lib/python3.11/ssl.py\", line 1314, in recv_into\n", - " return self.read(nbytes, buffer)\n", - " ^^^^^^^^^^^^^^^^^^^^^^^^^\n", - " File \"/home/inflaton/miniconda3/envs/llm-finetuning/lib/python3.11/ssl.py\", line 1166, in read\n", - " return self._sslobj.read(len, buffer)\n", - " ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n", - "KeyboardInterrupt\n", - "model-00001-of-00004.safetensors: 8%|▍ | 304M/3.95G [00:29<05:57, 10.2MB/s]\n", - "CPU times: user 901 ms, sys: 326 ms, total: 1.23 s\n", - "Wall time: 2min 55s\n" + "[INFO|modeling_utils.py:3634] 2024-09-22 08:59:33,644 >> loading weights file model.safetensors from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/model.safetensors.index.json\n", + "[INFO|modeling_utils.py:1572] 2024-09-22 08:59:33,659 >> Instantiating Qwen2ForCausalLM model under default dtype torch.bfloat16.\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 08:59:33,659 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645\n", + "}\n", + "\n", + "Loading checkpoint shards: 100%|██████████████████| 2/2 [00:51<00:00, 25.91s/it]\n", + "[INFO|modeling_utils.py:4463] 2024-09-22 09:00:25,834 >> All model checkpoint weights were used when initializing Qwen2ForCausalLM.\n", + "\n", + "[INFO|modeling_utils.py:4471] 2024-09-22 09:00:25,834 >> All the weights of Qwen2ForCausalLM were initialized from the model checkpoint at Qwen/Qwen2.5-3B-Instruct.\n", + "If your task is similar to the task the model of the checkpoint was trained on, you can already use Qwen2ForCausalLM for predictions without further training.\n", + "[INFO|configuration_utils.py:993] 2024-09-22 09:00:26,113 >> loading configuration file generation_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/generation_config.json\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 09:00:26,113 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"do_sample\": true,\n", + " \"eos_token_id\": [\n", + " 151645,\n", + " 151643\n", + " ],\n", + " \"pad_token_id\": 151643,\n", + " \"repetition_penalty\": 1.05,\n", + " \"temperature\": 0.7,\n", + " \"top_k\": 20,\n", + " \"top_p\": 0.8\n", + "}\n", + "\n", + "100%|███████████████████████████████████████| 1500/1500 [03:42<00:00, 6.75it/s]\n", + "Epoch 1\n", + "adapter path: llama-factory/saves/Qwen2.5-3B-Instruct/checkpoint-35\n", + "[INFO|configuration_utils.py:733] 2024-09-22 09:04:19,590 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 09:04:19,590 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-3B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 2048,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 11008,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 70,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 16,\n", + " \"num_hidden_layers\": 36,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:04:19,850 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:04:19,851 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:04:19,851 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:04:19,851 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:04:19,851 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:04:19,851 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 09:04:19,947 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 09:04:20,951 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 09:04:20,951 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-3B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 2048,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 11008,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 70,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 16,\n", + " \"num_hidden_layers\": 36,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:04:21,220 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:04:21,220 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:04:21,220 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:04:21,220 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:04:21,220 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:04:21,220 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 09:04:21,311 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 09:04:21,568 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 09:04:21,568 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-3B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 2048,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 11008,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 70,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 16,\n", + " \"num_hidden_layers\": 36,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:3634] 2024-09-22 09:04:21,578 >> loading weights file model.safetensors from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/model.safetensors.index.json\n", + "[INFO|modeling_utils.py:1572] 2024-09-22 09:04:21,579 >> Instantiating Qwen2ForCausalLM model under default dtype torch.bfloat16.\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 09:04:21,579 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645\n", + "}\n", + "\n", + "Loading checkpoint shards: 100%|██████████████████| 2/2 [00:36<00:00, 18.03s/it]\n", + "[INFO|modeling_utils.py:4463] 2024-09-22 09:04:58,002 >> All model checkpoint weights were used when initializing Qwen2ForCausalLM.\n", + "\n", + "[INFO|modeling_utils.py:4471] 2024-09-22 09:04:58,002 >> All the weights of Qwen2ForCausalLM were initialized from the model checkpoint at Qwen/Qwen2.5-3B-Instruct.\n", + "If your task is similar to the task the model of the checkpoint was trained on, you can already use Qwen2ForCausalLM for predictions without further training.\n", + "[INFO|configuration_utils.py:993] 2024-09-22 09:04:58,253 >> loading configuration file generation_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/generation_config.json\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 09:04:58,253 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"do_sample\": true,\n", + " \"eos_token_id\": [\n", + " 151645,\n", + " 151643\n", + " ],\n", + " \"pad_token_id\": 151643,\n", + " \"repetition_penalty\": 1.05,\n", + " \"temperature\": 0.7,\n", + " \"top_k\": 20,\n", + " \"top_p\": 0.8\n", + "}\n", + "\n", + "100%|███████████████████████████████████████| 1500/1500 [03:32<00:00, 7.07it/s]\n", + "Epoch 2\n", + "adapter path: llama-factory/saves/Qwen2.5-3B-Instruct/checkpoint-70\n", + "[INFO|configuration_utils.py:733] 2024-09-22 09:08:43,242 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 09:08:43,242 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-3B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 2048,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 11008,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 70,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 16,\n", + " \"num_hidden_layers\": 36,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:08:43,517 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:08:43,517 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:08:43,517 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:08:43,517 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:08:43,517 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:08:43,517 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 09:08:43,610 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 09:08:44,643 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 09:08:44,643 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-3B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 2048,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 11008,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 70,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 16,\n", + " \"num_hidden_layers\": 36,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:08:44,900 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:08:44,900 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:08:44,900 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:08:44,900 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:08:44,900 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:08:44,900 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 09:08:44,993 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 09:08:45,291 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 09:08:45,292 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-3B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 2048,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 11008,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 70,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 16,\n", + " \"num_hidden_layers\": 36,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:3634] 2024-09-22 09:08:45,302 >> loading weights file model.safetensors from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/model.safetensors.index.json\n", + "[INFO|modeling_utils.py:1572] 2024-09-22 09:08:45,302 >> Instantiating Qwen2ForCausalLM model under default dtype torch.bfloat16.\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 09:08:45,303 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645\n", + "}\n", + "\n", + "Loading checkpoint shards: 100%|██████████████████| 2/2 [00:35<00:00, 17.97s/it]\n", + "[INFO|modeling_utils.py:4463] 2024-09-22 09:09:21,601 >> All model checkpoint weights were used when initializing Qwen2ForCausalLM.\n", + "\n", + "[INFO|modeling_utils.py:4471] 2024-09-22 09:09:21,601 >> All the weights of Qwen2ForCausalLM were initialized from the model checkpoint at Qwen/Qwen2.5-3B-Instruct.\n", + "If your task is similar to the task the model of the checkpoint was trained on, you can already use Qwen2ForCausalLM for predictions without further training.\n", + "[INFO|configuration_utils.py:993] 2024-09-22 09:09:21,860 >> loading configuration file generation_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/generation_config.json\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 09:09:21,860 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"do_sample\": true,\n", + " \"eos_token_id\": [\n", + " 151645,\n", + " 151643\n", + " ],\n", + " \"pad_token_id\": 151643,\n", + " \"repetition_penalty\": 1.05,\n", + " \"temperature\": 0.7,\n", + " \"top_k\": 20,\n", + " \"top_p\": 0.8\n", + "}\n", + "\n", + "100%|███████████████████████████████████████| 1500/1500 [03:44<00:00, 6.69it/s]\n", + "Epoch 3\n", + "adapter path: llama-factory/saves/Qwen2.5-3B-Instruct/checkpoint-105\n", + "[INFO|configuration_utils.py:733] 2024-09-22 09:13:21,410 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 09:13:21,410 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-3B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 2048,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 11008,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 70,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 16,\n", + " \"num_hidden_layers\": 36,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:13:21,668 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:13:21,668 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:13:21,668 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:13:21,668 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:13:21,668 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:13:21,669 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 09:13:21,763 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 09:13:22,801 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 09:13:22,802 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-3B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 2048,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 11008,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 70,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 16,\n", + " \"num_hidden_layers\": 36,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:13:23,057 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:13:23,057 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:13:23,057 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:13:23,057 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:13:23,057 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:13:23,057 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 09:13:23,150 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 09:13:23,411 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 09:13:23,412 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-3B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 2048,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 11008,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 70,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 16,\n", + " \"num_hidden_layers\": 36,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:3634] 2024-09-22 09:13:23,422 >> loading weights file model.safetensors from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/model.safetensors.index.json\n", + "[INFO|modeling_utils.py:1572] 2024-09-22 09:13:23,422 >> Instantiating Qwen2ForCausalLM model under default dtype torch.bfloat16.\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 09:13:23,423 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645\n", + "}\n", + "\n", + "Loading checkpoint shards: 100%|██████████████████| 2/2 [00:36<00:00, 18.35s/it]\n", + "[INFO|modeling_utils.py:4463] 2024-09-22 09:14:13,170 >> All model checkpoint weights were used when initializing Qwen2ForCausalLM.\n", + "\n", + "[INFO|modeling_utils.py:4471] 2024-09-22 09:14:13,170 >> All the weights of Qwen2ForCausalLM were initialized from the model checkpoint at Qwen/Qwen2.5-3B-Instruct.\n", + "If your task is similar to the task the model of the checkpoint was trained on, you can already use Qwen2ForCausalLM for predictions without further training.\n", + "[INFO|configuration_utils.py:993] 2024-09-22 09:14:13,421 >> loading configuration file generation_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/generation_config.json\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 09:14:13,421 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"do_sample\": true,\n", + " \"eos_token_id\": [\n", + " 151645,\n", + " 151643\n", + " ],\n", + " \"pad_token_id\": 151643,\n", + " \"repetition_penalty\": 1.05,\n", + " \"temperature\": 0.7,\n", + " \"top_k\": 20,\n", + " \"top_p\": 0.8\n", + "}\n", + "\n", + "100%|███████████████████████████████████████| 1500/1500 [03:27<00:00, 7.23it/s]\n", + "Epoch 4\n", + "adapter path: llama-factory/saves/Qwen2.5-3B-Instruct/checkpoint-140\n", + "[INFO|configuration_utils.py:733] 2024-09-22 09:18:15,693 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 09:18:15,693 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-3B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 2048,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 11008,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 70,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 16,\n", + " \"num_hidden_layers\": 36,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:18:16,874 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:18:16,874 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:18:16,874 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:18:16,874 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:18:16,874 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:18:16,874 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 09:18:16,962 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 09:18:17,966 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 09:18:17,966 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-3B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 2048,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 11008,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 70,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 16,\n", + " \"num_hidden_layers\": 36,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:18:18,218 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:18:18,218 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:18:18,218 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:18:18,218 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:18:18,218 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:18:18,218 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 09:18:18,307 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 09:18:18,563 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 09:18:18,564 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-3B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 2048,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 11008,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 70,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 16,\n", + " \"num_hidden_layers\": 36,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:3634] 2024-09-22 09:18:18,573 >> loading weights file model.safetensors from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/model.safetensors.index.json\n", + "[INFO|modeling_utils.py:1572] 2024-09-22 09:18:18,574 >> Instantiating Qwen2ForCausalLM model under default dtype torch.bfloat16.\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 09:18:18,574 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645\n", + "}\n", + "\n", + "Loading checkpoint shards: 100%|██████████████████| 2/2 [00:35<00:00, 17.96s/it]\n", + "[INFO|modeling_utils.py:4463] 2024-09-22 09:18:54,831 >> All model checkpoint weights were used when initializing Qwen2ForCausalLM.\n", + "\n", + "[INFO|modeling_utils.py:4471] 2024-09-22 09:18:54,831 >> All the weights of Qwen2ForCausalLM were initialized from the model checkpoint at Qwen/Qwen2.5-3B-Instruct.\n", + "If your task is similar to the task the model of the checkpoint was trained on, you can already use Qwen2ForCausalLM for predictions without further training.\n", + "[INFO|configuration_utils.py:993] 2024-09-22 09:18:55,312 >> loading configuration file generation_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/generation_config.json\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 09:18:55,313 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"do_sample\": true,\n", + " \"eos_token_id\": [\n", + " 151645,\n", + " 151643\n", + " ],\n", + " \"pad_token_id\": 151643,\n", + " \"repetition_penalty\": 1.05,\n", + " \"temperature\": 0.7,\n", + " \"top_k\": 20,\n", + " \"top_p\": 0.8\n", + "}\n", + "\n", + "100%|███████████████████████████████████████| 1500/1500 [03:30<00:00, 7.13it/s]\n", + "Epoch 5\n", + "adapter path: llama-factory/saves/Qwen2.5-3B-Instruct/checkpoint-175\n", + "[INFO|configuration_utils.py:733] 2024-09-22 09:22:44,984 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 09:22:44,985 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-3B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 2048,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 11008,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 70,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 16,\n", + " \"num_hidden_layers\": 36,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:22:45,236 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:22:45,236 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:22:45,236 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:22:45,236 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:22:45,236 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:22:45,236 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 09:22:45,331 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 09:22:46,375 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 09:22:46,376 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-3B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 2048,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 11008,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 70,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 16,\n", + " \"num_hidden_layers\": 36,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:22:46,628 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:22:46,628 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:22:46,628 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:22:46,628 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:22:46,628 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:22:46,628 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 09:22:46,718 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 09:22:46,976 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 09:22:46,976 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-3B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 2048,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 11008,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 70,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 16,\n", + " \"num_hidden_layers\": 36,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:3634] 2024-09-22 09:22:46,986 >> loading weights file model.safetensors from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/model.safetensors.index.json\n", + "[INFO|modeling_utils.py:1572] 2024-09-22 09:22:46,987 >> Instantiating Qwen2ForCausalLM model under default dtype torch.bfloat16.\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 09:22:46,987 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645\n", + "}\n", + "\n", + "Loading checkpoint shards: 100%|██████████████████| 2/2 [00:35<00:00, 18.00s/it]\n", + "[INFO|modeling_utils.py:4463] 2024-09-22 09:23:23,311 >> All model checkpoint weights were used when initializing Qwen2ForCausalLM.\n", + "\n", + "[INFO|modeling_utils.py:4471] 2024-09-22 09:23:23,311 >> All the weights of Qwen2ForCausalLM were initialized from the model checkpoint at Qwen/Qwen2.5-3B-Instruct.\n", + "If your task is similar to the task the model of the checkpoint was trained on, you can already use Qwen2ForCausalLM for predictions without further training.\n", + "[INFO|configuration_utils.py:993] 2024-09-22 09:23:23,558 >> loading configuration file generation_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/generation_config.json\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 09:23:23,558 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"do_sample\": true,\n", + " \"eos_token_id\": [\n", + " 151645,\n", + " 151643\n", + " ],\n", + " \"pad_token_id\": 151643,\n", + " \"repetition_penalty\": 1.05,\n", + " \"temperature\": 0.7,\n", + " \"top_k\": 20,\n", + " \"top_p\": 0.8\n", + "}\n", + "\n", + "100%|███████████████████████████████████████| 1500/1500 [03:28<00:00, 7.18it/s]\n", + "Epoch 6\n", + "adapter path: llama-factory/saves/Qwen2.5-3B-Instruct/checkpoint-210\n", + "[INFO|configuration_utils.py:733] 2024-09-22 09:27:07,910 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 09:27:07,911 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-3B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 2048,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 11008,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 70,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 16,\n", + " \"num_hidden_layers\": 36,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:27:08,164 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:27:08,164 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:27:08,164 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:27:08,164 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:27:08,164 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:27:08,164 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 09:27:08,256 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 09:27:09,277 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 09:27:09,278 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-3B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 2048,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 11008,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 70,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 16,\n", + " \"num_hidden_layers\": 36,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:27:09,534 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:27:09,534 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:27:09,534 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:27:09,534 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:27:09,534 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:27:09,534 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 09:27:09,622 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 09:27:09,888 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 09:27:09,889 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-3B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 2048,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 11008,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 70,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 16,\n", + " \"num_hidden_layers\": 36,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:3634] 2024-09-22 09:27:09,899 >> loading weights file model.safetensors from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/model.safetensors.index.json\n", + "[INFO|modeling_utils.py:1572] 2024-09-22 09:27:09,899 >> Instantiating Qwen2ForCausalLM model under default dtype torch.bfloat16.\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 09:27:09,900 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645\n", + "}\n", + "\n", + "Loading checkpoint shards: 100%|██████████████████| 2/2 [00:35<00:00, 17.88s/it]\n", + "[INFO|modeling_utils.py:4463] 2024-09-22 09:27:46,691 >> All model checkpoint weights were used when initializing Qwen2ForCausalLM.\n", + "\n", + "[INFO|modeling_utils.py:4471] 2024-09-22 09:27:46,691 >> All the weights of Qwen2ForCausalLM were initialized from the model checkpoint at Qwen/Qwen2.5-3B-Instruct.\n", + "If your task is similar to the task the model of the checkpoint was trained on, you can already use Qwen2ForCausalLM for predictions without further training.\n", + "[INFO|configuration_utils.py:993] 2024-09-22 09:27:46,950 >> loading configuration file generation_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/generation_config.json\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 09:27:46,950 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"do_sample\": true,\n", + " \"eos_token_id\": [\n", + " 151645,\n", + " 151643\n", + " ],\n", + " \"pad_token_id\": 151643,\n", + " \"repetition_penalty\": 1.05,\n", + " \"temperature\": 0.7,\n", + " \"top_k\": 20,\n", + " \"top_p\": 0.8\n", + "}\n", + "\n", + "100%|███████████████████████████████████████| 1500/1500 [03:42<00:00, 6.73it/s]\n", + "Epoch 7\n", + "adapter path: llama-factory/saves/Qwen2.5-3B-Instruct/checkpoint-245\n", + "[INFO|configuration_utils.py:733] 2024-09-22 09:32:05,240 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 09:32:05,240 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-3B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 2048,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 11008,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 70,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 16,\n", + " \"num_hidden_layers\": 36,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:32:05,495 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:32:05,495 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:32:05,495 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:32:05,495 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:32:05,495 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:32:05,495 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 09:32:05,587 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 09:32:06,643 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 09:32:06,643 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-3B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 2048,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 11008,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 70,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 16,\n", + " \"num_hidden_layers\": 36,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:32:06,896 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:32:06,896 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:32:06,896 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:32:06,896 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:32:06,896 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:32:06,896 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 09:32:06,990 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 09:32:07,247 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 09:32:07,248 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-3B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 2048,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 11008,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 70,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 16,\n", + " \"num_hidden_layers\": 36,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:3634] 2024-09-22 09:32:07,258 >> loading weights file model.safetensors from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/model.safetensors.index.json\n", + "[INFO|modeling_utils.py:1572] 2024-09-22 09:32:07,259 >> Instantiating Qwen2ForCausalLM model under default dtype torch.bfloat16.\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 09:32:07,259 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645\n", + "}\n", + "\n", + "Loading checkpoint shards: 100%|██████████████████| 2/2 [00:36<00:00, 18.15s/it]\n", + "[INFO|modeling_utils.py:4463] 2024-09-22 09:32:43,888 >> All model checkpoint weights were used when initializing Qwen2ForCausalLM.\n", + "\n", + "[INFO|modeling_utils.py:4471] 2024-09-22 09:32:43,888 >> All the weights of Qwen2ForCausalLM were initialized from the model checkpoint at Qwen/Qwen2.5-3B-Instruct.\n", + "If your task is similar to the task the model of the checkpoint was trained on, you can already use Qwen2ForCausalLM for predictions without further training.\n", + "[INFO|configuration_utils.py:993] 2024-09-22 09:32:44,207 >> loading configuration file generation_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/generation_config.json\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 09:32:44,207 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"do_sample\": true,\n", + " \"eos_token_id\": [\n", + " 151645,\n", + " 151643\n", + " ],\n", + " \"pad_token_id\": 151643,\n", + " \"repetition_penalty\": 1.05,\n", + " \"temperature\": 0.7,\n", + " \"top_k\": 20,\n", + " \"top_p\": 0.8\n", + "}\n", + "\n", + "100%|███████████████████████████████████████| 1500/1500 [03:22<00:00, 7.42it/s]\n", + "Epoch 8\n", + "adapter path: llama-factory/saves/Qwen2.5-3B-Instruct/checkpoint-280\n", + "[INFO|configuration_utils.py:733] 2024-09-22 09:36:21,368 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 09:36:21,368 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-3B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 2048,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 11008,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 70,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 16,\n", + " \"num_hidden_layers\": 36,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:36:21,624 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:36:21,624 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:36:21,624 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:36:21,624 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:36:21,624 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:36:21,624 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 09:36:21,714 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 09:36:22,762 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 09:36:22,762 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-3B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 2048,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 11008,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 70,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 16,\n", + " \"num_hidden_layers\": 36,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:36:23,015 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:36:23,015 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:36:23,015 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:36:23,015 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:36:23,015 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:36:23,015 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 09:36:23,109 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 09:36:23,372 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 09:36:23,372 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-3B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 2048,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 11008,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 70,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 16,\n", + " \"num_hidden_layers\": 36,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:3634] 2024-09-22 09:36:23,382 >> loading weights file model.safetensors from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/model.safetensors.index.json\n", + "[INFO|modeling_utils.py:1572] 2024-09-22 09:36:23,383 >> Instantiating Qwen2ForCausalLM model under default dtype torch.bfloat16.\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 09:36:23,383 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645\n", + "}\n", + "\n", + "Loading checkpoint shards: 100%|██████████████████| 2/2 [00:36<00:00, 18.11s/it]\n", + "[INFO|modeling_utils.py:4463] 2024-09-22 09:37:00,155 >> All model checkpoint weights were used when initializing Qwen2ForCausalLM.\n", + "\n", + "[INFO|modeling_utils.py:4471] 2024-09-22 09:37:00,156 >> All the weights of Qwen2ForCausalLM were initialized from the model checkpoint at Qwen/Qwen2.5-3B-Instruct.\n", + "If your task is similar to the task the model of the checkpoint was trained on, you can already use Qwen2ForCausalLM for predictions without further training.\n", + "[INFO|configuration_utils.py:993] 2024-09-22 09:37:00,413 >> loading configuration file generation_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/generation_config.json\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 09:37:00,413 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"do_sample\": true,\n", + " \"eos_token_id\": [\n", + " 151645,\n", + " 151643\n", + " ],\n", + " \"pad_token_id\": 151643,\n", + " \"repetition_penalty\": 1.05,\n", + " \"temperature\": 0.7,\n", + " \"top_k\": 20,\n", + " \"top_p\": 0.8\n", + "}\n", + "\n", + "100%|███████████████████████████████████████| 1500/1500 [03:26<00:00, 7.26it/s]\n", + "Epoch 9\n", + "adapter path: llama-factory/saves/Qwen2.5-3B-Instruct/checkpoint-315\n", + "[INFO|configuration_utils.py:733] 2024-09-22 09:40:47,424 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 09:40:47,424 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-3B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 2048,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 11008,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 70,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 16,\n", + " \"num_hidden_layers\": 36,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:40:47,691 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:40:47,691 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:40:47,691 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:40:47,691 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:40:47,691 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:40:47,691 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 09:40:47,784 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 09:40:48,879 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 09:40:48,880 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-3B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 2048,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 11008,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 70,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 16,\n", + " \"num_hidden_layers\": 36,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:40:49,175 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:40:49,175 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:40:49,175 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:40:49,175 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:40:49,175 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:40:49,175 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 09:40:49,269 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 09:40:49,531 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 09:40:49,531 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-3B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 2048,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 11008,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 70,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 16,\n", + " \"num_hidden_layers\": 36,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:3634] 2024-09-22 09:40:49,542 >> loading weights file model.safetensors from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/model.safetensors.index.json\n", + "[INFO|modeling_utils.py:1572] 2024-09-22 09:40:49,542 >> Instantiating Qwen2ForCausalLM model under default dtype torch.bfloat16.\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 09:40:49,542 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645\n", + "}\n", + "\n", + "Loading checkpoint shards: 100%|██████████████████| 2/2 [00:36<00:00, 18.00s/it]\n", + "[INFO|modeling_utils.py:4463] 2024-09-22 09:41:25,923 >> All model checkpoint weights were used when initializing Qwen2ForCausalLM.\n", + "\n", + "[INFO|modeling_utils.py:4471] 2024-09-22 09:41:25,923 >> All the weights of Qwen2ForCausalLM were initialized from the model checkpoint at Qwen/Qwen2.5-3B-Instruct.\n", + "If your task is similar to the task the model of the checkpoint was trained on, you can already use Qwen2ForCausalLM for predictions without further training.\n", + "[INFO|configuration_utils.py:993] 2024-09-22 09:41:26,182 >> loading configuration file generation_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/generation_config.json\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 09:41:26,182 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"do_sample\": true,\n", + " \"eos_token_id\": [\n", + " 151645,\n", + " 151643\n", + " ],\n", + " \"pad_token_id\": 151643,\n", + " \"repetition_penalty\": 1.05,\n", + " \"temperature\": 0.7,\n", + " \"top_k\": 20,\n", + " \"top_p\": 0.8\n", + "}\n", + "\n", + "100%|███████████████████████████████████████| 1500/1500 [03:32<00:00, 7.04it/s]\n", + "Epoch 10\n", + "adapter path: llama-factory/saves/Qwen2.5-3B-Instruct/checkpoint-350\n", + "[INFO|configuration_utils.py:733] 2024-09-22 09:45:13,634 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 09:45:13,634 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-3B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 2048,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 11008,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 70,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 16,\n", + " \"num_hidden_layers\": 36,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:45:13,893 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:45:13,893 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:45:13,893 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:45:13,893 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:45:13,893 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:45:13,893 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 09:45:13,989 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 09:45:15,126 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 09:45:15,127 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-3B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 2048,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 11008,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 70,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 16,\n", + " \"num_hidden_layers\": 36,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:45:15,390 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:45:15,390 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:45:15,390 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:45:15,390 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:45:15,390 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:45:15,390 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 09:45:15,485 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 09:45:15,754 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 09:45:15,754 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-3B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 2048,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 11008,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 70,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 16,\n", + " \"num_hidden_layers\": 36,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:3634] 2024-09-22 09:45:15,764 >> loading weights file model.safetensors from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/model.safetensors.index.json\n", + "[INFO|modeling_utils.py:1572] 2024-09-22 09:45:15,764 >> Instantiating Qwen2ForCausalLM model under default dtype torch.bfloat16.\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 09:45:15,765 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645\n", + "}\n", + "\n", + "Loading checkpoint shards: 100%|██████████████████| 2/2 [00:36<00:00, 18.03s/it]\n", + "[INFO|modeling_utils.py:4463] 2024-09-22 09:45:52,220 >> All model checkpoint weights were used when initializing Qwen2ForCausalLM.\n", + "\n", + "[INFO|modeling_utils.py:4471] 2024-09-22 09:45:52,220 >> All the weights of Qwen2ForCausalLM were initialized from the model checkpoint at Qwen/Qwen2.5-3B-Instruct.\n", + "If your task is similar to the task the model of the checkpoint was trained on, you can already use Qwen2ForCausalLM for predictions without further training.\n", + "[INFO|configuration_utils.py:993] 2024-09-22 09:45:52,484 >> loading configuration file generation_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-3B-Instruct/snapshots/82f42baa094a9600e39ccd80d34058aeeb3abbc1/generation_config.json\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 09:45:52,484 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"do_sample\": true,\n", + " \"eos_token_id\": [\n", + " 151645,\n", + " 151643\n", + " ],\n", + " \"pad_token_id\": 151643,\n", + " \"repetition_penalty\": 1.05,\n", + " \"temperature\": 0.7,\n", + " \"top_k\": 20,\n", + " \"top_p\": 0.8\n", + "}\n", + "\n", + "100%|███████████████████████████████████████| 1500/1500 [04:22<00:00, 5.70it/s]\n", + "Current Directory:\n", + "/home/inflaton/code/logical-reasoning\n", + "Eval Qwen/Qwen2.5-1.5B-Instruct with llama-factory/saves/Qwen2.5-1.5B-Instruct\n", + "loading env vars from: /home/inflaton/code/logical-reasoning/.env\n", + "workding dir: /home/inflaton/code/logical-reasoning\n", + "adding /home/inflaton/code/logical-reasoning to sys.path\n", + "found 10 checkpoints: ['checkpoint-35', 'checkpoint-70', 'checkpoint-105', 'checkpoint-140', 'checkpoint-175', 'checkpoint-210', 'checkpoint-245', 'checkpoint-280', 'checkpoint-315', 'checkpoint-350']\n", + "Epoch 0\n", + "adapter path: None\n", + "[INFO|configuration_utils.py:733] 2024-09-22 09:50:38,086 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 09:50:38,087 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-1.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 1536,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 8960,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 12,\n", + " \"num_hidden_layers\": 28,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:50:38,339 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:50:38,339 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:50:38,339 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:50:38,339 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:50:38,339 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:50:38,339 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 09:50:38,486 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 09:50:39,521 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 09:50:39,522 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-1.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 1536,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 8960,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 12,\n", + " \"num_hidden_layers\": 28,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:50:39,773 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:50:39,773 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:50:39,773 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:50:39,773 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:50:39,773 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:50:39,773 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 09:50:39,866 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 09:50:40,132 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 09:50:40,132 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-1.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 1536,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 8960,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 12,\n", + " \"num_hidden_layers\": 28,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:3634] 2024-09-22 09:50:40,149 >> loading weights file model.safetensors from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/model.safetensors\n", + "[INFO|modeling_utils.py:1572] 2024-09-22 09:50:40,234 >> Instantiating Qwen2ForCausalLM model under default dtype torch.bfloat16.\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 09:50:40,235 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:4463] 2024-09-22 09:51:17,015 >> All model checkpoint weights were used when initializing Qwen2ForCausalLM.\n", + "\n", + "[INFO|modeling_utils.py:4471] 2024-09-22 09:51:17,015 >> All the weights of Qwen2ForCausalLM were initialized from the model checkpoint at Qwen/Qwen2.5-1.5B-Instruct.\n", + "If your task is similar to the task the model of the checkpoint was trained on, you can already use Qwen2ForCausalLM for predictions without further training.\n", + "[INFO|configuration_utils.py:993] 2024-09-22 09:51:17,276 >> loading configuration file generation_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/generation_config.json\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 09:51:17,276 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"do_sample\": true,\n", + " \"eos_token_id\": [\n", + " 151645,\n", + " 151643\n", + " ],\n", + " \"pad_token_id\": 151643,\n", + " \"repetition_penalty\": 1.1,\n", + " \"temperature\": 0.7,\n", + " \"top_k\": 20,\n", + " \"top_p\": 0.8\n", + "}\n", + "\n", + "100%|███████████████████████████████████████| 1500/1500 [03:24<00:00, 7.33it/s]\n", + "Epoch 1\n", + "adapter path: llama-factory/saves/Qwen2.5-1.5B-Instruct/checkpoint-35\n", + "[INFO|configuration_utils.py:733] 2024-09-22 09:54:57,204 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 09:54:57,204 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-1.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 1536,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 8960,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 12,\n", + " \"num_hidden_layers\": 28,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:54:57,465 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:54:57,466 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:54:57,466 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:54:57,466 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:54:57,466 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:54:57,466 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 09:54:57,568 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 09:54:58,627 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 09:54:58,627 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-1.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 1536,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 8960,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 12,\n", + " \"num_hidden_layers\": 28,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:54:58,883 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:54:58,883 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:54:58,883 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:54:58,883 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:54:58,883 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:54:58,883 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 09:54:58,975 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 09:54:59,253 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 09:54:59,253 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-1.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 1536,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 8960,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 12,\n", + " \"num_hidden_layers\": 28,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:3634] 2024-09-22 09:54:59,270 >> loading weights file model.safetensors from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/model.safetensors\n", + "[INFO|modeling_utils.py:1572] 2024-09-22 09:54:59,273 >> Instantiating Qwen2ForCausalLM model under default dtype torch.bfloat16.\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 09:54:59,274 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:4463] 2024-09-22 09:55:35,397 >> All model checkpoint weights were used when initializing Qwen2ForCausalLM.\n", + "\n", + "[INFO|modeling_utils.py:4471] 2024-09-22 09:55:35,397 >> All the weights of Qwen2ForCausalLM were initialized from the model checkpoint at Qwen/Qwen2.5-1.5B-Instruct.\n", + "If your task is similar to the task the model of the checkpoint was trained on, you can already use Qwen2ForCausalLM for predictions without further training.\n", + "[INFO|configuration_utils.py:993] 2024-09-22 09:55:35,657 >> loading configuration file generation_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/generation_config.json\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 09:55:35,657 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"do_sample\": true,\n", + " \"eos_token_id\": [\n", + " 151645,\n", + " 151643\n", + " ],\n", + " \"pad_token_id\": 151643,\n", + " \"repetition_penalty\": 1.1,\n", + " \"temperature\": 0.7,\n", + " \"top_k\": 20,\n", + " \"top_p\": 0.8\n", + "}\n", + "\n", + "100%|███████████████████████████████████████| 1500/1500 [02:29<00:00, 10.01it/s]\n", + "Epoch 2\n", + "adapter path: llama-factory/saves/Qwen2.5-1.5B-Instruct/checkpoint-70\n", + "[INFO|configuration_utils.py:733] 2024-09-22 09:58:22,352 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 09:58:22,353 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-1.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 1536,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 8960,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 12,\n", + " \"num_hidden_layers\": 28,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:58:22,649 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:58:22,649 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:58:22,649 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:58:22,649 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:58:22,649 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:58:22,649 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 09:58:22,739 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 09:58:23,772 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 09:58:23,772 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-1.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 1536,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 8960,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 12,\n", + " \"num_hidden_layers\": 28,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:58:24,030 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:58:24,030 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:58:24,030 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:58:24,030 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:58:24,030 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 09:58:24,030 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 09:58:24,124 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 09:58:24,389 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 09:58:24,390 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-1.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 1536,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 8960,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 12,\n", + " \"num_hidden_layers\": 28,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:3634] 2024-09-22 09:58:24,400 >> loading weights file model.safetensors from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/model.safetensors\n", + "[INFO|modeling_utils.py:1572] 2024-09-22 09:58:24,403 >> Instantiating Qwen2ForCausalLM model under default dtype torch.bfloat16.\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 09:58:24,404 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:4463] 2024-09-22 09:58:53,422 >> All model checkpoint weights were used when initializing Qwen2ForCausalLM.\n", + "\n", + "[INFO|modeling_utils.py:4471] 2024-09-22 09:58:53,422 >> All the weights of Qwen2ForCausalLM were initialized from the model checkpoint at Qwen/Qwen2.5-1.5B-Instruct.\n", + "If your task is similar to the task the model of the checkpoint was trained on, you can already use Qwen2ForCausalLM for predictions without further training.\n", + "[INFO|configuration_utils.py:993] 2024-09-22 09:58:53,684 >> loading configuration file generation_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/generation_config.json\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 09:58:53,684 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"do_sample\": true,\n", + " \"eos_token_id\": [\n", + " 151645,\n", + " 151643\n", + " ],\n", + " \"pad_token_id\": 151643,\n", + " \"repetition_penalty\": 1.1,\n", + " \"temperature\": 0.7,\n", + " \"top_k\": 20,\n", + " \"top_p\": 0.8\n", + "}\n", + "\n", + "100%|███████████████████████████████████████| 1500/1500 [02:24<00:00, 10.40it/s]\n", + "Epoch 3\n", + "adapter path: llama-factory/saves/Qwen2.5-1.5B-Instruct/checkpoint-105\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:01:33,063 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:01:33,064 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-1.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 1536,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 8960,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 12,\n", + " \"num_hidden_layers\": 28,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:01:33,327 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:01:33,327 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:01:33,327 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:01:33,327 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:01:33,327 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:01:33,327 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 10:01:33,420 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:01:34,454 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:01:34,455 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-1.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 1536,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 8960,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 12,\n", + " \"num_hidden_layers\": 28,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:01:34,741 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:01:34,741 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:01:34,741 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:01:34,741 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:01:34,741 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:01:34,741 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 10:01:34,831 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:01:35,096 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:01:35,096 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-1.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 1536,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 8960,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 12,\n", + " \"num_hidden_layers\": 28,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:3634] 2024-09-22 10:01:35,106 >> loading weights file model.safetensors from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/model.safetensors\n", + "[INFO|modeling_utils.py:1572] 2024-09-22 10:01:35,109 >> Instantiating Qwen2ForCausalLM model under default dtype torch.bfloat16.\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 10:01:35,110 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:4463] 2024-09-22 10:02:02,883 >> All model checkpoint weights were used when initializing Qwen2ForCausalLM.\n", + "\n", + "[INFO|modeling_utils.py:4471] 2024-09-22 10:02:02,884 >> All the weights of Qwen2ForCausalLM were initialized from the model checkpoint at Qwen/Qwen2.5-1.5B-Instruct.\n", + "If your task is similar to the task the model of the checkpoint was trained on, you can already use Qwen2ForCausalLM for predictions without further training.\n", + "[INFO|configuration_utils.py:993] 2024-09-22 10:02:03,140 >> loading configuration file generation_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/generation_config.json\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 10:02:03,140 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"do_sample\": true,\n", + " \"eos_token_id\": [\n", + " 151645,\n", + " 151643\n", + " ],\n", + " \"pad_token_id\": 151643,\n", + " \"repetition_penalty\": 1.1,\n", + " \"temperature\": 0.7,\n", + " \"top_k\": 20,\n", + " \"top_p\": 0.8\n", + "}\n", + "\n", + "100%|███████████████████████████████████████| 1500/1500 [02:36<00:00, 9.56it/s]\n", + "Epoch 4\n", + "adapter path: llama-factory/saves/Qwen2.5-1.5B-Instruct/checkpoint-140\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:04:57,320 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:04:57,321 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-1.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 1536,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 8960,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 12,\n", + " \"num_hidden_layers\": 28,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:04:57,581 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:04:57,581 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:04:57,581 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:04:57,581 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:04:57,581 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:04:57,581 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 10:04:57,676 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:04:58,715 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:04:58,716 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-1.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 1536,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 8960,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 12,\n", + " \"num_hidden_layers\": 28,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:04:58,977 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:04:58,977 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:04:58,977 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:04:58,977 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:04:58,977 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:04:58,977 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 10:04:59,068 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:04:59,334 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:04:59,334 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-1.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 1536,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 8960,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 12,\n", + " \"num_hidden_layers\": 28,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:3634] 2024-09-22 10:04:59,344 >> loading weights file model.safetensors from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/model.safetensors\n", + "[INFO|modeling_utils.py:1572] 2024-09-22 10:04:59,347 >> Instantiating Qwen2ForCausalLM model under default dtype torch.bfloat16.\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 10:04:59,348 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:4463] 2024-09-22 10:05:27,666 >> All model checkpoint weights were used when initializing Qwen2ForCausalLM.\n", + "\n", + "[INFO|modeling_utils.py:4471] 2024-09-22 10:05:27,666 >> All the weights of Qwen2ForCausalLM were initialized from the model checkpoint at Qwen/Qwen2.5-1.5B-Instruct.\n", + "If your task is similar to the task the model of the checkpoint was trained on, you can already use Qwen2ForCausalLM for predictions without further training.\n", + "[INFO|configuration_utils.py:993] 2024-09-22 10:05:27,923 >> loading configuration file generation_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/generation_config.json\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 10:05:27,923 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"do_sample\": true,\n", + " \"eos_token_id\": [\n", + " 151645,\n", + " 151643\n", + " ],\n", + " \"pad_token_id\": 151643,\n", + " \"repetition_penalty\": 1.1,\n", + " \"temperature\": 0.7,\n", + " \"top_k\": 20,\n", + " \"top_p\": 0.8\n", + "}\n", + "\n", + "100%|███████████████████████████████████████| 1500/1500 [02:39<00:00, 9.38it/s]\n", + "Epoch 5\n", + "adapter path: llama-factory/saves/Qwen2.5-1.5B-Instruct/checkpoint-175\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:08:25,461 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:08:25,461 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-1.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 1536,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 8960,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 12,\n", + " \"num_hidden_layers\": 28,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:08:25,726 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:08:25,726 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:08:25,726 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:08:25,726 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:08:25,726 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:08:25,726 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 10:08:25,820 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:08:26,927 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:08:26,928 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-1.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 1536,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 8960,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 12,\n", + " \"num_hidden_layers\": 28,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:08:27,187 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:08:27,187 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:08:27,187 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:08:27,187 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:08:27,187 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:08:27,187 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 10:08:27,276 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:08:27,543 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:08:27,543 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-1.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 1536,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 8960,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 12,\n", + " \"num_hidden_layers\": 28,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:3634] 2024-09-22 10:08:27,554 >> loading weights file model.safetensors from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/model.safetensors\n", + "[INFO|modeling_utils.py:1572] 2024-09-22 10:08:27,558 >> Instantiating Qwen2ForCausalLM model under default dtype torch.bfloat16.\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 10:08:27,559 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:4463] 2024-09-22 10:08:55,681 >> All model checkpoint weights were used when initializing Qwen2ForCausalLM.\n", + "\n", + "[INFO|modeling_utils.py:4471] 2024-09-22 10:08:55,681 >> All the weights of Qwen2ForCausalLM were initialized from the model checkpoint at Qwen/Qwen2.5-1.5B-Instruct.\n", + "If your task is similar to the task the model of the checkpoint was trained on, you can already use Qwen2ForCausalLM for predictions without further training.\n", + "[INFO|configuration_utils.py:993] 2024-09-22 10:08:55,939 >> loading configuration file generation_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/generation_config.json\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 10:08:55,939 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"do_sample\": true,\n", + " \"eos_token_id\": [\n", + " 151645,\n", + " 151643\n", + " ],\n", + " \"pad_token_id\": 151643,\n", + " \"repetition_penalty\": 1.1,\n", + " \"temperature\": 0.7,\n", + " \"top_k\": 20,\n", + " \"top_p\": 0.8\n", + "}\n", + "\n", + "100%|███████████████████████████████████████| 1500/1500 [02:19<00:00, 10.76it/s]\n", + "Epoch 6\n", + "adapter path: llama-factory/saves/Qwen2.5-1.5B-Instruct/checkpoint-210\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:11:25,983 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:11:25,984 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-1.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 1536,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 8960,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 12,\n", + " \"num_hidden_layers\": 28,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:11:26,243 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:11:26,243 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:11:26,243 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:11:26,243 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:11:26,243 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:11:26,243 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 10:11:26,340 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:11:27,427 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:11:27,427 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-1.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 1536,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 8960,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 12,\n", + " \"num_hidden_layers\": 28,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:11:27,685 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:11:27,685 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:11:27,685 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:11:27,685 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:11:27,685 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:11:27,685 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 10:11:27,775 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:11:28,045 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:11:28,046 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-1.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 1536,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 8960,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 12,\n", + " \"num_hidden_layers\": 28,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:3634] 2024-09-22 10:11:28,056 >> loading weights file model.safetensors from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/model.safetensors\n", + "[INFO|modeling_utils.py:1572] 2024-09-22 10:11:28,059 >> Instantiating Qwen2ForCausalLM model under default dtype torch.bfloat16.\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 10:11:28,060 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:4463] 2024-09-22 10:11:57,233 >> All model checkpoint weights were used when initializing Qwen2ForCausalLM.\n", + "\n", + "[INFO|modeling_utils.py:4471] 2024-09-22 10:11:57,233 >> All the weights of Qwen2ForCausalLM were initialized from the model checkpoint at Qwen/Qwen2.5-1.5B-Instruct.\n", + "If your task is similar to the task the model of the checkpoint was trained on, you can already use Qwen2ForCausalLM for predictions without further training.\n", + "[INFO|configuration_utils.py:993] 2024-09-22 10:11:57,545 >> loading configuration file generation_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/generation_config.json\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 10:11:57,545 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"do_sample\": true,\n", + " \"eos_token_id\": [\n", + " 151645,\n", + " 151643\n", + " ],\n", + " \"pad_token_id\": 151643,\n", + " \"repetition_penalty\": 1.1,\n", + " \"temperature\": 0.7,\n", + " \"top_k\": 20,\n", + " \"top_p\": 0.8\n", + "}\n", + "\n", + "100%|███████████████████████████████████████| 1500/1500 [02:29<00:00, 10.01it/s]\n", + "Epoch 7\n", + "adapter path: llama-factory/saves/Qwen2.5-1.5B-Instruct/checkpoint-245\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:14:42,252 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:14:42,253 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-1.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 1536,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 8960,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 12,\n", + " \"num_hidden_layers\": 28,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:14:42,504 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:14:42,504 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:14:42,504 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:14:42,504 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:14:42,504 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:14:42,504 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 10:14:42,602 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:14:43,612 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:14:43,613 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-1.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 1536,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 8960,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 12,\n", + " \"num_hidden_layers\": 28,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:14:43,861 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:14:43,861 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:14:43,861 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:14:43,861 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:14:43,861 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:14:43,861 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 10:14:43,952 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:14:44,210 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:14:44,211 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-1.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 1536,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 8960,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 12,\n", + " \"num_hidden_layers\": 28,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:3634] 2024-09-22 10:14:44,220 >> loading weights file model.safetensors from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/model.safetensors\n", + "[INFO|modeling_utils.py:1572] 2024-09-22 10:14:44,224 >> Instantiating Qwen2ForCausalLM model under default dtype torch.bfloat16.\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 10:14:44,225 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:4463] 2024-09-22 10:15:11,898 >> All model checkpoint weights were used when initializing Qwen2ForCausalLM.\n", + "\n", + "[INFO|modeling_utils.py:4471] 2024-09-22 10:15:11,898 >> All the weights of Qwen2ForCausalLM were initialized from the model checkpoint at Qwen/Qwen2.5-1.5B-Instruct.\n", + "If your task is similar to the task the model of the checkpoint was trained on, you can already use Qwen2ForCausalLM for predictions without further training.\n", + "[INFO|configuration_utils.py:993] 2024-09-22 10:15:12,147 >> loading configuration file generation_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/generation_config.json\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 10:15:12,147 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"do_sample\": true,\n", + " \"eos_token_id\": [\n", + " 151645,\n", + " 151643\n", + " ],\n", + " \"pad_token_id\": 151643,\n", + " \"repetition_penalty\": 1.1,\n", + " \"temperature\": 0.7,\n", + " \"top_k\": 20,\n", + " \"top_p\": 0.8\n", + "}\n", + "\n", + "100%|███████████████████████████████████████| 1500/1500 [02:25<00:00, 10.34it/s]\n", + "Epoch 8\n", + "adapter path: llama-factory/saves/Qwen2.5-1.5B-Instruct/checkpoint-280\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:17:51,538 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:17:51,538 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-1.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 1536,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 8960,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 12,\n", + " \"num_hidden_layers\": 28,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:17:51,826 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:17:51,826 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:17:51,826 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:17:51,826 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:17:51,826 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:17:51,826 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 10:17:51,934 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:17:52,992 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:17:52,993 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-1.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 1536,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 8960,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 12,\n", + " \"num_hidden_layers\": 28,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:17:53,246 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:17:53,246 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:17:53,246 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:17:53,246 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:17:53,246 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:17:53,246 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 10:17:53,334 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:17:53,600 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:17:53,600 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-1.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 1536,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 8960,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 12,\n", + " \"num_hidden_layers\": 28,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:3634] 2024-09-22 10:17:53,610 >> loading weights file model.safetensors from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/model.safetensors\n", + "[INFO|modeling_utils.py:1572] 2024-09-22 10:17:53,613 >> Instantiating Qwen2ForCausalLM model under default dtype torch.bfloat16.\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 10:17:53,614 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:4463] 2024-09-22 10:18:21,313 >> All model checkpoint weights were used when initializing Qwen2ForCausalLM.\n", + "\n", + "[INFO|modeling_utils.py:4471] 2024-09-22 10:18:21,313 >> All the weights of Qwen2ForCausalLM were initialized from the model checkpoint at Qwen/Qwen2.5-1.5B-Instruct.\n", + "If your task is similar to the task the model of the checkpoint was trained on, you can already use Qwen2ForCausalLM for predictions without further training.\n", + "[INFO|configuration_utils.py:993] 2024-09-22 10:18:21,578 >> loading configuration file generation_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/generation_config.json\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 10:18:21,578 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"do_sample\": true,\n", + " \"eos_token_id\": [\n", + " 151645,\n", + " 151643\n", + " ],\n", + " \"pad_token_id\": 151643,\n", + " \"repetition_penalty\": 1.1,\n", + " \"temperature\": 0.7,\n", + " \"top_k\": 20,\n", + " \"top_p\": 0.8\n", + "}\n", + "\n", + "100%|███████████████████████████████████████| 1500/1500 [02:54<00:00, 8.58it/s]\n", + "Epoch 9\n", + "adapter path: llama-factory/saves/Qwen2.5-1.5B-Instruct/checkpoint-315\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:21:35,759 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:21:35,759 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-1.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 1536,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 8960,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 12,\n", + " \"num_hidden_layers\": 28,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:21:36,035 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:21:36,035 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:21:36,035 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:21:36,035 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:21:36,035 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:21:36,035 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 10:21:36,131 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:21:37,178 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:21:37,178 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-1.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 1536,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 8960,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 12,\n", + " \"num_hidden_layers\": 28,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:21:37,441 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:21:37,441 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:21:37,441 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:21:37,441 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:21:37,441 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:21:37,441 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 10:21:37,528 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:21:37,800 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:21:37,800 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-1.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 1536,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 8960,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 12,\n", + " \"num_hidden_layers\": 28,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:3634] 2024-09-22 10:21:37,810 >> loading weights file model.safetensors from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/model.safetensors\n", + "[INFO|modeling_utils.py:1572] 2024-09-22 10:21:37,814 >> Instantiating Qwen2ForCausalLM model under default dtype torch.bfloat16.\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 10:21:37,814 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:4463] 2024-09-22 10:22:05,744 >> All model checkpoint weights were used when initializing Qwen2ForCausalLM.\n", + "\n", + "[INFO|modeling_utils.py:4471] 2024-09-22 10:22:05,745 >> All the weights of Qwen2ForCausalLM were initialized from the model checkpoint at Qwen/Qwen2.5-1.5B-Instruct.\n", + "If your task is similar to the task the model of the checkpoint was trained on, you can already use Qwen2ForCausalLM for predictions without further training.\n", + "[INFO|configuration_utils.py:993] 2024-09-22 10:22:06,017 >> loading configuration file generation_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/generation_config.json\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 10:22:06,018 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"do_sample\": true,\n", + " \"eos_token_id\": [\n", + " 151645,\n", + " 151643\n", + " ],\n", + " \"pad_token_id\": 151643,\n", + " \"repetition_penalty\": 1.1,\n", + " \"temperature\": 0.7,\n", + " \"top_k\": 20,\n", + " \"top_p\": 0.8\n", + "}\n", + "\n", + "100%|███████████████████████████████████████| 1500/1500 [02:30<00:00, 9.98it/s]\n", + "Epoch 10\n", + "adapter path: llama-factory/saves/Qwen2.5-1.5B-Instruct/checkpoint-350\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:24:49,119 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:24:49,120 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-1.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 1536,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 8960,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 12,\n", + " \"num_hidden_layers\": 28,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:24:49,383 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:24:49,384 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:24:49,384 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:24:49,384 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:24:49,384 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:24:49,384 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 10:24:49,492 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:24:50,564 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:24:50,564 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-1.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 1536,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 8960,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 12,\n", + " \"num_hidden_layers\": 28,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:24:50,823 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:24:50,823 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:24:50,823 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:24:50,823 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:24:50,823 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:24:50,823 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 10:24:50,915 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:24:51,195 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:24:51,195 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-1.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 1536,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 8960,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 12,\n", + " \"num_hidden_layers\": 28,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:3634] 2024-09-22 10:24:51,205 >> loading weights file model.safetensors from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/model.safetensors\n", + "[INFO|modeling_utils.py:1572] 2024-09-22 10:24:51,209 >> Instantiating Qwen2ForCausalLM model under default dtype torch.bfloat16.\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 10:24:51,210 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:4463] 2024-09-22 10:25:20,709 >> All model checkpoint weights were used when initializing Qwen2ForCausalLM.\n", + "\n", + "[INFO|modeling_utils.py:4471] 2024-09-22 10:25:20,710 >> All the weights of Qwen2ForCausalLM were initialized from the model checkpoint at Qwen/Qwen2.5-1.5B-Instruct.\n", + "If your task is similar to the task the model of the checkpoint was trained on, you can already use Qwen2ForCausalLM for predictions without further training.\n", + "[INFO|configuration_utils.py:993] 2024-09-22 10:25:20,970 >> loading configuration file generation_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-1.5B-Instruct/snapshots/5fee7c4ed634dc66c6e318c8ac2897b8b9154536/generation_config.json\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 10:25:20,970 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"do_sample\": true,\n", + " \"eos_token_id\": [\n", + " 151645,\n", + " 151643\n", + " ],\n", + " \"pad_token_id\": 151643,\n", + " \"repetition_penalty\": 1.1,\n", + " \"temperature\": 0.7,\n", + " \"top_k\": 20,\n", + " \"top_p\": 0.8\n", + "}\n", + "\n", + "100%|███████████████████████████████████████| 1500/1500 [03:04<00:00, 8.11it/s]\n", + "Current Directory:\n", + "/home/inflaton/code/logical-reasoning\n", + "Eval Qwen/Qwen2.5-0.5B-Instruct with llama-factory/saves/Qwen2.5-0.5B-Instruct\n", + "loading env vars from: /home/inflaton/code/logical-reasoning/.env\n", + "workding dir: /home/inflaton/code/logical-reasoning\n", + "adding /home/inflaton/code/logical-reasoning to sys.path\n", + "found 10 checkpoints: ['checkpoint-35', 'checkpoint-70', 'checkpoint-105', 'checkpoint-140', 'checkpoint-175', 'checkpoint-210', 'checkpoint-245', 'checkpoint-280', 'checkpoint-315', 'checkpoint-350']\n", + "Epoch 0\n", + "adapter path: None\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:28:36,672 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:28:36,673 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-0.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 896,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 4864,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 14,\n", + " \"num_hidden_layers\": 24,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:28:36,950 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:28:36,950 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:28:36,950 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:28:36,950 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:28:36,950 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:28:36,950 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 10:28:37,107 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:28:38,136 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:28:38,136 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-0.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 896,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 4864,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 14,\n", + " \"num_hidden_layers\": 24,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:28:38,383 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:28:38,383 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:28:38,383 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:28:38,383 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:28:38,383 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:28:38,383 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 10:28:38,478 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:28:38,754 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:28:38,755 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-0.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 896,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 4864,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 14,\n", + " \"num_hidden_layers\": 24,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:3634] 2024-09-22 10:28:38,771 >> loading weights file model.safetensors from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/model.safetensors\n", + "[INFO|modeling_utils.py:1572] 2024-09-22 10:28:38,777 >> Instantiating Qwen2ForCausalLM model under default dtype torch.bfloat16.\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 10:28:38,778 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:4463] 2024-09-22 10:29:03,239 >> All model checkpoint weights were used when initializing Qwen2ForCausalLM.\n", + "\n", + "[INFO|modeling_utils.py:4471] 2024-09-22 10:29:03,239 >> All the weights of Qwen2ForCausalLM were initialized from the model checkpoint at Qwen/Qwen2.5-0.5B-Instruct.\n", + "If your task is similar to the task the model of the checkpoint was trained on, you can already use Qwen2ForCausalLM for predictions without further training.\n", + "[INFO|configuration_utils.py:993] 2024-09-22 10:29:03,497 >> loading configuration file generation_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/generation_config.json\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 10:29:03,497 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"do_sample\": true,\n", + " \"eos_token_id\": [\n", + " 151645,\n", + " 151643\n", + " ],\n", + " \"pad_token_id\": 151643,\n", + " \"repetition_penalty\": 1.1,\n", + " \"temperature\": 0.7,\n", + " \"top_k\": 20,\n", + " \"top_p\": 0.8\n", + "}\n", + "\n", + "100%|███████████████████████████████████████| 1500/1500 [01:48<00:00, 13.81it/s]\n", + "/home/inflaton/miniconda3/envs/llm-finetuning/lib/python3.11/site-packages/sklearn/metrics/_classification.py:1517: UndefinedMetricWarning: Precision is ill-defined and being set to 0.0 in labels with no predicted samples. Use `zero_division` parameter to control this behavior.\n", + " _warn_prf(average, modifier, f\"{metric.capitalize()} is\", len(result))\n", + "Epoch 1\n", + "adapter path: llama-factory/saves/Qwen2.5-0.5B-Instruct/checkpoint-35\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:31:02,339 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:31:02,339 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-0.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 896,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 4864,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 14,\n", + " \"num_hidden_layers\": 24,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:31:02,609 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:31:02,609 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:31:02,609 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:31:02,609 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:31:02,609 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:31:02,609 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 10:31:02,707 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:31:04,518 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:31:04,519 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-0.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 896,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 4864,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 14,\n", + " \"num_hidden_layers\": 24,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:31:04,779 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:31:04,779 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:31:04,779 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:31:04,779 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:31:04,779 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:31:04,779 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 10:31:04,898 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:31:05,173 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:31:05,173 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-0.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 896,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 4864,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 14,\n", + " \"num_hidden_layers\": 24,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:3634] 2024-09-22 10:31:05,184 >> loading weights file model.safetensors from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/model.safetensors\n", + "[INFO|modeling_utils.py:1572] 2024-09-22 10:31:05,189 >> Instantiating Qwen2ForCausalLM model under default dtype torch.bfloat16.\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 10:31:05,190 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:4463] 2024-09-22 10:31:38,898 >> All model checkpoint weights were used when initializing Qwen2ForCausalLM.\n", + "\n", + "[INFO|modeling_utils.py:4471] 2024-09-22 10:31:38,898 >> All the weights of Qwen2ForCausalLM were initialized from the model checkpoint at Qwen/Qwen2.5-0.5B-Instruct.\n", + "If your task is similar to the task the model of the checkpoint was trained on, you can already use Qwen2ForCausalLM for predictions without further training.\n", + "[INFO|configuration_utils.py:993] 2024-09-22 10:31:39,153 >> loading configuration file generation_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/generation_config.json\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 10:31:39,154 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"do_sample\": true,\n", + " \"eos_token_id\": [\n", + " 151645,\n", + " 151643\n", + " ],\n", + " \"pad_token_id\": 151643,\n", + " \"repetition_penalty\": 1.1,\n", + " \"temperature\": 0.7,\n", + " \"top_k\": 20,\n", + " \"top_p\": 0.8\n", + "}\n", + "\n", + "100%|███████████████████████████████████████| 1500/1500 [01:10<00:00, 21.15it/s]\n", + "Epoch 2\n", + "adapter path: llama-factory/saves/Qwen2.5-0.5B-Instruct/checkpoint-70\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:32:56,366 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:32:56,366 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-0.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 896,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 4864,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 14,\n", + " \"num_hidden_layers\": 24,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:32:56,640 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:32:56,640 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:32:56,640 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:32:56,640 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:32:56,640 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:32:56,640 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 10:32:56,736 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:32:57,817 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:32:57,818 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-0.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 896,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 4864,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 14,\n", + " \"num_hidden_layers\": 24,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:32:58,065 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:32:58,065 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:32:58,065 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:32:58,065 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:32:58,065 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:32:58,066 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 10:32:58,159 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:32:58,421 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:32:58,421 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-0.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 896,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 4864,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 14,\n", + " \"num_hidden_layers\": 24,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:3634] 2024-09-22 10:32:58,432 >> loading weights file model.safetensors from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/model.safetensors\n", + "[INFO|modeling_utils.py:1572] 2024-09-22 10:32:58,435 >> Instantiating Qwen2ForCausalLM model under default dtype torch.bfloat16.\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 10:32:58,436 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:4463] 2024-09-22 10:33:22,411 >> All model checkpoint weights were used when initializing Qwen2ForCausalLM.\n", + "\n", + "[INFO|modeling_utils.py:4471] 2024-09-22 10:33:22,412 >> All the weights of Qwen2ForCausalLM were initialized from the model checkpoint at Qwen/Qwen2.5-0.5B-Instruct.\n", + "If your task is similar to the task the model of the checkpoint was trained on, you can already use Qwen2ForCausalLM for predictions without further training.\n", + "[INFO|configuration_utils.py:993] 2024-09-22 10:33:22,665 >> loading configuration file generation_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/generation_config.json\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 10:33:22,665 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"do_sample\": true,\n", + " \"eos_token_id\": [\n", + " 151645,\n", + " 151643\n", + " ],\n", + " \"pad_token_id\": 151643,\n", + " \"repetition_penalty\": 1.1,\n", + " \"temperature\": 0.7,\n", + " \"top_k\": 20,\n", + " \"top_p\": 0.8\n", + "}\n", + "\n", + "100%|███████████████████████████████████████| 1500/1500 [01:17<00:00, 19.42it/s]\n", + "Epoch 3\n", + "adapter path: llama-factory/saves/Qwen2.5-0.5B-Instruct/checkpoint-105\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:34:53,861 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:34:53,862 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-0.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 896,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 4864,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 14,\n", + " \"num_hidden_layers\": 24,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:34:54,131 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:34:54,131 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:34:54,131 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:34:54,131 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:34:54,131 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:34:54,131 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 10:34:54,228 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:34:55,349 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:34:55,349 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-0.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 896,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 4864,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 14,\n", + " \"num_hidden_layers\": 24,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:34:55,627 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:34:55,627 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:34:55,627 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:34:55,627 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:34:55,627 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:34:55,627 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 10:34:55,723 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:34:56,055 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:34:56,055 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-0.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 896,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 4864,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 14,\n", + " \"num_hidden_layers\": 24,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:3634] 2024-09-22 10:34:56,066 >> loading weights file model.safetensors from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/model.safetensors\n", + "[INFO|modeling_utils.py:1572] 2024-09-22 10:34:56,070 >> Instantiating Qwen2ForCausalLM model under default dtype torch.bfloat16.\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 10:34:56,071 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:4463] 2024-09-22 10:35:20,895 >> All model checkpoint weights were used when initializing Qwen2ForCausalLM.\n", + "\n", + "[INFO|modeling_utils.py:4471] 2024-09-22 10:35:20,895 >> All the weights of Qwen2ForCausalLM were initialized from the model checkpoint at Qwen/Qwen2.5-0.5B-Instruct.\n", + "If your task is similar to the task the model of the checkpoint was trained on, you can already use Qwen2ForCausalLM for predictions without further training.\n", + "[INFO|configuration_utils.py:993] 2024-09-22 10:35:21,169 >> loading configuration file generation_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/generation_config.json\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 10:35:21,169 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"do_sample\": true,\n", + " \"eos_token_id\": [\n", + " 151645,\n", + " 151643\n", + " ],\n", + " \"pad_token_id\": 151643,\n", + " \"repetition_penalty\": 1.1,\n", + " \"temperature\": 0.7,\n", + " \"top_k\": 20,\n", + " \"top_p\": 0.8\n", + "}\n", + "\n", + "100%|███████████████████████████████████████| 1500/1500 [01:20<00:00, 18.55it/s]\n", + "Epoch 4\n", + "adapter path: llama-factory/saves/Qwen2.5-0.5B-Instruct/checkpoint-140\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:36:58,434 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:36:58,434 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-0.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 896,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 4864,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 14,\n", + " \"num_hidden_layers\": 24,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:36:58,687 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:36:58,687 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:36:58,687 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:36:58,687 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:36:58,687 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:36:58,687 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 10:36:58,786 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:36:59,870 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:36:59,870 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-0.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 896,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 4864,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 14,\n", + " \"num_hidden_layers\": 24,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:37:00,121 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:37:00,121 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:37:00,121 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:37:00,121 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:37:00,121 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:37:00,121 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 10:37:00,210 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:37:00,474 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:37:00,474 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-0.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 896,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 4864,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 14,\n", + " \"num_hidden_layers\": 24,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:3634] 2024-09-22 10:37:00,484 >> loading weights file model.safetensors from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/model.safetensors\n", + "[INFO|modeling_utils.py:1572] 2024-09-22 10:37:00,487 >> Instantiating Qwen2ForCausalLM model under default dtype torch.bfloat16.\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 10:37:00,488 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:4463] 2024-09-22 10:37:25,505 >> All model checkpoint weights were used when initializing Qwen2ForCausalLM.\n", + "\n", + "[INFO|modeling_utils.py:4471] 2024-09-22 10:37:25,505 >> All the weights of Qwen2ForCausalLM were initialized from the model checkpoint at Qwen/Qwen2.5-0.5B-Instruct.\n", + "If your task is similar to the task the model of the checkpoint was trained on, you can already use Qwen2ForCausalLM for predictions without further training.\n", + "[INFO|configuration_utils.py:993] 2024-09-22 10:37:25,758 >> loading configuration file generation_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/generation_config.json\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 10:37:25,758 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"do_sample\": true,\n", + " \"eos_token_id\": [\n", + " 151645,\n", + " 151643\n", + " ],\n", + " \"pad_token_id\": 151643,\n", + " \"repetition_penalty\": 1.1,\n", + " \"temperature\": 0.7,\n", + " \"top_k\": 20,\n", + " \"top_p\": 0.8\n", + "}\n", + "\n", + "100%|███████████████████████████████████████| 1500/1500 [01:21<00:00, 18.42it/s]\n", + "Epoch 5\n", + "adapter path: llama-factory/saves/Qwen2.5-0.5B-Instruct/checkpoint-175\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:38:58,338 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:38:58,339 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-0.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 896,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 4864,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 14,\n", + " \"num_hidden_layers\": 24,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:38:58,601 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:38:58,601 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:38:58,601 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:38:58,601 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:38:58,601 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:38:58,601 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 10:38:58,702 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:38:59,788 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:38:59,789 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-0.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 896,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 4864,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 14,\n", + " \"num_hidden_layers\": 24,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:39:00,055 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:39:00,056 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:39:00,056 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:39:00,056 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:39:00,056 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:39:00,056 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 10:39:00,150 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:39:00,490 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:39:00,490 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-0.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 896,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 4864,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 14,\n", + " \"num_hidden_layers\": 24,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:3634] 2024-09-22 10:39:00,501 >> loading weights file model.safetensors from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/model.safetensors\n", + "[INFO|modeling_utils.py:1572] 2024-09-22 10:39:00,504 >> Instantiating Qwen2ForCausalLM model under default dtype torch.bfloat16.\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 10:39:00,505 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:4463] 2024-09-22 10:39:24,238 >> All model checkpoint weights were used when initializing Qwen2ForCausalLM.\n", + "\n", + "[INFO|modeling_utils.py:4471] 2024-09-22 10:39:24,238 >> All the weights of Qwen2ForCausalLM were initialized from the model checkpoint at Qwen/Qwen2.5-0.5B-Instruct.\n", + "If your task is similar to the task the model of the checkpoint was trained on, you can already use Qwen2ForCausalLM for predictions without further training.\n", + "[INFO|configuration_utils.py:993] 2024-09-22 10:39:24,529 >> loading configuration file generation_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/generation_config.json\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 10:39:24,530 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"do_sample\": true,\n", + " \"eos_token_id\": [\n", + " 151645,\n", + " 151643\n", + " ],\n", + " \"pad_token_id\": 151643,\n", + " \"repetition_penalty\": 1.1,\n", + " \"temperature\": 0.7,\n", + " \"top_k\": 20,\n", + " \"top_p\": 0.8\n", + "}\n", + "\n", + "100%|███████████████████████████████████████| 1500/1500 [01:19<00:00, 18.96it/s]\n", + "Epoch 6\n", + "adapter path: llama-factory/saves/Qwen2.5-0.5B-Instruct/checkpoint-210\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:40:58,327 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:40:58,327 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-0.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 896,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 4864,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 14,\n", + " \"num_hidden_layers\": 24,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:40:58,581 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:40:58,582 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:40:58,582 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:40:58,582 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:40:58,582 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:40:58,582 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 10:40:58,681 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:40:59,748 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:40:59,748 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-0.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 896,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 4864,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 14,\n", + " \"num_hidden_layers\": 24,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:41:00,011 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:41:00,011 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:41:00,011 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:41:00,011 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:41:00,011 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:41:00,011 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 10:41:00,105 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:41:00,375 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:41:00,375 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-0.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 896,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 4864,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 14,\n", + " \"num_hidden_layers\": 24,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:3634] 2024-09-22 10:41:00,385 >> loading weights file model.safetensors from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/model.safetensors\n", + "[INFO|modeling_utils.py:1572] 2024-09-22 10:41:00,388 >> Instantiating Qwen2ForCausalLM model under default dtype torch.bfloat16.\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 10:41:00,389 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:4463] 2024-09-22 10:41:23,971 >> All model checkpoint weights were used when initializing Qwen2ForCausalLM.\n", + "\n", + "[INFO|modeling_utils.py:4471] 2024-09-22 10:41:23,971 >> All the weights of Qwen2ForCausalLM were initialized from the model checkpoint at Qwen/Qwen2.5-0.5B-Instruct.\n", + "If your task is similar to the task the model of the checkpoint was trained on, you can already use Qwen2ForCausalLM for predictions without further training.\n", + "[INFO|configuration_utils.py:993] 2024-09-22 10:41:24,744 >> loading configuration file generation_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/generation_config.json\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 10:41:24,744 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"do_sample\": true,\n", + " \"eos_token_id\": [\n", + " 151645,\n", + " 151643\n", + " ],\n", + " \"pad_token_id\": 151643,\n", + " \"repetition_penalty\": 1.1,\n", + " \"temperature\": 0.7,\n", + " \"top_k\": 20,\n", + " \"top_p\": 0.8\n", + "}\n", + "\n", + "100%|███████████████████████████████████████| 1500/1500 [01:14<00:00, 20.26it/s]\n", + "Epoch 7\n", + "adapter path: llama-factory/saves/Qwen2.5-0.5B-Instruct/checkpoint-245\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:43:17,337 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:43:17,338 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-0.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 896,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 4864,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 14,\n", + " \"num_hidden_layers\": 24,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:43:17,635 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:43:17,635 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:43:17,636 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:43:17,636 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:43:17,636 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:43:17,636 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 10:43:17,731 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:43:18,764 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:43:18,765 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-0.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 896,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 4864,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 14,\n", + " \"num_hidden_layers\": 24,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:43:19,025 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:43:19,025 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:43:19,025 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:43:19,025 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:43:19,025 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:43:19,025 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 10:43:19,118 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:43:19,393 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:43:19,393 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-0.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 896,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 4864,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 14,\n", + " \"num_hidden_layers\": 24,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:3634] 2024-09-22 10:43:19,403 >> loading weights file model.safetensors from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/model.safetensors\n", + "[INFO|modeling_utils.py:1572] 2024-09-22 10:43:19,406 >> Instantiating Qwen2ForCausalLM model under default dtype torch.bfloat16.\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 10:43:19,407 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:4463] 2024-09-22 10:43:42,915 >> All model checkpoint weights were used when initializing Qwen2ForCausalLM.\n", + "\n", + "[INFO|modeling_utils.py:4471] 2024-09-22 10:43:42,915 >> All the weights of Qwen2ForCausalLM were initialized from the model checkpoint at Qwen/Qwen2.5-0.5B-Instruct.\n", + "If your task is similar to the task the model of the checkpoint was trained on, you can already use Qwen2ForCausalLM for predictions without further training.\n", + "[INFO|configuration_utils.py:993] 2024-09-22 10:43:43,171 >> loading configuration file generation_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/generation_config.json\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 10:43:43,172 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"do_sample\": true,\n", + " \"eos_token_id\": [\n", + " 151645,\n", + " 151643\n", + " ],\n", + " \"pad_token_id\": 151643,\n", + " \"repetition_penalty\": 1.1,\n", + " \"temperature\": 0.7,\n", + " \"top_k\": 20,\n", + " \"top_p\": 0.8\n", + "}\n", + "\n", + "100%|███████████████████████████████████████| 1500/1500 [01:18<00:00, 19.15it/s]\n", + "Epoch 8\n", + "adapter path: llama-factory/saves/Qwen2.5-0.5B-Instruct/checkpoint-280\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:45:12,998 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:45:12,998 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-0.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 896,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 4864,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 14,\n", + " \"num_hidden_layers\": 24,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:45:13,251 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:45:13,251 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:45:13,251 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:45:13,251 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:45:13,251 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:45:13,251 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 10:45:13,350 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:45:14,370 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:45:14,370 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-0.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 896,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 4864,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 14,\n", + " \"num_hidden_layers\": 24,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:45:14,621 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:45:14,621 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:45:14,621 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:45:14,621 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:45:14,622 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:45:14,622 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 10:45:14,726 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:45:14,984 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:45:14,985 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-0.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 896,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 4864,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 14,\n", + " \"num_hidden_layers\": 24,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:3634] 2024-09-22 10:45:14,995 >> loading weights file model.safetensors from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/model.safetensors\n", + "[INFO|modeling_utils.py:1572] 2024-09-22 10:45:14,998 >> Instantiating Qwen2ForCausalLM model under default dtype torch.bfloat16.\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 10:45:14,999 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:4463] 2024-09-22 10:45:38,509 >> All model checkpoint weights were used when initializing Qwen2ForCausalLM.\n", + "\n", + "[INFO|modeling_utils.py:4471] 2024-09-22 10:45:38,509 >> All the weights of Qwen2ForCausalLM were initialized from the model checkpoint at Qwen/Qwen2.5-0.5B-Instruct.\n", + "If your task is similar to the task the model of the checkpoint was trained on, you can already use Qwen2ForCausalLM for predictions without further training.\n", + "[INFO|configuration_utils.py:993] 2024-09-22 10:45:38,762 >> loading configuration file generation_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/generation_config.json\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 10:45:38,762 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"do_sample\": true,\n", + " \"eos_token_id\": [\n", + " 151645,\n", + " 151643\n", + " ],\n", + " \"pad_token_id\": 151643,\n", + " \"repetition_penalty\": 1.1,\n", + " \"temperature\": 0.7,\n", + " \"top_k\": 20,\n", + " \"top_p\": 0.8\n", + "}\n", + "\n", + "100%|███████████████████████████████████████| 1500/1500 [01:21<00:00, 18.33it/s]\n", + "Epoch 9\n", + "adapter path: llama-factory/saves/Qwen2.5-0.5B-Instruct/checkpoint-315\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:47:12,668 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:47:12,669 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-0.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 896,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 4864,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 14,\n", + " \"num_hidden_layers\": 24,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:47:12,926 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:47:12,926 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:47:12,926 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:47:12,926 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:47:12,926 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:47:12,926 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 10:47:13,026 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:47:14,076 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:47:14,076 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-0.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 896,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 4864,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 14,\n", + " \"num_hidden_layers\": 24,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:47:14,332 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:47:14,332 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:47:14,332 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:47:14,332 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:47:14,332 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:47:14,332 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 10:47:14,428 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:47:14,694 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:47:14,694 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-0.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 896,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 4864,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 14,\n", + " \"num_hidden_layers\": 24,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:3634] 2024-09-22 10:47:14,704 >> loading weights file model.safetensors from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/model.safetensors\n", + "[INFO|modeling_utils.py:1572] 2024-09-22 10:47:14,708 >> Instantiating Qwen2ForCausalLM model under default dtype torch.bfloat16.\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 10:47:14,708 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:4463] 2024-09-22 10:47:38,356 >> All model checkpoint weights were used when initializing Qwen2ForCausalLM.\n", + "\n", + "[INFO|modeling_utils.py:4471] 2024-09-22 10:47:38,357 >> All the weights of Qwen2ForCausalLM were initialized from the model checkpoint at Qwen/Qwen2.5-0.5B-Instruct.\n", + "If your task is similar to the task the model of the checkpoint was trained on, you can already use Qwen2ForCausalLM for predictions without further training.\n", + "[INFO|configuration_utils.py:993] 2024-09-22 10:47:38,623 >> loading configuration file generation_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/generation_config.json\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 10:47:38,623 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"do_sample\": true,\n", + " \"eos_token_id\": [\n", + " 151645,\n", + " 151643\n", + " ],\n", + " \"pad_token_id\": 151643,\n", + " \"repetition_penalty\": 1.1,\n", + " \"temperature\": 0.7,\n", + " \"top_k\": 20,\n", + " \"top_p\": 0.8\n", + "}\n", + "\n", + "100%|███████████████████████████████████████| 1500/1500 [01:17<00:00, 19.43it/s]\n", + "Epoch 10\n", + "adapter path: llama-factory/saves/Qwen2.5-0.5B-Instruct/checkpoint-350\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:49:39,361 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:49:39,362 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-0.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 896,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 4864,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 14,\n", + " \"num_hidden_layers\": 24,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:49:39,616 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:49:39,616 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:49:39,616 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:49:39,616 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:49:39,616 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:49:39,616 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 10:49:39,715 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:49:40,769 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:49:40,769 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-0.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 896,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 4864,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 14,\n", + " \"num_hidden_layers\": 24,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:49:41,037 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:49:41,037 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:49:41,037 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:49:41,037 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:49:41,037 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:49:41,037 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 10:49:41,130 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:49:41,393 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:49:41,393 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-0.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 896,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 4864,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 14,\n", + " \"num_hidden_layers\": 24,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:3634] 2024-09-22 10:49:41,403 >> loading weights file model.safetensors from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/model.safetensors\n", + "[INFO|modeling_utils.py:1572] 2024-09-22 10:49:41,407 >> Instantiating Qwen2ForCausalLM model under default dtype torch.bfloat16.\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 10:49:41,407 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:4463] 2024-09-22 10:50:05,610 >> All model checkpoint weights were used when initializing Qwen2ForCausalLM.\n", + "\n", + "[INFO|modeling_utils.py:4471] 2024-09-22 10:50:05,610 >> All the weights of Qwen2ForCausalLM were initialized from the model checkpoint at Qwen/Qwen2.5-0.5B-Instruct.\n", + "If your task is similar to the task the model of the checkpoint was trained on, you can already use Qwen2ForCausalLM for predictions without further training.\n", + "[INFO|configuration_utils.py:993] 2024-09-22 10:50:05,868 >> loading configuration file generation_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/generation_config.json\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 10:50:05,868 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"do_sample\": true,\n", + " \"eos_token_id\": [\n", + " 151645,\n", + " 151643\n", + " ],\n", + " \"pad_token_id\": 151643,\n", + " \"repetition_penalty\": 1.1,\n", + " \"temperature\": 0.7,\n", + " \"top_k\": 20,\n", + " \"top_p\": 0.8\n", + "}\n", + "\n", + "100%|███████████████████████████████████████| 1500/1500 [01:22<00:00, 18.29it/s]\n", + "Current Directory:\n", + "/home/inflaton/code/logical-reasoning\n", + "Evaluating Qwen/Qwen2.5-0.5B-Instruct with few-shot learning\n", + "loading env vars from: /home/inflaton/code/logical-reasoning/.env\n", + "Adding /home/inflaton/code/logical-reasoning to sys.path\n", + "loading /home/inflaton/code/logical-reasoning/llm_toolkit/logical_reasoning_utils.py\n", + "CUDA is available, we have found 1 GPU(s)\n", + "NVIDIA GeForce RTX 4090\n", + "CUDA version: 12.1\n", + "Qwen/Qwen2.5-0.5B-Instruct None False datasets/mgtv data/Qwen2.5-0.5B-Instruct_results.csv 2048 1\n", + "(0) GPU = NVIDIA GeForce RTX 4090. Max memory = 23.988 GB.\n", + "0.0 GB of memory reserved.\n", + "loading model: Qwen/Qwen2.5-0.5B-Instruct with adapter: None\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:51:36,635 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:51:36,636 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-0.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 896,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 4864,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 14,\n", + " \"num_hidden_layers\": 24,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:51:36,894 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:51:36,894 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:51:36,894 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:51:36,894 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:51:36,894 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:51:36,894 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 10:51:36,991 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:51:38,045 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:51:38,046 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-0.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 896,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 4864,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 14,\n", + " \"num_hidden_layers\": 24,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:51:38,329 >> loading file vocab.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/vocab.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:51:38,329 >> loading file merges.txt from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/merges.txt\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:51:38,329 >> loading file tokenizer.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/tokenizer.json\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:51:38,329 >> loading file added_tokens.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:51:38,329 >> loading file special_tokens_map.json from cache at None\n", + "[INFO|tokenization_utils_base.py:2289] 2024-09-22 10:51:38,329 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/tokenizer_config.json\n", + "[INFO|tokenization_utils_base.py:2533] 2024-09-22 10:51:38,429 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.\n", + "09/22/2024 10:51:38 - INFO - llamafactory.data.template - Replace eos token: <|im_end|>\n", + "[INFO|configuration_utils.py:733] 2024-09-22 10:51:38,707 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/config.json\n", + "[INFO|configuration_utils.py:800] 2024-09-22 10:51:38,707 >> Model config Qwen2Config {\n", + " \"_name_or_path\": \"Qwen/Qwen2.5-0.5B-Instruct\",\n", + " \"architectures\": [\n", + " \"Qwen2ForCausalLM\"\n", + " ],\n", + " \"attention_dropout\": 0.0,\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645,\n", + " \"hidden_act\": \"silu\",\n", + " \"hidden_size\": 896,\n", + " \"initializer_range\": 0.02,\n", + " \"intermediate_size\": 4864,\n", + " \"max_position_embeddings\": 32768,\n", + " \"max_window_layers\": 21,\n", + " \"model_type\": \"qwen2\",\n", + " \"num_attention_heads\": 14,\n", + " \"num_hidden_layers\": 24,\n", + " \"num_key_value_heads\": 2,\n", + " \"rms_norm_eps\": 1e-06,\n", + " \"rope_theta\": 1000000.0,\n", + " \"sliding_window\": null,\n", + " \"tie_word_embeddings\": true,\n", + " \"torch_dtype\": \"bfloat16\",\n", + " \"transformers_version\": \"4.43.3\",\n", + " \"use_cache\": true,\n", + " \"use_sliding_window\": false,\n", + " \"vocab_size\": 151936\n", + "}\n", + "\n", + "09/22/2024 10:51:38 - INFO - llamafactory.model.patcher - Using KV cache for faster generation.\n", + "[INFO|modeling_utils.py:3634] 2024-09-22 10:51:38,717 >> loading weights file model.safetensors from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/model.safetensors\n", + "[INFO|modeling_utils.py:1572] 2024-09-22 10:51:38,720 >> Instantiating Qwen2ForCausalLM model under default dtype torch.bfloat16.\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 10:51:38,721 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"eos_token_id\": 151645\n", + "}\n", + "\n", + "[INFO|modeling_utils.py:4463] 2024-09-22 10:52:02,551 >> All model checkpoint weights were used when initializing Qwen2ForCausalLM.\n", + "\n", + "[INFO|modeling_utils.py:4471] 2024-09-22 10:52:02,551 >> All the weights of Qwen2ForCausalLM were initialized from the model checkpoint at Qwen/Qwen2.5-0.5B-Instruct.\n", + "If your task is similar to the task the model of the checkpoint was trained on, you can already use Qwen2ForCausalLM for predictions without further training.\n", + "[INFO|configuration_utils.py:993] 2024-09-22 10:52:02,800 >> loading configuration file generation_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/a8b602d9dafd3a75d382e62757d83d89fca3be54/generation_config.json\n", + "[INFO|configuration_utils.py:1038] 2024-09-22 10:52:02,800 >> Generate config GenerationConfig {\n", + " \"bos_token_id\": 151643,\n", + " \"do_sample\": true,\n", + " \"eos_token_id\": [\n", + " 151645,\n", + " 151643\n", + " ],\n", + " \"pad_token_id\": 151643,\n", + " \"repetition_penalty\": 1.1,\n", + " \"temperature\": 0.7,\n", + " \"top_k\": 20,\n", + " \"top_p\": 0.8\n", + "}\n", + "\n", + "09/22/2024 10:52:02 - INFO - llamafactory.model.model_utils.attention - Using torch SDPA for faster training and inference.\n", + "09/22/2024 10:52:02 - INFO - llamafactory.model.loader - all params: 494,032,768\n", + "(2) GPU = NVIDIA GeForce RTX 4090. Max memory = 23.988 GB.\n", + "1.418 GB of memory reserved.\n", + "Evaluating model: Qwen/Qwen2.5-0.5B-Instruct on cuda\n", + "*** Evaluating with num_shots: 40\n", + "loading train/test data files\n", + "Map: 100%|███████████████████████| 25000/25000 [00:13<00:00, 1903.03 examples/s]\n", + "Map: 100%|█████████████████████████| 3000/3000 [00:01<00:00, 2256.49 examples/s]\n", + "DatasetDict({\n", + " train: Dataset({\n", + " features: ['text', 'label', 'answer', 'title', 'puzzle', 'truth', 'train_text', 'prompt'],\n", + " num_rows: 25000\n", + " })\n", + " test: Dataset({\n", + " features: ['text', 'label', 'answer', 'title', 'puzzle', 'truth', 'train_text', 'prompt'],\n", + " num_rows: 3000\n", + " })\n", + "})\n", + "--------------------------------------------------\n", + "text: 甄加索是自杀吗\n", + "--------------------------------------------------\n", + "label: 不是\n", + "--------------------------------------------------\n", + "answer: nan\n", + "--------------------------------------------------\n", + "title: 海岸之谜\n", + "--------------------------------------------------\n", + "puzzle: 在远离城市喧嚣的海边小屋,一天清晨,邻居发现甄加索僵卧在沙滩上,已无生命迹象。现场没有发现任何打斗的迹象。请问甄加索的死因是什么?\n", + "--------------------------------------------------\n", + "truth: 甄加索是一位热爱自然的画家,他每年都会来到这个海边小屋寻找灵感。在他生命的最后几天,他一直在创作一幅描绘海洋生物的画作。在画即将完成的前一��晚上,他骑着自行车外出,打算在海边观赏夜景。然而,他在沙滩上意外发现了一只搁浅的海豚,为了救助这只海豚,他耗费了极大的体力,最终成功将其送回海中。筋疲力尽的甄加索在沙滩上睡着了,由于他患有严重的心脏病,却未告知旁人,在寒冷的海风中,他的心脏停止了跳动。因此,警方在现场只发现了车轮痕迹和未完成的画作,而没有发现任何他杀的迹象。\n", + "--------------------------------------------------\n", + "train_text: <|im_start|>system\n", + "You are an expert in logical reasoning.<|im_end|>\n", + "<|im_start|>user\n", + "你是一个情景猜谜游戏的主持人。游戏规则如下:\n", + "\n", + "1. 参与者会得到一个谜面,谜面会描述一个简单又难以理解的事件。\n", + "2. 主持人知道谜底,谜底是谜面的答案。\n", + "3. 参与者可以询问任何封闭式问题来找寻事件的真相。\n", + "4. 对于每个问题,主持人将根据实际情况回答以下五个选项之一:是、不是、不重要、回答正确、问法错误。各回答的判断标准如下:\n", + " - 若谜面和谜底能找到问题的答案,回答:是或者不是\n", + " - 若谜面和谜底不能直接或者间接推断出问题的答案,回答:不重要\n", + " - 若参与者提问不是一个封闭式问题或者问题难以理解,回答:问法错误\n", + " - 若参与者提问基本还原了谜底真相,回答:回答正确\n", + "5. 回答中不能添加任何其它信息,也不能省略选项中的任何一个字。例如,不可以把“不是”省略成“不”。\n", + "\n", + "请严格按照这些规则回答参与者提出的问题。\n", + "\n", + "示例输入和输出: \n", + "谜面: 在甄家村里,有一个古老的传说:每年南瓜丰收的季节,南瓜田里总有一个最大的南瓜会不翼而飞,村民们对此现象困惑不解。请找出南瓜失踪背后的原因。\n", + "谜底: 真相原来与一位年迈的农夫有关。这位农夫年轻时,曾与一位美丽的姑娘相恋。他们约定在南瓜丰收的季节结婚。然而,命运弄人,姑娘在婚礼前的一场意外中离世。悲伤的农夫为了纪念心爱的姑娘,每年都会将最大的南瓜偷走,放到姑娘的墓前,以此寄托自己的哀思。这一行为延续了多年,成为了乡村里一个神秘的传说。\n", + "参与者提出的问题: 偷的人信神吗\n", + "回答: 不是\n", + "\n", + "谜面: 在甄家村里,有一个古老的传说:每年南瓜丰收的季节,南瓜田里总有一个最大的南瓜会不翼而飞,村民们对此现象困惑不解。请找出南瓜失踪背后的原因。\n", + "谜底: 真相原来与一位年迈的农夫有关。这位农夫年轻时,曾与一位美丽的姑娘相恋。他们约定在南瓜丰收的季节结婚。然而,命运弄人,姑娘在婚礼前的一场意外中离世。悲伤的农夫为了纪念心爱的姑娘,每年都会将最大的南瓜偷走,放到姑娘的墓前,以此寄托自己的哀思。这一行为延续了多年,成为了乡村里一个神秘的传说。\n", + "参与者提出的问题: 村庄里的人喜欢南瓜嘛\n", + "回答: 不重要\n", + "\n", + "谜面: 在甄家村里,有一个古老的传说:每年南瓜丰收的季节,南瓜田里总有一个最大的南瓜会不翼而飞,村民们对此现象困惑不解。请找出南瓜失踪背后的原因。\n", + "谜底: 真相原来与一位年迈的农夫有关。这位农夫年轻时,曾与一位美丽的姑娘相恋。他们约定在南瓜丰收的季节结婚。然而,命运弄人,姑娘在婚礼前的一场意外中离世。悲伤的农夫为了纪念心爱的姑娘,每年都会将最大的南瓜偷走,放到姑娘的墓前,以此寄托自己的哀思。这一行为延续了多年,成为了乡村里一个神秘的传说。\n", + "参与者提出的问题: 是村里的人偷的么\n", + "回答: 是\n", + "\n", + "谜面: 在一个炎热的夏日,乡村的甄家大院的西瓜突然全部不翼而飞。据了解,甄家大院周围并没有其他人家,而且门窗都完好无损,没有任何被撬的痕迹。村民们议论纷纷���猜测这批西瓜究竟去了哪里。你知道西瓜去了哪里吗?\n", + "谜底: 原来,这批西瓜是被一只巨大的乌鸦偷走了。这只乌鸦为了给自己的孩子们准备食物,它趁着夜色,竟然将甄家大院的西瓜一颗颗地带回了巢穴。第二天,村民们发现了乌鸦的巢穴,里面堆满了西瓜,而这个意外的真相让所有人都忍俊不禁。甄家老爷也感慨地说:“真是世界大了,什么奇事都有!”\n", + "参与者提出的问题: 挖地道\n", + "回答: 问法错误\n", + "\n", + "谜面: 在一个炎热的夏日,乡村的甄家大院的西瓜突然全部不翼而飞。据了解,甄家大院周围并没有其他人家,而且门窗都完好无损,没有任何被撬的痕迹。村民们议论纷纷,猜测这批西瓜究竟去了哪里。你知道西瓜去了哪里吗?\n", + "谜底: 原来,这批西瓜是被一只巨大的乌鸦偷走了。这只乌鸦为了给自己的孩子们准备食物,它趁着夜色,竟然将甄家大院的西瓜一颗颗地带回了巢穴。第二天,村民们发现了乌鸦的巢穴,里面堆满了西瓜,而这个意外的真相让所有人都忍俊不禁。甄家老爷也感慨地说:“真是世界大了,什么奇事都有!”\n", + "参与者提出的问题: 鸟觅食时发现甄家大院有西瓜,飞入大院一颗一颗把西瓜带走\n", + "回答: 回答正确\n", + "\n", + "谜面: 在一个安静的夜晚,小镇上的钟楼突然停止了报时。第二天早晨,人们发现钟楼的管理员甄大勇失踪了,而钟楼的门紧闭,从外面看起来一切正常。小镇上的人们议论纷纷,不知道发生了什么事情。\n", + "谜底: 真相是,钟楼的管理员甄大勇在夜晚进行例行的钟楼维护时,不慎从钟楼的顶部摔落,但并未死亡,只是昏迷。由于他跌落时砸到了控制时钟报时的机械装置,导致钟声停止。他躺在钟楼底部,但由于门从内部反锁,外面的人无法进入。甄大勇在第二天中午苏醒后,自己打开了门,这才知道自己引发了小镇上的恐慌。\n", + "参与者提出的问题: 有人身亡吗?\n", + "回答: 不是\n", + "\n", + "谜面: 在一个安静的夜晚,小镇上的钟楼突然停止了报时。第二天早晨,人们发现钟楼的管理员甄大勇失踪了,而钟楼的门紧闭,从外面看起来一切正常。小镇上的人们议论纷纷,不知道发生了什么事情。\n", + "谜底: 真相是,钟楼的管理员甄大勇在夜晚进行例行的钟楼维护时,不慎从钟楼的顶部摔落,但并未死亡,只是昏迷。由于他跌落时砸到了控制时钟报时的机械装置,导致钟声停止。他躺在钟楼底部,但由于门从内部反锁,外面的人无法进入。甄大勇在第二天中午苏醒后,自己打开了门,这才知道自己引发了小镇上的恐慌。\n", + "参与者提出的问题: 有人跟甄大勇有仇吗\n", + "回答: 不重要\n", + "\n", + "谜面: 在一个安静的夜晚,小镇上的钟楼突然停止了报时。第二天早晨,人们发现钟楼的管理员甄大勇失踪了,而钟楼的门紧闭,从外面看起来一切正常。小镇上的人们议论纷纷,不知道发生了什么事情。\n", + "谜底: 真相是,钟楼的管理员甄大勇在夜晚进行例行的钟楼维护时,不慎从钟楼的顶部摔落,但并未死亡,只是昏迷。由于他跌落时砸到了控制时钟报时的机械装置,导致钟声停止。他躺在钟楼底部,但由于门从内部反锁,外面的人无法进入。甄大勇在第二天中午苏醒后,自己打开了门,这才知道自己引发了小镇上的恐慌。\n", + "参与者提出的问题: 他仅仅是在修钟楼吗\n", + "回答: 是\n", + "\n", + "谜面: 在一个安静的夜晚,小镇上的钟楼突然停止了报时。第二天早晨,人们发现钟楼的管理员甄大勇失踪了,而钟楼的门紧闭,从外面看起来一切正常。小镇上的人们议论纷纷,不知道发生了什么事情。\n", + "谜底: 真相是,钟楼的管理员甄大勇在夜晚进行例行的钟楼维护时,不慎从钟楼的顶部摔落,但并未死亡,只是昏迷。由于他跌落时砸到了控制时钟报时的机械装置,导致钟声停止。他躺在钟楼底部,但由于门从内部反锁,外面的人无法进入。甄大勇在第二天中午苏醒后,自己打开了门,这才知道自���引发了小镇上的恐慌。\n", + "参与者提出的问题: 是自然意外还是人为意外\n", + "回答: 问法错误\n", + "\n", + "谜面: 在一个安静的夜晚,小镇上的钟楼突然停止了报时。第二天早晨,人们发现钟楼的管理员甄大勇失踪了,而钟楼的门紧闭,从外面看起来一切正常。小镇上的人们议论纷纷,不知道发生了什么事情。\n", + "谜底: 真相是,钟楼的管理员甄大勇在夜晚进行例行的钟楼维护时,不慎从钟楼的顶部摔落,但并未死亡,只是昏迷。由于他跌落时砸到了控制时钟报时的机械装置,导致钟声停止。他躺在钟楼底部,但由于门从内部反锁,外面的人无法进入。甄大勇在第二天中午苏醒后,自己打开了门,这才知道自己引发了小镇上的恐慌。\n", + "参与者提出的问题: 因为甄在钟楼里维修然后昏迷了导致钟楼停止报时\n", + "回答: 回答正确\n", + "\n", + "谜面: 在一个安静的夜晚,小镇上的钟楼突然停止了报时。第二天早晨,人们发现钟楼的管理员甄大勇失踪了,而钟楼的门紧闭,从外面看起来一切正常。小镇上的人们议论纷纷,不知道发生了什么事情。\n", + "谜底: 真相是,钟楼的管理员甄大勇在夜晚进行例行的钟楼维护时,不慎从钟楼的顶部摔落,但并未死亡,只是昏迷。由于他跌落时砸到了控制时钟报时的机械装置,导致钟声停止。他躺在钟楼底部,但由于门从内部反锁,外面的人无法进入。甄大勇在第二天中午苏醒后,自己打开了门,这才知道自己引发了小镇上的恐慌。\n", + "参与者提出的问题: 是停止报时才导致甄大勇失踪吗?\n", + "回答: 不是\n", + "\n", + "谜面: 在一个安静的夜晚,小镇上的钟楼突然停止了报时。第二天早晨,人们发现钟楼的管理员甄大勇失踪了,而钟楼的门紧闭,从外面看起来一切正常。小镇上的人们议论纷纷,不知道发生了什么事情。\n", + "谜底: 真相是,钟楼的管理员甄大勇在夜晚进行例行的钟楼维护时,不慎从钟楼的顶部摔落,但并未死亡,只是昏迷。由于他跌落时砸到了控制时钟报时的机械装置,导致钟声停止。他躺在钟楼底部,但由于门从内部反锁,外面的人无法进入。甄大勇在第二天中午苏醒后,自己打开了门,这才知道自己引发了小镇上的恐慌。\n", + "参与者提出的问题: 甄大勇是和他人一起离开吗?\n", + "回答: 不重要\n", + "\n", + "谜面: 在一个安静的夜晚,小镇上的钟楼突然停止了报时。第二天早晨,人们发现钟楼的管理员甄大勇失踪了,而钟楼的门紧闭,从外面看起来一切正常。小镇上的人们议论纷纷,不知道发生了什么事情。\n", + "谜底: 真相是,钟楼的管理员甄大勇在夜晚进行例行的钟楼维护时,不慎从钟楼的顶部摔落,但并未死亡,只是昏迷。由于他跌落时砸到了控制时钟报时的机械装置,导致钟声停止。他躺在钟楼底部,但由于门从内部反锁,外面的人无法进入。甄大勇在第二天中午苏醒后,自己打开了门,这才知道自己引发了小镇上的恐慌。\n", + "参与者提出的问题: 是甄大勇弄坏了钟楼的报时器吗?\n", + "回答: 是\n", + "\n", + "谜面: 在一个安静的夜晚,小镇上的钟楼突然停止了报时。第二天早晨,人们发现钟楼的管理员甄大勇失踪了,而钟楼的门紧闭,从外面看起来一切正常。小镇上的人们议论纷纷,不知道发生了什么事情。\n", + "谜底: 真相是,钟楼的管理员甄大勇在夜晚进行例行的钟楼维护时,不慎从钟楼的顶部摔落,但并未死亡,只是昏迷。由于他跌落时砸到了控制时钟报时的机械装置,导致钟声停止。他躺在钟楼底部,但由于门从内部反锁,外面的人无法进入。甄大勇在第二天中午苏醒后,自己打开了门,这才知道自己引发了小镇上的恐慌。\n", + "参与者提出的问题: 别被人当时在此地吗?。\n", + "回答: 问法错误\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 老人认为的椅子其实一个梯子,被人拿走去维修东西了\n", + "回答: 回答正确\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 长椅坏了吗\n", + "回答: 不是\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 被人偷走了\n", + "回答: 不重要\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 他那张椅子是公园的吗\n", + "回答: 是\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 椅子是其他人受过伤吗\n", + "回答: 问法错误\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 维修工人在公园维修物品,用老人这个椅子来垫脚\n", + "回答: 回答正确\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 把椅子搬走是为了老头吗\n", + "回答: 不是\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 老头的椅子有不一样的地方\n", + "回答: 不重要\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 公园的灯坏了吗\n", + "回答: 是\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 原本有这个意思吗\n", + "回答: 问法错误\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 工人为了修东西搬来了椅子,老人认为椅子是固定的,一直用椅子,工人修好东西后把椅子搬走了\n", + "回答: 回答正确\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 是他的子女搬走他的椅子吗\n", + "回答: 不是\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 老人有疾病吗\n", + "回答: 不重要\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 他的椅子是正常椅子吗\n", + "回答: 是\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 这是一个悲伤的故事\n", + "回答: 问法错误\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 公园的这把椅子实际上是个梯子,公园里有东西坏了,所以梯子被维修人员拿去用了\n", + "回答: 回答正确\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 是重新刷漆了吗\n", + "回答: 不是\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 椅子很重要\n", + "回答: 不重要\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 椅子被挪走了\n", + "回答: 是\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 一块木头\n", + "回答: 问法错误\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 老人的椅子是个梯子,有人要维修东西将椅子搬走了\n", + "回答: 回答正确\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 椅子是被人偷走了吗\n", + "回答: 不是\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 椅子被盖住了吗\n", + "回答: 不重要\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 椅子在公园的其他地方\n", + "回答: 是\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 能给我点提示吗\n", + "回答: 问法错误\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 真正的椅子在公园那个椅子不是椅子被拿去维修了\n", + "回答: 回答正确\n", + "\n", + "\n", + "谜面: 在远离城市喧嚣的海边小屋,一天清晨,邻居发现甄加索僵卧在沙滩���,已无生命迹象。现场没有发现任何打斗的迹象。请问甄加索的死因是什么?\n", + "谜底: 甄加索是一位热爱自然的画家,他每年都会来到这个海边小屋寻找灵感。在他生命的最后几天,他一直在创作一幅描绘海洋生物的画作。在画即将完成的前一天晚上,他骑着自行车外出,打算在海边观赏夜景。然而,他在沙滩上意外发现了一只搁浅的海豚,为了救助这只海豚,他耗费了极大的体力,最终成功将其送回海中。筋疲力尽的甄加索在沙滩上睡着了,由于他患有严重的心脏病,却未告知旁人,在寒冷的海风中,他的心脏停止了跳动。因此,警方在现场只发现了车轮痕迹和未完成的画作,而没有发现任何他杀的迹象。\n", + "参与者提出的问题: 甄加索是自杀吗\n", + "回答: \n", + "<|im_end|>\n", + "<|im_start|>assistant\n", + "不是<|im_end|>\n", + "--------------------------------------------------\n", + "prompt: <|im_start|>system\n", + "You are an expert in logical reasoning.<|im_end|>\n", + "<|im_start|>user\n", + "你是一个情景猜谜游戏的主持人。游戏规则如下:\n", + "\n", + "1. 参与者会得到一个谜面,谜面会描述一个简单又难以理解的事件。\n", + "2. 主持人知道谜底,谜底是谜面的答案。\n", + "3. 参与者可以询问任何封闭式问题来找寻事件的真相。\n", + "4. 对于每个问题,主持人将根据实际情况回答以下五个选项之一:是、不是、不重要、回答正确、问法错误。各回答的判断标准如下:\n", + " - 若谜面和谜底能找到问题的答案,回答:是或者不是\n", + " - 若谜面和谜底不能直接或者间接推断出问题的答案,回答:不重要\n", + " - 若参与者提问不是一个封闭式问题或者问题难以理解,回答:问法错误\n", + " - 若参与者提问基本还原了谜底真相,回答:回答正确\n", + "5. 回答中不能添加任何其它信息,也不能省略选项中的任何一个字。例如,不可以把“不是”省略成“不”。\n", + "\n", + "请严格按照这些规则回答参与者提出的问题。\n", + "\n", + "示例输入和输出: \n", + "谜面: 在甄家村里,有一个古老的传说:每年南瓜丰收的季节,南瓜田里总有一个最大的南瓜会不翼而飞,村民们对此现象困惑不解。请找出南瓜失踪背后的原因。\n", + "谜底: 真相原来与一位年迈的农夫有关。这位农夫年轻时,曾与一位美丽的姑娘相恋。他们约定在南瓜丰收的季节结婚。然而,命运弄人,姑娘在婚礼前的一场意外中离世。悲伤的农夫为了纪念心爱的姑娘,每年都会将最大的南瓜偷走,放到姑娘的墓前,以此寄托自己的哀思。这一行为延续了多年,成为了乡村里一个神秘的传说。\n", + "参与者提出的问题: 偷的人信神吗\n", + "回答: 不是\n", + "\n", + "谜面: 在甄家村里,有一个古老的传说:每年南瓜丰收的季节,南瓜田里总有一个最大的南瓜会不翼而飞,村民们对此现象困惑不解。请找出南瓜失踪背后的原因。\n", + "谜底: 真相原来与一位年迈的农夫有关。这位农夫年轻时,曾与一位美丽的姑娘相恋。他们约定在南瓜丰收的季节结婚。然而,命运弄人,姑娘在婚礼前的一场意外中离世。悲伤的农夫为了纪念心爱的姑娘,每年都会将最大的南瓜偷走,放到姑娘的墓前,以此寄托自己的哀思。这一行为延续了多年,成为了乡村里一个神秘的传说。\n", + "参与者提出的问题: 村庄里的人喜欢南瓜嘛\n", + "回答: 不重要\n", + "\n", + "谜面: 在甄家村里,有一个古老的传说:每年南瓜丰收的季节,南瓜田里总有一个最大的南瓜会不翼而飞,村民们对此现象困惑不解。请找出南瓜失踪背后的原因。\n", + "谜底: 真相原来与一位年迈的农夫有关。这位农夫年轻时,曾与一位美丽的姑娘相恋。他们约定在南瓜丰收的季节结婚。然而,命运弄人,姑娘在婚礼前的一场意外中离世。悲伤的农夫为了纪念心爱的姑娘,每年都会将最大的南瓜偷走,放到姑娘的墓前,以此寄托自己的哀思。这一行为延续了多年,成为了乡村里一个神秘的传说。\n", + "参与者提出的问题: 是村里的人偷的么\n", + "回答: 是\n", + "\n", + "谜面: 在一个炎热的夏日,乡村的甄家大院的西瓜突然全部不翼而飞。据了解,甄家大院周围并没有其他人家,而且门窗都完好无损,没有任何被撬的痕迹。村民们议论纷纷,猜测这批西瓜究竟去了哪里。你知道西瓜去了哪里吗?\n", + "谜底: 原来,这批西瓜是被一只巨大的乌鸦偷走了。这只乌鸦为了给自己的孩子们准备食物,它趁着夜色,竟然将甄家大院的西瓜一颗颗地带回了巢穴。第二天,村民们发现了乌鸦的巢穴,里面堆满了西瓜,而这个意外的真相让所有人都忍俊不禁。甄家老爷也感慨地说:“真是世界大了,什么奇事都有!”\n", + "参与者提出的问题: 挖地道\n", + "回答: 问法错误\n", + "\n", + "谜面: 在一个炎热的夏日,乡村的甄家大院的西瓜突然全部不翼而飞。据了解,甄家大院周围并没有其他人家,而且门窗都完好无损,没有任何被撬的痕迹。村民们议论纷纷,猜测这批西瓜究竟去了哪里。你知道西瓜去了哪里吗?\n", + "谜底: 原来,这批西瓜是被一只巨大的乌鸦偷走了。这只乌鸦为了给自己的孩子们准备食物,它趁着夜色,竟然将甄家大院的西瓜一颗颗地带回了巢穴。第二天,村民们发现了乌鸦的巢穴,里面堆满了西瓜,而这个意外的真相让所有人都忍俊不禁。甄家老爷也感慨地说:“真是世界大了,什么奇事都有!”\n", + "参与者提出的问题: 鸟觅食时发现甄家大院有西瓜,飞入大院一颗一颗把西瓜带走\n", + "回答: 回答正确\n", + "\n", + "谜面: 在一个安静的夜晚,小镇上的钟楼突然停止了报时。第二天早晨,人们发现钟楼的管理员甄大勇失踪了,而钟楼的门紧闭,从外面看起来一切正常。小镇上的人们议论纷纷,不知道发生了什么事情。\n", + "谜底: 真相是,钟楼的管理员甄大勇在夜晚进行例行的钟楼维护时,不慎从钟楼的顶部摔落,但并未死亡,只是昏迷。由于他跌落时砸到了控制时钟报时的机械装置,导致钟声停止。他躺在钟楼底部,但由于门从内部反锁,外面的人无法进入。甄大勇在第二天中午苏醒后,自己打开了门,这才知道自己引发了小镇上的恐慌。\n", + "参与者提出的问题: 有人身亡吗?\n", + "回答: 不是\n", + "\n", + "谜面: 在一个安静的夜晚,小镇上的钟楼突然停止了报时。第二天早晨,人们发现钟楼的管理员甄大勇失踪了,而钟楼的门紧闭,从外面看起来一切正常。小镇上的人们议论纷纷,不知道发生了什么事情。\n", + "谜底: 真相是,钟楼的管理员甄大勇在夜晚进行例行的钟楼维护时,不慎从钟楼的顶部摔落,但并未死亡,只是昏迷。由于他跌落时砸到了控制时钟报时的机械装置,导致钟声停止。他躺在钟楼底部,但由于门从内部反锁,外面的人无法进入。甄大勇在第二天中午苏醒后,自己打开了门,这才知道自己引发了小镇上的恐慌。\n", + "参与者提出的问题: 有人跟甄大勇有仇吗\n", + "回答: 不重要\n", + "\n", + "谜面: 在一个安静的夜晚,小镇上的钟楼突然停止了报时。第二天早晨,人们发现钟楼的管理员甄大勇失踪了,而钟楼的门紧闭,从外面看起来一切正常。小镇上的人们议论纷纷,不知道发生了什么事情。\n", + "谜底: 真相是,钟楼的管理员甄大勇在夜晚进行例行的钟楼维护时,不慎从钟楼的顶部摔落,但并未死亡,只是昏迷。由于他跌落时砸到了控制时钟报时的机械装置,导致钟声停止。他躺在钟楼底部,但由于门从内部反锁,外面的人无法进入。甄大勇在第二天中午苏醒后,自己打开了门,这才知道自己引发了小镇上的恐慌。\n", + "参与者提出的问题: 他仅仅是在修钟楼吗\n", + "回答: 是\n", + "\n", + "谜面: 在一个安静的夜晚,小镇上的钟楼突然停止了报时。第二天早��,人们发现钟楼的管理员甄大勇失踪了,而钟楼的门紧闭,从外面看起来一切正常。小镇上的人们议论纷纷,不知道发生了什么事情。\n", + "谜底: 真相是,钟楼的管理员甄大勇在夜晚进行例行的钟楼维护时,不慎从钟楼的顶部摔落,但并未死亡,只是昏迷。由于他跌落时砸到了控制时钟报时的机械装置,导致钟声停止。他躺在钟楼底部,但由于门从内部反锁,外面的人无法进入。甄大勇在第二天中午苏醒后,自己打开了门,这才知道自己引发了小镇上的恐慌。\n", + "参与者提出的问题: 是自然意外还是人为意外\n", + "回答: 问法错误\n", + "\n", + "谜面: 在一个安静的夜晚,小镇上的钟楼突然停止了报时。第二天早晨,人们发现钟楼的管理员甄大勇失踪了,而钟楼的门紧闭,从外面看起来一切正常。小镇上的人们议论纷纷,不知道发生了什么事情。\n", + "谜底: 真相是,钟楼的管理员甄大勇在夜晚进行例行的钟楼维护时,不慎从钟楼的顶部摔落,但并未死亡,只是昏迷。由于他跌落时砸到了控制时钟报时的机械装置,导致钟声停止。他躺在钟楼底部,但由于门从内部反锁,外面的人无法进入。甄大勇在第二天中午苏醒后,自己打开了门,这才知道自己引发了小镇上的恐慌。\n", + "参与者提出的问题: 因为甄在钟楼里维修然后昏迷了导致钟楼停止报时\n", + "回答: 回答正确\n", + "\n", + "谜面: 在一个安静的夜晚,小镇上的钟楼突然停止了报时。第二天早晨,人们发现钟楼的管理员甄大勇失踪了,而钟楼的门紧闭,从外面看起来一切正常。小镇上的人们议论纷纷,不知道发生了什么事情。\n", + "谜底: 真相是,钟楼的管理员甄大勇在夜晚进行例行的钟楼维护时,不慎从钟楼的顶部摔落,但并未死亡,只是昏迷。由于他跌落时砸到了控制时钟报时的机械装置,导致钟声停止。他躺在钟楼底部,但由于门从内部反锁,外面的人无法进入。甄大勇在第二天中午苏醒后,自己打开了门,这才知道自己引发了小镇上的恐慌。\n", + "参与者提出的问题: 是停止报时才导致甄大勇失踪吗?\n", + "回���: 不是\n", + "\n", + "谜面: 在一个安静的夜晚,小镇上的钟楼突然停止了报时。第二天早晨,人们发现钟楼的管理员甄大勇失踪了,而钟楼的门紧闭,从外面看起来一切正常。小镇上的人们议论纷纷,不知道发生了什么事情。\n", + "谜底: 真相是,钟楼的管理员甄大勇在夜晚进行例行的钟楼维护时,不慎从钟楼的顶部摔落,但并未死亡,只是昏迷。由于他跌落时砸到了控制时钟报时的机械装置,导致钟声停止。他躺在钟楼底部,但由于门从内部反锁,外面的人无法进入。甄大勇在第二天中午苏醒后,自己打开了门,这才知道自己引发了小镇上的恐慌。\n", + "参与者提出的问题: 甄大勇是和他人一起离开吗?\n", + "回答: 不重要\n", + "\n", + "谜面: 在一个安静的夜晚,小镇上的钟楼突然停止了报时。第二天早晨,人们发现钟楼的管理员甄大勇失踪了,而钟楼的门紧闭,从外面看起来一切正常。小镇上的人们议论纷纷,不知道发生了什么事情。\n", + "谜底: 真相是,钟楼的管理员甄大勇在夜晚进行例行的钟楼维护时,不慎从钟楼的顶部摔落,但并未死亡,只是昏迷。由于他跌落时砸到了控制时钟报时的机械装置,导致钟声停止。他躺在钟楼底部,但由于门从内部反锁,外面的人无法进入。甄大勇在第二天中午苏醒后,自己打开了门,这才知道自己引发了小镇上的恐慌。\n", + "参与者提出的问题: 是甄大勇弄坏了钟楼的报时器吗?\n", + "回答: 是\n", + "\n", + "谜面: 在一个安静的夜晚,小镇上的钟楼突然停止了报时。第二天早晨,人们发现钟楼的管理员甄大勇失踪了,而钟楼的门紧闭,从外面看起来一切正常。小镇上的人们议论纷纷,不知道发生了什么事情。\n", + "谜底: 真相是,钟楼的管理员甄大勇在夜晚进行例行的钟楼维护时,不慎从钟楼的顶部摔落,但并未死亡,只是昏迷。由于他跌落时砸到了控制时钟报时的��械装置,导致钟声停止。他躺在钟楼底部,但由于门从内部反锁,外面的人无法进入。甄大勇在第二天中午苏醒后,自己打开了门,这才知道自己引发了小镇上的恐慌。\n", + "参与者提出的问题: 别被人当时在此地吗?。\n", + "回答: 问法错误\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 老人认为的椅子其实一个梯子,被人拿走去维修东西了\n", + "回答: 回答正确\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 长椅坏了吗\n", + "回答: 不是\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 被人偷走了\n", + "回答: 不重要\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 他那张椅子是公园的吗\n", + "回答: 是\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时��子。”\n", + "参与者提出的问题: 椅子是其他人受过伤吗\n", + "回答: 问法错误\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 维修工人在公园维修物品,用老人这个椅子来垫脚\n", + "回答: 回答正确\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 把椅子搬走是为了老头吗\n", + "回答: 不是\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 老头的椅子有不一样的地方\n", + "回答: 不重要\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 公园的灯坏了吗\n", + "回答: 是\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 原本有这个意思吗\n", + "回答: 问法错误\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 工人为了修东西搬来了椅子,老人认为椅子是固定的,一直用椅子,工人修好东西后把椅子搬走了\n", + "回答: 回答正确\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 是他的子女搬走他的椅子吗\n", + "回答: 不是\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 老人有疾病吗\n", + "回答: 不重要\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 他的椅子是正常椅子吗\n", + "回答: 是\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 这是一个悲伤的故事\n", + "回答: 问法错误\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个���园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 公园的这把椅子实际上是个梯子,公园里有东西坏了,所以梯子被维修人员拿去用了\n", + "回答: 回答正确\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 是重新刷漆了吗\n", + "回答: 不是\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 椅子很重要\n", + "回答: 不重要\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 椅子被挪走了\n", + "回答: 是\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 一块木头\n", + "回答: 问法错误\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子���见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 老人的椅子是个梯子,有人要维修东西将椅子搬走了\n", + "回答: 回答正确\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 椅子是被人偷走了吗\n", + "回答: 不是\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 椅子被盖住了吗\n", + "回答: 不重要\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 椅子在公园的其他地方\n", + "回答: 是\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 能给我点提示吗\n", + "回答: 问法错误\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬���去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 真正的椅子在公园那个椅子不是椅子被拿去维修了\n", + "回答: 回答正确\n", + "\n", + "\n", + "谜面: 在远离城市喧嚣的海边小屋,一天清晨,邻居发现甄加索僵卧在沙滩上,已无生命迹象。现场没有发现任何打斗的迹象。请问甄加索的死因是什么?\n", + "谜底: 甄加索是一位热爱自然的画家,他每年都会来到这个海边小屋寻找灵感。在他生命的最后几天,他一直在创作一幅描绘海洋生物的画作。在画即将完成的前一天晚上,他骑着自行车外出,打算在海边观赏夜景。然而,他在沙滩上意外发现了一只搁浅的海豚,为了救助这只海豚,他耗费了极大的体力,最终成功将其送回海中。筋疲力尽的甄加索在沙滩上睡着了,由于他患有严重的心脏病,却未告知旁人,在寒冷的海风中,他的心脏停止了跳动。因此,警方在现场只发现了车轮痕迹和未完成的画作,而没有发现任何他杀的迹象。\n", + "参与者提出的问题: 甄加索是自杀吗\n", + "回答: \n", + "<|im_end|>\n", + "<|im_start|>assistant\n", + "\n", + "--------------------------------------------------\n", + "text: 死者受伤了吗\n", + "--------------------------------------------------\n", + "label: 不是\n", + "--------------------------------------------------\n", + "answer: nan\n", + "--------------------------------------------------\n", + "title: 甄庄哭声\n", + "--------------------------------------------------\n", + "puzzle: 在一个安静的夜晚,小村庄的湖边突然传来了阵阵哭泣声。第二天早晨,村长甄锐发现湖边的石头上放着一顶破旧的帽子,但没有人知道这顶帽子是从哪里来的,哭泣声又是为何。请还原故事真相。\n", + "--------------------------------------------------\n", + "truth: 原来,这顶破旧的帽子属于一个小男孩,他小时候与爷爷在湖边生活。爷爷教他钓鱼、游泳,还告诉他湖中的海龟是他们的朋友。后来,小男孩随父母去了城市生活,但每年夏天都会回到村子探望爷爷。然而,去年夏天,爷爷因病去世,小男孩伤心欲绝。今年夏天,他回到村子,来到湖边,想起和爷爷的美好回忆,忍不住哭泣。他将爷爷的帽子放在湖边的石头上,希望能让爷爷的在天之灵得到安慰。那晚的哭泣声正是小男孩在祭莫他亲爱的爷爷。\n", + "--------------------------------------------------\n", + "train_text: <|im_start|>system\n", + "You are an expert in logical reasoning.<|im_end|>\n", + "<|im_start|>user\n", + "你是一个情景猜谜游戏的主持人。游戏规则如下:\n", + "\n", + "1. 参与者会得到一个谜面,谜面会描述一个简单又难以理解的事件。\n", + "2. 主持人知道谜底,谜底是谜面的答案。\n", + "3. 参与者可以询问任何封闭式问题来找寻事件的真相。\n", + "4. 对于每个问题,主持人将根据实际情况回答以下五个选项之一:是、不是、不重要、回答正确、问法错误。各回答的判断标准如下:\n", + " - 若谜面和谜底能找到问题的答案,回答:是或者不是\n", + " - 若谜面和谜底不能直接或者间接推断出问题的答案,回答:不重要\n", + " - 若参与者提问不是一个封闭式问题或者问题难以理解,回答:问法错误\n", + " - 若参与者提问基本还原了谜底真相,回答:回答正确\n", + "5. 回答中不能添加任何其它信息,也不能省略选项中的任何一个字。例如,不可以把“不是”省略成“不”。\n", + "\n", + "请严格按照这些规则回答参与者提出的问题。\n", + "\n", + "示例输入和输出: \n", + "谜面: 在甄家村里,��一个古老的传说:每年南瓜丰收的季节,南瓜田里总有一个最大的南瓜会不翼而飞,村民们对此现象困惑不解。请找出南瓜失踪背后的原因。\n", + "谜底: 真相原来与一位年迈的农夫有关。这位农夫年轻时,曾与一位美丽的姑娘相恋。他们约定在南瓜丰收的季节结婚。然而,命运弄人,姑娘在婚礼前的一场意外中离世。悲伤的农夫为了纪念心爱的姑娘,每年都会将最大的南瓜偷走,放到姑娘的墓前,以此寄托自己的哀思。这一行为延续了多年,成为了乡村里一个神秘的传说。\n", + "参与者提出的问题: 偷的人信神吗\n", + "回答: 不是\n", + "\n", + "谜面: 在甄家村里,有一个古老的传说:每年南瓜丰收的季节,南瓜田里总有一个最大的南瓜会不翼而飞,村民们对此现象困惑不解。请找出南瓜失踪背后的原因。\n", + "谜底: 真相原来与一位年迈的农夫有关。这位农夫年轻时,曾与一位美丽的姑娘相恋。他们约定在南瓜丰收的季节结婚。然而,命运弄人,姑娘在婚礼前的一场意外中离世。悲伤的农夫为了纪念心爱的姑娘,每年都会将最大的南瓜偷走,放到姑娘的墓前,以此寄托自己的哀思。这一行为延续了多年,成为了乡村里一个神秘的传说。\n", + "参与者提出的问题: 村庄里的人喜欢南瓜嘛\n", + "回答: 不重要\n", + "\n", + "谜面: 在甄家村里,有一个古老的传说:每年南瓜丰收的季节,南瓜田里总有一个最大的南瓜会不翼而飞,村民们对此现象困惑不解。请找出南瓜失踪背后的原因。\n", + "谜底: 真相原来与一位年迈的农夫有关。这位农夫年轻时,曾与一位美丽的姑娘相恋。他们约定在南瓜丰收的季节结婚。然而,命运弄人,姑娘在婚礼前的一场意外中离世。悲伤的农夫为了纪念心爱的姑娘,每年都会将最大的南瓜偷走,放到姑娘的墓前,以此寄托自己的哀思。这一行为延续了多年,成为了乡村里一个神秘的传说。\n", + "参与者提出的问题: 是村里的人偷的么\n", + "回答: 是\n", + "\n", + "谜面: 在一个炎热的夏日,乡村的甄家大院的西瓜突然全部不翼而飞。据了解,甄家大院周围并没有其他人家,而且门窗都完好无损,没有任何被撬的痕迹。村民们议论纷纷,猜测这批西瓜究竟去了哪里。你知道西瓜去了哪里吗?\n", + "谜底: 原来,这批西瓜是被一只巨大的乌鸦偷走了。这只乌鸦为了给自己的孩子们准备食物,它趁着夜色,竟然将甄家大院的西瓜一颗颗地带回了巢穴。第二天,村民们发现了乌鸦的巢穴,里面堆满了西瓜,而这个意外的真相让所有人都忍俊不禁。甄家老爷也感慨地说:“真是世界大了,什么奇事都有!”\n", + "参与者提出的问题: 挖地道\n", + "回答: 问法错误\n", + "\n", + "谜面: 在一个炎热的夏日,乡村的甄家大院的西瓜突然全部不翼而飞。据了解,甄家大院周围并没有其他人家,而且门窗都完好无损,没有任何被撬的痕迹。村民们议论纷纷,猜测这批西瓜究竟去了哪里。你知道西瓜去了哪里吗?\n", + "谜底: 原来,这批西瓜是被一只巨大的乌鸦偷走了。这只乌鸦为了给自己的孩子们准备食物,它趁着夜色,竟然将甄家大院的西瓜一颗颗地带回了巢穴。第二天,村民们发现了乌鸦的巢穴,里面堆满了西瓜,而这个意外的真相让所有人都忍俊不禁。甄家老爷也感慨地说:“真是世界大了,什么奇事都有!”\n", + "参与者提出的问题: 鸟觅食时发现甄家大院有西瓜,飞入大院一颗一颗把西瓜带走\n", + "回答: 回答正确\n", + "\n", + "谜面: 在一个安静的夜晚,小镇上的钟楼突然停止了报时。第二天早晨,人们发现钟楼的管理员甄大勇失踪了,而钟楼的门紧闭,从外面看起来一切正常。小镇上的人们议论纷纷,不知道发生了什么事情。\n", + "谜底: 真相是,钟楼的管理员甄大勇在夜晚进行例行的钟楼维护时,不慎从钟楼的顶部摔落,但并未死亡,只是昏迷。由于他跌落时砸到了控制时钟报时的机械装置,导致钟声停止。他躺在钟楼底部,但由于门从内部反锁,外面的人无法进入。甄大勇��第二天中午苏醒后,自己打开了门,这才知道自己引发了小镇上的恐慌。\n", + "参与者提出的问题: 有人身亡吗?\n", + "回答: 不是\n", + "\n", + "谜面: 在一个安静的夜晚,小镇上的钟楼突然停止了报时。第二天早晨,人们发现钟楼的管理员甄大勇失踪了,而钟楼的门紧闭,从外面看起来一切正常。小镇上的人们议论纷纷,不知道发生了什么事情。\n", + "谜底: 真相是,钟楼的管理员甄大勇在夜晚进行例行的钟楼维护时,不慎从钟楼的顶部摔落,但并未死亡,只是昏迷。由于他跌落时砸到了控制时钟报时的机械装置,导致钟声停止。他躺在钟楼底部,但由于门从内部反锁,外面的人无法进入。甄大勇在第二天中午苏醒后,自己打开了门,这才知道自己引发了小镇上的恐慌。\n", + "参与者提出的问题: 有人跟甄大勇有仇吗\n", + "回答: 不重要\n", + "\n", + "谜面: 在一个安静的夜晚,小镇上的钟楼突然停止了报时。第二天早晨,人们发现钟楼的管理员甄大勇失踪了,而钟楼的门紧闭,从外面看起来一切正常。小镇上的人们议论纷纷,不知道发生了什么事情。\n", + "谜底: 真相是,钟楼的管理员甄大勇在夜晚进行例行的钟楼维护时,不慎从钟楼的顶部摔落,但并未死亡,只是昏迷。由于他跌落时砸到了控制时钟报时的机械装置,导致钟声停止。他躺在钟楼底部,但由于门从内部反锁,外面的人无法进入。甄大勇在第二天中午苏醒后,自己打开了门,这才知道自己引发了小镇上的恐慌。\n", + "参与者提出的问题: 他仅仅是在修钟楼吗\n", + "回答: 是\n", + "\n", + "谜面: 在一个安静的夜晚,小镇上的钟楼突然停止了报时。第二天早晨,人们发现钟楼的管理员甄大勇失踪了,而钟楼的门紧闭,从外面看起来一切正常。小镇上的人们议论纷纷,不知道发生了什么事情。\n", + "谜底: 真相是,钟楼的管理员甄大勇在夜晚进行例行的钟楼维护时,不慎从钟楼的顶部摔落,但并未死亡,只是昏迷。由于他跌落时砸到了控制时钟报时的机械装置,导致钟声停止。他躺在钟楼底部,但由于门从内部反锁,外面的人无法进入。甄大勇在第二天中午苏醒后,自己打开了门,这才知道自己引发了小镇上的恐慌。\n", + "参与者提出的问题: 是自然意外还是人为意外\n", + "回答: 问法错误\n", + "\n", + "谜面: 在一个安静的夜晚,小镇上的钟楼突然停止了报时。第二天早晨,人们发现钟楼的管理员甄大勇失踪了,而钟楼的门紧闭,从外面看起来一切正常。小镇上的人们议论纷纷,不知道发生了什么事情。\n", + "谜底: 真相是,钟楼的管理员甄大勇在夜晚进行例行的钟楼维护时,不慎从钟楼的顶部摔落,但并未死亡,只是昏迷。由于他跌落时砸到了控制时钟报时的机械装置,导致钟声停止。他躺在钟楼底部,但由于门从内部反锁,外面的人无法进入。甄大勇在第二天中午苏醒后,自己打开了门,这才知道自己引发了小镇上的恐慌。\n", + "参与者提出的问题: 因为甄在钟楼里维修然后昏迷了导致钟楼停止报时\n", + "回答: 回答正确\n", + "\n", + "谜面: 在一个安静的夜晚,小镇上的钟楼突然停止了报时。第二天早晨,人们发现钟楼的管理员甄大勇失踪了,而钟楼的门紧闭,从外面看起来一切正常。小镇上的人们议论纷纷,不知道发生了什么事情。\n", + "谜底: 真相是,钟楼的管理员甄大勇在夜晚进行例行的钟楼维护时,不慎从钟楼的顶部摔落,但并未死亡,只是昏迷。由于他跌落时砸到了控制时钟报时的机械装置,导致钟声停止。他躺在钟楼底部,但由于门从内部反锁,外面的人无法进入。甄大勇在第二天中午苏醒后,自己打开了门,这才知道自己引发了小镇上的恐慌。\n", + "参与者提出的问题: 是停止报时才导致甄大勇失踪吗?\n", + "回答: 不是\n", + "\n", + "谜面: 在一个安静的夜晚,小镇上的钟楼突然停止了报时。第二天早晨,人们发现钟楼的管理员甄大勇失踪了,而钟楼的门紧闭,���外面看起来一切正常。小镇上的人们议论纷纷,不知道发生了什么事情。\n", + "谜底: 真相是,钟楼的管理员甄大勇在夜晚进行例行的钟楼维护时,不慎从钟楼的顶部摔落,但并未死亡,只是昏迷。由于他跌落时砸到了控制时钟报时的机械装置,导致钟声停止。他躺在钟楼底部,但由于门从内部反锁,外面的人无法进入。甄大勇在第二天中午苏醒后,自己打开了门,这才知道自己引发了小镇上的恐慌。\n", + "参与者提出的问题: 甄大勇是和他人一起离开吗?\n", + "回答: 不重要\n", + "\n", + "谜面: 在一个安静的夜晚,小镇上的钟楼突然停止了报时。第二天早晨,人们发现钟楼的管理员甄大勇失踪了,而钟楼的门紧闭,从外面看起来一切正常。小镇上的人们议论纷纷,不知道发生了什么事情。\n", + "谜底: 真相是,钟楼的管理员甄大勇在夜晚进行例行的钟楼维护时,不慎从钟楼的顶部摔落,但并未死亡,只是昏迷。由于他跌落时砸到了控制时钟报时的机械装置,导致钟声停止。他躺在钟楼底部,但由于门从内部反锁,外面的人无法进入。甄大勇在第二天中午苏醒后,自己打开了门,这才知道自己引发了小镇上的恐慌。\n", + "参与者提出的问题: 是甄大勇弄坏了钟楼的报时器吗?\n", + "回答: 是\n", + "\n", + "谜面: 在一个安静的夜晚,小镇上的钟楼突然停止了报时。第二天早晨,人们发现钟楼的管理员甄大勇失踪了,而钟楼的门紧闭,从外面看起来一切正常。小镇上的人们议论纷纷,不知道发生了什么事情。\n", + "谜底: 真相是,钟楼的管理员甄大勇在夜晚进行例行的钟楼维护时,不慎从钟楼的顶部摔落,但并未死亡,只是昏迷。由于他跌落时砸到了控制时钟报时的机械装置,导致钟声停止。他躺在钟楼底部,但由于门从内部反锁,外面的人无法进入。甄大勇在第二天中午苏醒后,自己打开了门,这才知道自己引发了小镇上的恐慌。\n", + "参与者提出的问题: 别被人当时在此地吗?。\n", + "回答: 问法错误\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 老人认为的椅子其实一个梯子,被人拿走去维修东西了\n", + "回答: 回答正确\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 长椅坏了吗\n", + "回答: 不是\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路���,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 被人偷走了\n", + "回答: 不重要\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 他那张椅子是公园的吗\n", + "回答: 是\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 椅子是其他人受过伤吗\n", + "回答: 问法错误\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 维修工人在公园维修物品,用老人这个椅子来垫脚\n", + "回答: 回答正确\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 把椅子搬走是为了老头吗\n", + "回答: 不是\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角���里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 老头的椅子有不一样的地方\n", + "回答: 不重要\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 公园的灯坏了吗\n", + "回答: 是\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 原本有这个意思吗\n", + "回答: 问法错误\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 工人为了修东西搬来了椅子,老人认为椅子是固定的,一直用椅子,工人修好东西后把椅子搬走了\n", + "回答: 回答正确\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 是他的子女搬走他的椅子吗\n", + "回答: 不是\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 老���有疾病吗\n", + "回答: 不重要\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 他的椅子是正常椅子吗\n", + "回答: 是\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 这是一个悲伤的故事\n", + "回答: 问法错误\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 公园的这把椅子实际上是个梯子,公园里有东西坏了,所以梯子被维修人员拿去用了\n", + "回答: 回答正确\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 是重新刷漆了吗\n", + "回答: 不是\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 椅子很重要\n", + "回答: 不重要\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 椅子被挪走了\n", + "回答: 是\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 一块木头\n", + "回答: 问法错误\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 老人的椅子是个梯子,有人要维修东西将椅子搬走了\n", + "回答: 回答正确\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 椅子是被人偷走了吗\n", + "回答: 不是\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 椅子被盖住了吗\n", + "回答: 不重要\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 椅子在公园的其他地方\n", + "回答: 是\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 能给我点提示吗\n", + "回答: 问法错误\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 真正的椅子在公园那个椅子不是椅子被拿去维修了\n", + "回答: 回答正确\n", + "\n", + "\n", + "谜面: 在一个安静的夜晚,小村庄的湖边突然传来了阵阵哭泣声。第二天早晨,村长甄锐发现湖边的石头上放着一顶破旧的帽子,但没有人知道这顶帽子是从哪里来的,哭泣声又是为何。请还原故事真相。\n", + "谜底: 原来,这顶破旧的帽子属于一个小男孩,他小时候与爷爷在湖边生活。爷爷教他钓鱼、游泳,还告诉他湖中的海龟是他们的朋友。后来,小男孩随父母去了城市生活,但每年夏天都会回到村子探望爷爷。然而,去年夏天,爷爷因病去世,小男孩伤心欲绝。今年夏天,他回到村子,来到湖边,想起和爷爷的美好回忆,忍不住哭泣。他将爷爷的帽子放在湖边的石头上,希望能让爷爷的在天之灵得到安慰。那晚的哭泣声正是小男孩在祭莫他亲爱的爷爷。\n", + "参与者提出的问题: 死者受伤了吗\n", + "回答: \n", + "<|im_end|>\n", + "<|im_start|>assistant\n", + "不是<|im_end|>\n", + "--------------------------------------------------\n", + "prompt: <|im_start|>system\n", + "You are an expert in logical reasoning.<|im_end|>\n", + "<|im_start|>user\n", + "你是一个情景猜谜游戏的主持人。游戏规则如下:\n", + "\n", + "1. 参与者会得到一个谜面,谜面会描述一个简单又难以理解的事件。\n", + "2. 主持人知道谜底,谜底是谜面的答案。\n", + "3. 参与者可以询问任何封闭式问题来找寻事件的真相。\n", + "4. 对于每个问题,主持人将根据实际情况回答以下五个选项之一:是、不是、不重要、回答正确、问法错误。各回答的判断标准如下:\n", + " - 若谜面和谜底能找到问题的答案,回答:是或者不是\n", + " - 若谜面和谜底不能直接或者间接推断出问题的答案,回答:不重要\n", + " - 若参与者提问不是一个封闭式问题或者问题难以理解,回���:问法错误\n", + " - 若参与者提问基本还原了谜底真相,回答:回答正确\n", + "5. 回答中不能添加任何其它信息,也不能省略选项中的任何一个字。例如,不可以把“不是”省略成“不”。\n", + "\n", + "请严格按照这些规则回答参与者提出的问题。\n", + "\n", + "示例输入和输出: \n", + "谜面: 在甄家村里,有一个古老的传说:每年南瓜丰收的季节,南瓜田里总有一个最大的南瓜会不翼而飞,村民们对此现象困惑不解。请找出南瓜失踪背后的原因。\n", + "谜底: 真相原来与一位年迈的农夫有关。这位农夫年轻时,曾与一位美丽的姑娘相恋。他们约定在南瓜丰收的季节结婚。然而,命运弄人,姑娘在婚礼前的一场意外中离世。悲伤的农夫为了纪念心爱的姑娘,每年都会将最大的南瓜偷走,放到姑娘的墓前,以此寄托自己的哀思。这一行为延续了多年,成为了乡村里一个神秘的传说。\n", + "参与者提出的问题: 偷的人信神吗\n", + "回答: 不是\n", + "\n", + "谜面: 在甄家村里,有一个古老的传说:每年南瓜丰收的季节,南瓜田里总有一个最大的南瓜会不翼而飞,村民们对此现象困惑不解。请找出南瓜失踪背后的原因。\n", + "谜底: 真相原来与一位年迈的农夫有关。这位农夫年轻时,曾与一位美丽的姑娘相恋。他们约定在南瓜丰收的季节结婚。然而,命运弄人,姑娘在婚礼前的一场意外中离世。悲伤的农夫为了纪念心爱的姑娘,每年都会将最大的南瓜偷走,放到姑娘的墓前,以此寄托自己的哀思。这一行为延续了多年,成为了乡村里一个神秘的传说。\n", + "参与者提出的问题: 村庄里的人喜欢南瓜嘛\n", + "回答: 不重要\n", + "\n", + "谜面: 在甄家村里,有一个古老的传说:每年南瓜丰收的季节,南瓜田里总有一个最大的南瓜会不翼而飞,村民们对此现象困惑不解。请找出南瓜失踪背后的原因。\n", + "谜底: 真相原来与一位年迈的农夫有关。这位农夫年轻时,曾与一位美丽的姑娘相恋。他们约定在南瓜丰收的季节结婚。然而,命运弄人,姑娘在婚礼前的一场意外中离世。悲伤的农夫为了纪念心爱的姑娘,每年都会将最大的南瓜偷走,放到姑娘的墓前,以此寄托自己的哀思。这一行为延续了多年,成为了乡村里一个神秘的传说。\n", + "参与者提出的问题: 是村里的人偷的么\n", + "回答: 是\n", + "\n", + "谜面: 在一个炎热的夏日,乡村的甄家大院的西瓜突然全部不翼而飞。据了解,甄家大院周围并没有其他人家,而且门窗都完好无损,没有任何被撬的痕迹。村民们议论纷纷,猜测这批西瓜究竟去了哪里。你知道西瓜去了哪里吗?\n", + "谜底: 原来,这批西瓜是被一只巨大的乌鸦偷走了。这只乌鸦为了给自己的孩子们准备食物,它趁着夜色,竟然将甄家大院的西瓜一颗颗地带回了巢穴。第二天,村民们发现了乌鸦的巢穴,里面堆满了西瓜,而这个意外的真相让所有人都忍俊不禁。甄家老爷也感慨地说:“真是世界大了,什么奇事都有!”\n", + "参与者提出的问题: 挖地道\n", + "回答: 问法错误\n", + "\n", + "谜面: 在一个炎热的夏日,乡村的甄家大院的西瓜突然全部不翼而飞。据了解,甄家大院周围并没有其他人家,而且门窗都完好无损,没有任何被撬的痕迹。村民们议论纷纷,猜测这批西瓜究竟去了哪里。你知道西瓜去了哪里吗?\n", + "谜底: 原来,这批西瓜是被一只巨大的乌鸦偷走了。这只乌鸦为了给自己的孩子们准备食物,它趁着夜色,竟然将甄家大院的西瓜一颗颗地带回了巢穴。第二天,村民们发现了乌鸦的巢穴,里面堆满了西瓜,而这个意外的真相让所有人都忍俊不禁。甄家老爷也感慨地说:“真是世界大了,什么奇事都有!”\n", + "参与者提出的问题: 鸟觅食时发现甄家大院有西瓜,飞入大院一颗一颗把西瓜带走\n", + "回答: 回答正确\n", + "\n", + "谜面: 在一个安静的夜晚,小镇上的钟楼突然停止了报时。第二天早晨,人们发现钟楼的管理员甄大勇失踪了,而钟楼的门紧闭,从外面看起来一切正常。小镇上的人们议论纷纷,不知道发生了什么事情。\n", + "谜底: 真相是,钟楼的管理员甄大勇在夜晚进行例行的钟楼维护时,不慎从钟楼的顶部摔落,但并未死亡,只是昏迷。由于他跌落时砸到了控制时钟报时的机械装置,导致钟声停止。他躺在钟楼底部,但由于门从内部反锁,外面的人无法进入。甄大勇在第二天中午苏醒后,自己打开了门,这才知道自己引发了小镇上的恐慌。\n", + "参与者提出的问题: 有人身亡吗?\n", + "回答: 不是\n", + "\n", + "谜面: 在一个安静的夜晚,小镇上的钟楼突然停止了报时。第二天早晨,人们发现钟楼的管理员甄大勇失踪了,而钟楼的门紧闭,从外面看起来一切正常。小镇上的人们议论纷纷,不知道发生了什么事情。\n", + "谜底: 真相是,钟楼的管理员甄大勇在夜晚进行例行的钟楼维护时,不慎从钟楼的顶部摔落,但并未死亡,只是昏迷。由于他跌落时砸到了控制时钟报时的机械装置,导致钟声停止。他躺在钟楼底部,但由于门从内部反锁,外面的人无法进入。甄大勇在第二天中午苏醒后,自己打开了门,这才知道自己引发了小镇上的恐慌。\n", + "参与者提出的问题: 有人跟甄大勇有仇吗\n", + "回答: 不重要\n", + "\n", + "谜面: 在一个安静的夜晚,小镇上的钟楼突然停止了报时。第二天早晨,人们发现钟楼的管理员甄大勇失踪了,而钟楼的门紧闭,从外面看起来一切正常。小镇上的人们议论纷纷,不知道发生了什么事情。\n", + "谜底: 真相是,钟楼的管理员甄大勇在夜晚进行例行的钟楼维护时,不慎从钟楼的顶部摔落,但并未死亡,只是昏迷。由于他跌落时砸到了控制时钟报时的机械装置,导致钟声停止。他躺在钟楼底部,但由于门从内部反锁,外面的人无法进入。甄大勇在第二天中午苏醒后,自己打开了门,这才知道自己引发了小镇上的恐慌。\n", + "参与者提出的问题: 他仅仅是在修钟楼吗\n", + "回答: 是\n", + "\n", + "谜面: 在一个安静的夜晚,小镇上的钟楼突然停止了报时。第二天早晨,人们发现钟楼的管理员甄大勇失踪了,而钟楼的门紧闭,从外面看起来一切正常。小镇上的人们议论纷纷,不知道发生了什么事情。\n", + "谜底: 真相是,钟楼的管理员甄大勇在夜晚进行例行的钟楼维护时,不慎从钟楼的顶部摔落,但并未死亡,只是昏迷。由于他跌落时砸到了控制时钟报时的机械装置,导致钟声停止。他躺在钟楼底部,但由于门从内部反锁,外面的人无法进入。甄大勇在第二天中午苏醒后,自己打开了门,这才知道自己引发了小镇上的恐慌。\n", + "参与者提出的问题: 是自然意外还是人为意外\n", + "回答: 问法错误\n", + "\n", + "谜面: 在一个安静的夜晚,小镇上的钟楼突然停止了报时。第二天早晨,人们发现钟楼的管理员甄大勇失踪了,而钟楼的门紧闭,从外面看起来一切正常。小镇上的人们议论纷纷,不知道发生了什么事情。\n", + "谜底: 真相是,钟楼的管理员甄大勇在夜晚进行例行的钟楼维护时,不慎从钟楼的顶部摔落,但并未死亡,只是昏迷。由于他跌落时砸到了控制时钟报时的机械装置,导致钟声停止。他躺在钟楼底部,但由于门从内部反锁,外面的人无法进入。甄大勇在第二天中午苏醒后,自己打开了门,这才知道自己引发了小镇上的恐慌。\n", + "参与者提出的问题: 因为甄在钟楼里维修然后昏迷了导致钟楼停止报时\n", + "回答: 回答正确\n", + "\n", + "谜面: 在一个安静的夜晚,小镇上的钟楼突然停止了报时。第二天早晨,人们发现钟楼的管理员甄大勇失踪了,而钟楼的门紧闭,从外面看起来一切正常。小镇上的人们议论纷纷,不知道发生了什么事情。\n", + "谜底: 真相是,钟楼的管理员甄大勇在夜晚进行例行的钟楼维护时,不慎从钟楼的顶部摔落,但并未死亡,只是昏迷。由于他跌落时砸到了控制时钟报时的机械装置,导致钟声停止。他躺在钟楼底部,但由于门从内部反锁,外面的人无法进入。甄大勇在第二天中午苏醒后,自己打开了门,这才知道自己引发了小镇上的恐慌。\n", + "参与者提出的问题: 是停止报时才导致甄大勇失踪吗?\n", + "回答: 不是\n", + "\n", + "谜面: 在一个安静的夜晚,小镇上的钟楼突然停止了报时。第二天早晨,人们发现钟楼的管理员甄大勇失踪了,而钟楼的门紧闭,从外面看起来一切正常。小镇上的人们议论纷纷,不知道发生了什么事情。\n", + "谜底: 真相是,钟楼的管理员甄大勇在夜晚进行例行的钟楼维护时,不慎从钟楼的顶部摔落,但并未死亡,只是昏迷。由于他跌落时砸到了控制时钟报时的机械装置,导致钟声停止。他躺在钟楼底部,但由于门从内部反锁,外面的人无法进入。甄大勇在第二天中午苏醒后,自己打开了门,这才知道自己引发了小镇上的恐慌。\n", + "参与者提出的问题: 甄大勇是和他人一起离开吗?\n", + "回答: 不重要\n", + "\n", + "谜面: 在一个安静的夜晚,小镇上的钟楼突然停止了报时。第二天早晨,人们发现钟楼的管理员甄大勇失踪了,而钟楼的门紧闭,从外面看起来一切正常。小镇上的人们议论纷纷,不知道发生了什么事情。\n", + "谜底: 真相是,钟楼的管理员甄大勇在夜晚进行例行的钟楼维护时,不慎从钟楼的顶部摔落,但并未死亡,只是昏迷。由于他跌落时砸到了控制时钟报时的机械装置,导致钟声停止。他躺在钟楼底部,但由于门从内部反锁,外面的人无法进入。甄大勇在第二天中午苏醒后,自己打开了门,这才知道自己引发了小镇上的恐慌。\n", + "参与者提出的问题: 是甄大勇弄坏了钟楼的报时器吗?\n", + "回答: 是\n", + "\n", + "谜面: 在一个安静的夜晚,小镇上的钟楼突然停止了报时。第二天早晨,人们发现钟楼的管理员甄大勇失踪了,而钟楼的门紧闭,从外面看起来一切正常。小镇上的人们议论纷纷,不知道发生了什么事情。\n", + "谜底: 真相是,钟楼的管理员甄大勇在夜晚进行例行的钟楼维护时,不慎从钟楼的顶部摔落,但并未死亡,只是昏迷。由于他跌落时砸到了控制时钟报时的机械装置,导致钟声停止。他躺在钟楼底部,但由于门从内部反锁,外面的人无法进入。甄大勇在第二天中午苏醒后,自己打开了门,这才知道自己引发了小镇上的恐慌。\n", + "参与者提出的问题: 别被人当时在此地吗?。\n", + "回答: 问法错误\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 老人认为的椅子其实一个梯子,被人拿走去维修东西了\n", + "回答: 回答正确\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 长椅坏了吗\n", + "回答: 不是\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 被人偷走了\n", + "回答: 不重要\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 他那张椅子是公园的吗\n", + "回答: 是\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 椅子是其他人受过伤吗\n", + "回答: 问法错误\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 维修工人在公园维修物品,用老人这个椅子来垫脚\n", + "回答: 回答正确\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 把椅子搬走是为了老头吗\n", + "回答: 不是\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长��上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 老头的椅子有不一样的地方\n", + "回答: 不重要\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 公园的灯坏了吗\n", + "回答: 是\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 原本有这个意思吗\n", + "回答: 问法错误\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 工人为了修东西搬来了椅子,老人认为椅子是固定的,一直用椅子,工人修好东西后把椅子搬走了\n", + "回答: 回答正确\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 是他的子女搬走他的椅子吗\n", + "回答: 不是\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其��椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 老人有疾病吗\n", + "回答: 不重要\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 他的椅子是正常椅子吗\n", + "回答: 是\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 这是一个悲伤的故事\n", + "回答: 问法错误\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 公园的这把椅子实际上是个梯子,公园里有东西坏了,所以梯子被维修人员拿去用了\n", + "回答: 回答正确\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 是重新刷漆了吗\n", + "回答: 不是\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝��不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 椅子很重要\n", + "回答: 不重要\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 椅子被挪走了\n", + "回答: 是\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 一块木头\n", + "回答: 问法错误\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 老人的椅子是个梯子,有人要维修东西将椅子搬走了\n", + "回答: 回答正确\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 椅子是被人偷走了吗\n", + "回答: 不是\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无��地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 椅子被盖住了吗\n", + "回答: 不重要\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 椅子在公园的其他地方\n", + "回答: 是\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 能给我点提示吗\n", + "回答: 问法错误\n", + "\n", + "谜面: 一个晴朗的周末,公园的长椅上发生了一件奇怪的事情。一位老人每天下午都会来到这个公园的同一张长椅上看书。但今天,当他像往常一样来到公园时,却发现他的椅子不见了。更奇怪的是,公园里的其他椅子都完好无损,唯独他那张椅子不见了踪影。\n", + "谜底: 真相是,公园里的园艺工人前一天晚上在修剪树枝时不小心弄坏了一盏路灯。他们为了修理路灯,临时把那张长椅搬过去当作梯子使用。修理完毕后,他们忘了把长椅放回原位,导致老人找不到自己的椅子。而那张长椅,此刻正无辜地躺在公园的角落里,上面还写着:“我不是椅子,我是临时梯子。”\n", + "参与者提出的问题: 真正的椅子在公园那个椅子不是椅子被拿去维修了\n", + "回答: 回答正确\n", + "\n", + "\n", + "谜面: 在一个安静的夜晚,小村庄的湖边突然传来了阵阵哭泣声。第二天早晨,村长甄锐发现湖边的石头上放着一顶破旧的帽子,但没有人知道这顶帽子是从哪里来的,哭泣声又是为何。请还原故事真相。\n", + "谜底: 原来,这顶破旧的帽子属于一个小男孩,他小时候与爷爷在湖边生活。爷爷教他钓鱼、游泳,还告诉他湖中的海龟是他们的朋友。后来,小男孩随父母去了城市生活,但每年夏天都会回到村子探望爷爷。然而,去年夏天,爷爷因病去世,小男孩伤心欲绝。今年夏天,他回到村子,来到湖边,想起和爷爷的美好回忆,忍不住哭泣。他将爷爷的帽子放在湖边的石头上,希望能让爷爷的在天之灵得到安慰。那晚的哭泣声正是小男孩在祭莫他亲爱的爷爷。\n", + "参与者提出的问题: 死者受伤了吗\n", + "回答: \n", + "<|im_end|>\n", + "<|im_start|>assistant\n", + "\n", + " 0%| | 0/3000 [00:00