vilarin commited on
Commit
8951355
Β·
verified Β·
1 Parent(s): e18b750

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +3 -4
app.py CHANGED
@@ -5,8 +5,7 @@ import os
5
  from huggingface_hub import hf_hub_download
6
  import base64
7
  from llama_cpp import Llama
8
- from llama_cpp.llama_chat_format import LlamaChatCompletionHandler
9
-
10
 
11
  os.environ["HF_HUB_ENABLE_HF_TRANSFER"] = "1"
12
  MODEL_LIST = ["openbmb/MiniCPM-Llama3-V-2_5","openbmb/MiniCPM-Llama3-V-2_5-int4"]
@@ -27,7 +26,7 @@ CSS = """
27
  }
28
  """
29
 
30
- chat_handler = LlamaChatCompletionHandler.from_pretrained(
31
  repo_id="openbmb/MiniCPM-Llama3-V-2_5-gguf",
32
  filename="*mmproj*",
33
  )
@@ -35,7 +34,7 @@ chat_handler = LlamaChatCompletionHandler.from_pretrained(
35
  llm = Llama.from_pretrained(
36
  repo_id="openbmb/MiniCPM-Llama3-V-2_5-gguf",
37
  filename="ggml-model-Q5_K_M.gguf",
38
- chat_handler=chat_handler,
39
  n_ctx=4096,
40
  verbose=True
41
  )
 
5
  from huggingface_hub import hf_hub_download
6
  import base64
7
  from llama_cpp import Llama
8
+ from llama_cpp.llama_tokenizer import LlamaHFTokenizer
 
9
 
10
  os.environ["HF_HUB_ENABLE_HF_TRANSFER"] = "1"
11
  MODEL_LIST = ["openbmb/MiniCPM-Llama3-V-2_5","openbmb/MiniCPM-Llama3-V-2_5-int4"]
 
26
  }
27
  """
28
 
29
+ tokenizer = LlamaHFTokenizer.from_pretrained(
30
  repo_id="openbmb/MiniCPM-Llama3-V-2_5-gguf",
31
  filename="*mmproj*",
32
  )
 
34
  llm = Llama.from_pretrained(
35
  repo_id="openbmb/MiniCPM-Llama3-V-2_5-gguf",
36
  filename="ggml-model-Q5_K_M.gguf",
37
+ tokenizer=tokenizer,
38
  n_ctx=4096,
39
  verbose=True
40
  )