Copain22 commited on
Commit
61d97ef
·
verified ·
1 Parent(s): b36fdc1

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +5 -6
app.py CHANGED
@@ -49,18 +49,17 @@ _state = {"chat_engine": None}
49
  def get_chat_engine():
50
  if _state["chat_engine"] is None:
51
  llm = HuggingFaceLLM(
52
- tokenizer_name="meta-llama/Llama-2-7b-chat-hf",
53
- model_name="meta-llama/Llama-2-7b-chat-hf",
54
- context_window=3900,
55
  max_new_tokens=256,
56
  generate_kwargs={"temperature": 0.2, "do_sample": True},
57
  device_map="auto",
58
  model_kwargs={
59
  "use_auth_token": os.environ["HF_TOKEN"]
60
- },
61
  system_prompt=SYSTEM_PROMPT,
62
- query_wrapper_prompt=WRAPPER_PROMPT,
63
- )
64
  Settings.llm = llm
65
 
66
  memory = ChatMemoryBuffer.from_defaults(token_limit=2000)
 
49
  def get_chat_engine():
50
  if _state["chat_engine"] is None:
51
  llm = HuggingFaceLLM(
52
+ tokenizer_name="TinyLlama/TinyLlama-1.1B-Chat-v1.0",
53
+ model_name="TinyLlama/TinyLlama-1.1B-Chat-v1.0",
54
+ context_window=2048,
55
  max_new_tokens=256,
56
  generate_kwargs={"temperature": 0.2, "do_sample": True},
57
  device_map="auto",
58
  model_kwargs={
59
  "use_auth_token": os.environ["HF_TOKEN"]
60
+ },
61
  system_prompt=SYSTEM_PROMPT,
62
+ query_wrapper_prompt=WRAPPER_PROMPT,)
 
63
  Settings.llm = llm
64
 
65
  memory = ChatMemoryBuffer.from_defaults(token_limit=2000)