peterpeter8585 commited on
Commit
90de6d2
·
verified ·
1 Parent(s): 3b6cd5e

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +0 -2
app.py CHANGED
@@ -94,9 +94,7 @@ llm = ChatLlamaCpp(
94
  temperature=0,
95
  model_path="./llama-3-open-ko-8b-instruct-preview-q5_k_m.gguf",
96
  n_ctx=10000,
97
- n_gpu_layers=8,
98
  n_batch=300, # Should be between 1 and n_ctx, consider the amount of VRAM in your GPU.
99
- max_tokens=512,
100
  n_threads=multiprocessing.cpu_count() - 1,
101
  repeat_penalty=1.5,
102
  top_p=0.5,
 
94
  temperature=0,
95
  model_path="./llama-3-open-ko-8b-instruct-preview-q5_k_m.gguf",
96
  n_ctx=10000,
 
97
  n_batch=300, # Should be between 1 and n_ctx, consider the amount of VRAM in your GPU.
 
98
  n_threads=multiprocessing.cpu_count() - 1,
99
  repeat_penalty=1.5,
100
  top_p=0.5,