sims2k commited on
Commit
0156a02
Β·
verified Β·
1 Parent(s): 5eb9d36

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +2 -2
app.py CHANGED
@@ -48,13 +48,13 @@ whisper_model = WhisperModel("large-v3", device="cuda", compute_type="float16")
48
  print("Loading Mistral LLM")
49
  hf_hub_download(repo_id="TheBloke/Mistral-7B-Instruct-v0.1-GGUF", local_dir=".", filename="mistral-7b-instruct-v0.1.Q5_K_M.gguf")
50
  mistral_model_path="./mistral-7b-instruct-v0.1.Q5_K_M.gguf"
51
- mistral_llm = Llama(model_path=mistral_model_path,n_gpu_layers=35,max_new_tokens=256, context_window=4096, n_ctx=4096,n_batch=128,verbose=False)
52
 
53
  # Load Saul-Instruct-v1-GGUF.Q4_K_M
54
  print("Loading Saul-Instruct-v1-GGUF.Q4_K_M")
55
  hf_hub_download(repo_id="MaziyarPanahi/Saul-Instruct-v1-GGUF", local_dir=".", filename="Saul-Instruct-v1.Q4_K_M.gguf")
56
  saul_model_path="./Saul-Instruct-v1.Q4_K_M.gguf"
57
- saul_instruct_llm = Llama(model_path=saul_model_path,n_gpu_layers=35,max_new_tokens=256, context_window=4096, n_ctx=32768,n_batch=128,verbose=False)
58
 
59
  # Load XTTS Model
60
  print("Loading XTTS model")
 
48
  print("Loading Mistral LLM")
49
  hf_hub_download(repo_id="TheBloke/Mistral-7B-Instruct-v0.1-GGUF", local_dir=".", filename="mistral-7b-instruct-v0.1.Q5_K_M.gguf")
50
  mistral_model_path="./mistral-7b-instruct-v0.1.Q5_K_M.gguf"
51
+ mistral_llm = Llama(model_path=mistral_model_path,n_gpu_layers=35,max_new_tokens=256, context_window=4096, n_ctx=4096, n_batch=128,verbose=False)
52
 
53
  # Load Saul-Instruct-v1-GGUF.Q4_K_M
54
  print("Loading Saul-Instruct-v1-GGUF.Q4_K_M")
55
  hf_hub_download(repo_id="MaziyarPanahi/Saul-Instruct-v1-GGUF", local_dir=".", filename="Saul-Instruct-v1.Q4_K_M.gguf")
56
  saul_model_path="./Saul-Instruct-v1.Q4_K_M.gguf"
57
+ saul_instruct_llm = Llama(model_path=saul_model_path,n_gpu_layers=35,max_new_tokens=256, context_window=4096, n_ctx=4096, n_batch=128,verbose=False)
58
 
59
  # Load XTTS Model
60
  print("Loading XTTS model")