Update app.py
Browse files
app.py
CHANGED
@@ -48,13 +48,13 @@ whisper_model = WhisperModel("large-v3", device="cuda", compute_type="float16")
|
|
48 |
print("Loading Mistral LLM")
|
49 |
hf_hub_download(repo_id="TheBloke/Mistral-7B-Instruct-v0.1-GGUF", local_dir=".", filename="mistral-7b-instruct-v0.1.Q5_K_M.gguf")
|
50 |
mistral_model_path="./mistral-7b-instruct-v0.1.Q5_K_M.gguf"
|
51 |
-
mistral_llm = Llama(model_path=mistral_model_path,n_gpu_layers=35,max_new_tokens=256, context_window=4096, n_ctx=4096,n_batch=128,verbose=False)
|
52 |
|
53 |
# Load Saul-Instruct-v1-GGUF.Q4_K_M
|
54 |
print("Loading Saul-Instruct-v1-GGUF.Q4_K_M")
|
55 |
hf_hub_download(repo_id="MaziyarPanahi/Saul-Instruct-v1-GGUF", local_dir=".", filename="Saul-Instruct-v1.Q4_K_M.gguf")
|
56 |
saul_model_path="./Saul-Instruct-v1.Q4_K_M.gguf"
|
57 |
-
saul_instruct_llm = Llama(model_path=saul_model_path,n_gpu_layers=35,max_new_tokens=256, context_window=4096, n_ctx=
|
58 |
|
59 |
# Load XTTS Model
|
60 |
print("Loading XTTS model")
|
|
|
48 |
print("Loading Mistral LLM")
|
49 |
hf_hub_download(repo_id="TheBloke/Mistral-7B-Instruct-v0.1-GGUF", local_dir=".", filename="mistral-7b-instruct-v0.1.Q5_K_M.gguf")
|
50 |
mistral_model_path="./mistral-7b-instruct-v0.1.Q5_K_M.gguf"
|
51 |
+
mistral_llm = Llama(model_path=mistral_model_path,n_gpu_layers=35,max_new_tokens=256, context_window=4096, n_ctx=4096, n_batch=128,verbose=False)
|
52 |
|
53 |
# Load Saul-Instruct-v1-GGUF.Q4_K_M
|
54 |
print("Loading Saul-Instruct-v1-GGUF.Q4_K_M")
|
55 |
hf_hub_download(repo_id="MaziyarPanahi/Saul-Instruct-v1-GGUF", local_dir=".", filename="Saul-Instruct-v1.Q4_K_M.gguf")
|
56 |
saul_model_path="./Saul-Instruct-v1.Q4_K_M.gguf"
|
57 |
+
saul_instruct_llm = Llama(model_path=saul_model_path,n_gpu_layers=35,max_new_tokens=256, context_window=4096, n_ctx=4096, n_batch=128,verbose=False)
|
58 |
|
59 |
# Load XTTS Model
|
60 |
print("Loading XTTS model")
|