FluentQ / models /local_llm.py
tommytracx's picture
Upload local_llm.py
d134f64 verified
raw
history blame
221 Bytes
import subprocess
def run_llm(prompt: str) -> str:
result = subprocess.run([
"./main", "-m", "models/ggml-model.bin", "-p", prompt, "-n", "128"
], capture_output=True, text=True)
return result.stdout