|
import time |
|
import gradio as gr |
|
from ctransformers import AutoModelForCausalLM |
|
|
|
|
|
llm = AutoModelForCausalLM.from_pretrained("TheBloke/WizardLM-7B-uncensored-GGUF", model_file="WizardLM-7B-uncensored.Q4_K_M.gguf", model_type="llama", stream=True) |
|
|
|
|
|
def generate_response(message): |
|
|
|
for text in llm(message): |
|
yield text |
|
yi |
|
|
|
def chatbot(message, history): |
|
response_generator = generate_response(message) |
|
for response in response_generator: |
|
time.sleep(0.3) |
|
|
|
x= "".join(response) |
|
yield x |
|
|
|
iface = gr.ChatInterface(chatbot) |
|
iface.launch() |
|
|