|
import time |
|
import gradio as gr |
|
from ctransformers import AutoModelForCausalLM |
|
|
|
|
|
llm = AutoModelForCausalLM.from_pretrained("TheBloke/WizardLM-7B-uncensored-GGUF", model_file="WizardLM-7B-uncensored.Q3_K_M.gguf", model_type="llama", stream=True) |
|
history = [""] |
|
|
|
def generate_response(message): |
|
global history |
|
for text in llm(message): |
|
|
|
response = ''.join(text) |
|
time.sleep(2) |
|
history.append(response) |
|
yield ' '.join(history) |
|
|
|
history = ["Chatbot:"] |
|
|
|
def chatbot(message, history): |
|
response_generator = generate_response(message) |
|
for response in response_generator: |
|
time.sleep(0.1) |
|
yield response |
|
|
|
iface = gr.ChatInterface(chatbot) |
|
iface.launch() |
|
|