syedmoinms commited on
Commit
f12abd4
Β·
verified Β·
1 Parent(s): 24b7fd6

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +20 -25
app.py CHANGED
@@ -1,29 +1,24 @@
1
- import gradio as gr
2
- from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
3
- from memory import update_memory, check_memory
4
 
5
- with open("persona.txt", "r", encoding="utf-8") as f:
6
- personality = f.read()
7
 
8
- model_name = "TheBloke/Pygmalion-7B-GPTQ"
9
  tokenizer = AutoTokenizer.from_pretrained(model_name)
10
- model = AutoModelForCausalLM.from_pretrained(model_name, device_map="auto", trust_remote_code=True)
11
 
12
- generator = pipeline("text-generation", model=model, tokenizer=tokenizer)
13
-
14
- def chatbot(input_text):
15
- memory_response = check_memory(input_text)
16
- if memory_response:
17
- return memory_response
18
-
19
- prompt = f"{personality}\nUser: {input_text}\nAI:"
20
- outputs = generator(prompt, max_length=200, do_sample=True, temperature=0.7, top_p=0.9)
21
- response = outputs[0]["generated_text"].split("AI:")[-1].strip()
22
-
23
- update_memory(input_text, response)
24
- return response
25
-
26
- iface = gr.Interface(fn=chatbot, inputs="text", outputs="text", title="MoinRomanticBot")
27
-
28
- if __name__ == "__main__":
29
- iface.launch(server_name="0.0.0.0", server_port=7860)
 
1
+ from transformers import AutoTokenizer, AutoModelForCausalLM
2
+ import torch
 
3
 
4
+ # βœ… Model name update karo
5
+ model_name = "TheBloke/Pygmalion-13B-SuperHOT-8K-GPTQ"
6
 
7
+ # βœ… Tokenizer Load karo
8
  tokenizer = AutoTokenizer.from_pretrained(model_name)
 
9
 
10
+ # βœ… Model Load karo (Automatic Device Selection)
11
+ model = AutoModelForCausalLM.from_pretrained(model_name, device_map="auto", torch_dtype=torch.float16)
12
+
13
+ def chat(prompt):
14
+ inputs = tokenizer(prompt, return_tensors="pt").to("cuda")
15
+ output = model.generate(**inputs, max_new_tokens=200, do_sample=True, temperature=0.7, top_p=0.9)
16
+ return tokenizer.decode(output[0], skip_special_tokens=True)
17
+
18
+ # βœ… Test Chat
19
+ while True:
20
+ user_input = input("You: ")
21
+ if user_input.lower() in ["exit", "quit"]:
22
+ break
23
+ response = chat(user_input)
24
+ print("AI:", response)