Update app.py
Browse files
app.py
CHANGED
@@ -38,6 +38,7 @@ model = AutoModelForCausalLM.from_pretrained(
|
|
38 |
device_map="auto",
|
39 |
ignore_mismatched_sizes=True)
|
40 |
|
|
|
41 |
|
42 |
@spaces.GPU()
|
43 |
def stream_chat(
|
@@ -75,7 +76,8 @@ def stream_chat(
|
|
75 |
top_k=top_k,
|
76 |
temperature=temperature,
|
77 |
streamer=streamer,
|
78 |
-
pad_token_id=
|
|
|
79 |
)
|
80 |
|
81 |
with torch.no_grad():
|
@@ -86,6 +88,8 @@ def stream_chat(
|
|
86 |
for new_text in streamer:
|
87 |
buffer += new_text
|
88 |
yield buffer
|
|
|
|
|
89 |
|
90 |
|
91 |
chatbot = gr.Chatbot(height=600, placeholder=PLACEHOLDER)
|
|
|
38 |
device_map="auto",
|
39 |
ignore_mismatched_sizes=True)
|
40 |
|
41 |
+
eos_token_id = tokenizer.eos_token_id
|
42 |
|
43 |
@spaces.GPU()
|
44 |
def stream_chat(
|
|
|
76 |
top_k=top_k,
|
77 |
temperature=temperature,
|
78 |
streamer=streamer,
|
79 |
+
pad_token_id=eos_token_id,
|
80 |
+
eos_token_id=eos_token_id,
|
81 |
)
|
82 |
|
83 |
with torch.no_grad():
|
|
|
88 |
for new_text in streamer:
|
89 |
buffer += new_text
|
90 |
yield buffer
|
91 |
+
if eos_token_id in tokenizer.encode(new_text):
|
92 |
+
break
|
93 |
|
94 |
|
95 |
chatbot = gr.Chatbot(height=600, placeholder=PLACEHOLDER)
|