Update app.py
Browse files
app.py
CHANGED
@@ -60,7 +60,7 @@ SYSTEM_PROMPT = "You are a friendly Chatbot."
|
|
60 |
@spaces.GPU
|
61 |
def generate(prompt, history):
|
62 |
messages = [
|
63 |
-
{"role": "system", "content": SYSTEM_PROMPT},
|
64 |
{"role": "user", "content": prompt}
|
65 |
]
|
66 |
# text = tokenizer.apply_chat_template(
|
@@ -87,10 +87,10 @@ def generate(prompt, history):
|
|
87 |
# tokenize=True,
|
88 |
return_dict=True,
|
89 |
return_tensors="pt",
|
90 |
-
)
|
91 |
outputs = model.generate(
|
92 |
-
**inputs,
|
93 |
-
max_new_tokens=
|
94 |
)
|
95 |
response = processor.batch_decode(outputs[:, inputs["input_ids"].shape[-1]:])[0]
|
96 |
|
|
|
60 |
@spaces.GPU
|
61 |
def generate(prompt, history):
|
62 |
messages = [
|
63 |
+
# {"role": "system", "content": SYSTEM_PROMPT},
|
64 |
{"role": "user", "content": prompt}
|
65 |
]
|
66 |
# text = tokenizer.apply_chat_template(
|
|
|
87 |
# tokenize=True,
|
88 |
return_dict=True,
|
89 |
return_tensors="pt",
|
90 |
+
)
|
91 |
outputs = model.generate(
|
92 |
+
**inputs.to(model.device),
|
93 |
+
max_new_tokens=100,
|
94 |
)
|
95 |
response = processor.batch_decode(outputs[:, inputs["input_ids"].shape[-1]:])[0]
|
96 |
|