fix prompt
Browse files
app.py
CHANGED
@@ -71,7 +71,6 @@ def chat_with_model(messages):
|
|
71 |
current_model.to(device).half()
|
72 |
|
73 |
# 1. Tokenize prompt
|
74 |
-
prompt = "Your input here"
|
75 |
inputs = current_tokenizer(prompt, return_tensors="pt").to(device)
|
76 |
prompt_len = inputs["input_ids"].shape[-1]
|
77 |
|
|
|
71 |
current_model.to(device).half()
|
72 |
|
73 |
# 1. Tokenize prompt
|
|
|
74 |
inputs = current_tokenizer(prompt, return_tensors="pt").to(device)
|
75 |
prompt_len = inputs["input_ids"].shape[-1]
|
76 |
|