Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
@@ -106,7 +106,7 @@ if st.session_state["qa_pipeline"]:
|
|
106 |
with st.spinner("Generating response..."):
|
107 |
try:
|
108 |
# Generate the model response for general QA (T5 model)
|
109 |
-
response = st.session_state["qa_pipeline"](f"Q: {user_input}", max_length=
|
110 |
generated_text = response[0]["generated_text"]
|
111 |
|
112 |
# Update the conversation
|
@@ -128,7 +128,7 @@ else:
|
|
128 |
value=st.session_state["user_input"],
|
129 |
key="calculation_input",
|
130 |
)
|
131 |
-
if st.button("
|
132 |
if user_input:
|
133 |
with st.spinner("Generating response..."):
|
134 |
try:
|
@@ -140,7 +140,7 @@ else:
|
|
140 |
output = st.session_state["model"].generate(
|
141 |
input_ids=input_ids,
|
142 |
attention_mask=attention_mask,
|
143 |
-
max_length=
|
144 |
pad_token_id=st.session_state["tokenizer"].pad_token_id,
|
145 |
eos_token_id=st.session_state["tokenizer"].eos_token_id,
|
146 |
do_sample=False
|
|
|
106 |
with st.spinner("Generating response..."):
|
107 |
try:
|
108 |
# Generate the model response for general QA (T5 model)
|
109 |
+
response = st.session_state["qa_pipeline"](f"Q: {user_input}", max_length=250)
|
110 |
generated_text = response[0]["generated_text"]
|
111 |
|
112 |
# Update the conversation
|
|
|
128 |
value=st.session_state["user_input"],
|
129 |
key="calculation_input",
|
130 |
)
|
131 |
+
if st.button("Submit", key="send_calculation_button"):
|
132 |
if user_input:
|
133 |
with st.spinner("Generating response..."):
|
134 |
try:
|
|
|
140 |
output = st.session_state["model"].generate(
|
141 |
input_ids=input_ids,
|
142 |
attention_mask=attention_mask,
|
143 |
+
max_length=250,
|
144 |
pad_token_id=st.session_state["tokenizer"].pad_token_id,
|
145 |
eos_token_id=st.session_state["tokenizer"].eos_token_id,
|
146 |
do_sample=False
|