import asyncio from ollama import AsyncClient import streamlit as st async def chat(prompt): message = {'role': 'user', 'content': prompt} async for part in await AsyncClient().chat(model='deepseek-r1:1.5b', messages=[message], stream=True): yield part['message']['content'] async def get_full_response(prompt): response = "" async for part in chat(prompt): response += part return response def main(): st.title("Deepseek-R1 Chat") prompt = st.text_input("Enter your prompt:") if st.button("Generate"): if prompt: with st.spinner("Generating response..."): full_response = asyncio.run(get_full_response(prompt)) st.write(full_response) if __name__ == '__main__': main()