Update app.py
Browse files
app.py
CHANGED
@@ -32,6 +32,7 @@ class ChatResponse(BaseModel):
|
|
32 |
@app.post("/chat", response_model=ChatResponse)
|
33 |
async def chat(request: ChatRequest):
|
34 |
try:
|
|
|
35 |
# messages = [
|
36 |
# {"role": "system", "content": request.system_message},
|
37 |
# {"role": "user", "content": request.message},
|
@@ -43,7 +44,7 @@ async def chat(request: ChatRequest):
|
|
43 |
# temperature=request.temperature,
|
44 |
# top_p=request.top_p,
|
45 |
# )
|
46 |
-
response = ollama.chat(model='llama3.2', messages=[{'role': 'user', 'content': 'Hello!'}])
|
47 |
|
48 |
|
49 |
# return {"response": response.choices[0].message.content}
|
|
|
32 |
@app.post("/chat", response_model=ChatResponse)
|
33 |
async def chat(request: ChatRequest):
|
34 |
try:
|
35 |
+
response = "".join(reversed(request.message))
|
36 |
# messages = [
|
37 |
# {"role": "system", "content": request.system_message},
|
38 |
# {"role": "user", "content": request.message},
|
|
|
44 |
# temperature=request.temperature,
|
45 |
# top_p=request.top_p,
|
46 |
# )
|
47 |
+
# response = ollama.chat(model='llama3.2', messages=[{'role': 'user', 'content': 'Hello!'}])
|
48 |
|
49 |
|
50 |
# return {"response": response.choices[0].message.content}
|