Update app.py
Browse files
app.py
CHANGED
@@ -30,7 +30,7 @@ prompt_template = f""""""
|
|
30 |
@app.post("/chat", response_model=ChatResponse)
|
31 |
async def chat(request: ChatRequest):
|
32 |
try:
|
33 |
-
if model_choice == "HF":
|
34 |
if hf_token:
|
35 |
client = InferenceClient("meta-llama/Llama-3.2-3B-Instruct", token=hf_token)
|
36 |
else:
|
@@ -50,7 +50,7 @@ async def chat(request: ChatRequest):
|
|
50 |
|
51 |
return {"response": response.choices[0].message.content}
|
52 |
|
53 |
-
if model_choice == "google":
|
54 |
genai.configure(api_key=google_api_key)
|
55 |
model = genai.GenerativeModel("gemini-2.0-flash")
|
56 |
|
|
|
30 |
@app.post("/chat", response_model=ChatResponse)
|
31 |
async def chat(request: ChatRequest):
|
32 |
try:
|
33 |
+
if request.model_choice == "HF":
|
34 |
if hf_token:
|
35 |
client = InferenceClient("meta-llama/Llama-3.2-3B-Instruct", token=hf_token)
|
36 |
else:
|
|
|
50 |
|
51 |
return {"response": response.choices[0].message.content}
|
52 |
|
53 |
+
if request.model_choice == "google":
|
54 |
genai.configure(api_key=google_api_key)
|
55 |
model = genai.GenerativeModel("gemini-2.0-flash")
|
56 |
|