VitaliiPuzhenko commited on
Commit
4e5b21c
·
verified ·
1 Parent(s): 923d7e0

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +17 -12
app.py CHANGED
@@ -1,25 +1,30 @@
1
- from fastapi import FastAPI, Request
2
  from transformers import pipeline
3
  import os
4
 
5
  app = FastAPI()
6
 
7
- # Встановлення директорії кешу
8
  cache_dir = os.path.expanduser("~/.cache/huggingface/transformers")
9
  os.environ["TRANSFORMERS_CACHE"] = cache_dir
10
 
11
- pipe = pipeline(
12
- "text-generation",
13
- model="google/gemma-3-1b-it",
14
- token=os.environ["HF_TOKEN"],
15
- device="cpu",
16
- max_new_tokens=256,
17
- temperature=0.7
18
- )
 
 
 
19
 
20
  @app.post("/generate")
21
  async def generate(request: Request):
22
  data = await request.json()
23
  prompt = data.get("prompt", "")
24
- result = pipe(prompt)
25
- return {"response": result[0]["generated_text"]}
 
 
 
 
1
+ from fastapi import FastAPI, Request, HTTPException
2
  from transformers import pipeline
3
  import os
4
 
5
  app = FastAPI()
6
 
 
7
  cache_dir = os.path.expanduser("~/.cache/huggingface/transformers")
8
  os.environ["TRANSFORMERS_CACHE"] = cache_dir
9
 
10
+ try:
11
+ pipe = pipeline(
12
+ "text-generation",
13
+ model="google/gemma-3-1b-it",
14
+ token=os.environ["HF_TOKEN"],
15
+ device="cpu",
16
+ max_new_tokens=256,
17
+ temperature=0.7
18
+ )
19
+ except Exception as e:
20
+ raise HTTPException(status_code=500, detail=str(e))
21
 
22
  @app.post("/generate")
23
  async def generate(request: Request):
24
  data = await request.json()
25
  prompt = data.get("prompt", "")
26
+ try:
27
+ result = pipe(prompt)
28
+ return {"response": result[0]["generated_text"]}
29
+ except Exception as e:
30
+ raise HTTPException(status_code=500, detail=str(e))