Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
@@ -230,14 +230,12 @@ from fastapi.responses import HTMLResponse, JSONResponse
|
|
230 |
from fastapi.staticfiles import StaticFiles
|
231 |
from llama_index.core import StorageContext, load_index_from_storage, VectorStoreIndex, SimpleDirectoryReader, ChatPromptTemplate, Settings
|
232 |
from llama_index.embeddings.huggingface import HuggingFaceEmbedding
|
|
|
233 |
from pydantic import BaseModel
|
234 |
from fastapi.middleware.cors import CORSMiddleware
|
235 |
from fastapi.templating import Jinja2Templates
|
236 |
-
from huggingface_hub import InferenceClient
|
237 |
-
import json
|
238 |
import datetime
|
239 |
from simple_salesforce import Salesforce, SalesforceLogin
|
240 |
-
from llama_index.llms.huggingface import HuggingFaceLLM
|
241 |
|
242 |
# Pydantic model for request body
|
243 |
class MessageRequest(BaseModel):
|
@@ -279,7 +277,7 @@ async def add_security_headers(request: Request, call_next):
|
|
279 |
# CORS configuration
|
280 |
app.add_middleware(
|
281 |
CORSMiddleware,
|
282 |
-
allow_origins=["*"], #
|
283 |
allow_credentials=True,
|
284 |
allow_methods=["*"],
|
285 |
allow_headers=["*"],
|
@@ -290,13 +288,11 @@ app.mount("/static", StaticFiles(directory="static"), name="static")
|
|
290 |
templates = Jinja2Templates(directory="static")
|
291 |
|
292 |
# LlamaIndex configuration
|
293 |
-
Settings.llm =
|
294 |
model_name="meta-llama/Meta-Llama-3-8B-Instruct",
|
295 |
-
tokenizer_name="meta-llama/Meta-Llama-3-8B-Instruct",
|
296 |
-
context_window=3000,
|
297 |
token=os.getenv("HF_TOKEN"),
|
298 |
max_new_tokens=512,
|
299 |
-
|
300 |
)
|
301 |
|
302 |
Settings.embed_model = HuggingFaceEmbedding(
|
|
|
230 |
from fastapi.staticfiles import StaticFiles
|
231 |
from llama_index.core import StorageContext, load_index_from_storage, VectorStoreIndex, SimpleDirectoryReader, ChatPromptTemplate, Settings
|
232 |
from llama_index.embeddings.huggingface import HuggingFaceEmbedding
|
233 |
+
from llama_index.llms.huggingface_api import HuggingFaceInferenceAPI
|
234 |
from pydantic import BaseModel
|
235 |
from fastapi.middleware.cors import CORSMiddleware
|
236 |
from fastapi.templating import Jinja2Templates
|
|
|
|
|
237 |
import datetime
|
238 |
from simple_salesforce import Salesforce, SalesforceLogin
|
|
|
239 |
|
240 |
# Pydantic model for request body
|
241 |
class MessageRequest(BaseModel):
|
|
|
277 |
# CORS configuration
|
278 |
app.add_middleware(
|
279 |
CORSMiddleware,
|
280 |
+
allow_origins=["*"], # Specify allowed origins in production
|
281 |
allow_credentials=True,
|
282 |
allow_methods=["*"],
|
283 |
allow_headers=["*"],
|
|
|
288 |
templates = Jinja2Templates(directory="static")
|
289 |
|
290 |
# LlamaIndex configuration
|
291 |
+
Settings.llm = HuggingFaceInferenceAPI(
|
292 |
model_name="meta-llama/Meta-Llama-3-8B-Instruct",
|
|
|
|
|
293 |
token=os.getenv("HF_TOKEN"),
|
294 |
max_new_tokens=512,
|
295 |
+
temperature=0.1
|
296 |
)
|
297 |
|
298 |
Settings.embed_model = HuggingFaceEmbedding(
|