Drag2121 commited on
Commit
f98b266
·
1 Parent(s): 4e963c4
Files changed (2) hide show
  1. app.py +1 -1
  2. start.sh +2 -2
app.py CHANGED
@@ -12,7 +12,7 @@ logging.basicConfig(level=logging.INFO)
12
  logger = logging.getLogger(__name__)
13
 
14
  app = FastAPI()
15
- MODEL_NAME = 'phi3:mini'
16
 
17
  @lru_cache()
18
  def get_llm():
 
12
  logger = logging.getLogger(__name__)
13
 
14
  app = FastAPI()
15
+ MODEL_NAME = 'gemma2:2b'
16
 
17
  @lru_cache()
18
  def get_llm():
start.sh CHANGED
@@ -9,8 +9,8 @@ export CUDA_VISIBLE_DEVICES=0 # Use the first GPU if available
9
  ollama serve & # Use GPU 0 if available
10
 
11
  # Pull the model if not already present
12
- if ! ollama list | grep -q "phi3:mini"; then
13
- ollama pull phi3:mini
14
  fi
15
 
16
  # Wait for Ollama to start up (use a more robust check)
 
9
  ollama serve & # Use GPU 0 if available
10
 
11
  # Pull the model if not already present
12
+ if ! ollama list | grep -q "gemma2:2b"; then
13
+ ollama pull gemma2:2b
14
  fi
15
 
16
  # Wait for Ollama to start up (use a more robust check)