gemma
Browse files
app.py
CHANGED
@@ -12,7 +12,7 @@ logging.basicConfig(level=logging.INFO)
|
|
12 |
logger = logging.getLogger(__name__)
|
13 |
|
14 |
app = FastAPI()
|
15 |
-
MODEL_NAME = '
|
16 |
|
17 |
@lru_cache()
|
18 |
def get_llm():
|
|
|
12 |
logger = logging.getLogger(__name__)
|
13 |
|
14 |
app = FastAPI()
|
15 |
+
MODEL_NAME = 'gemma2:2b'
|
16 |
|
17 |
@lru_cache()
|
18 |
def get_llm():
|
start.sh
CHANGED
@@ -9,8 +9,8 @@ export CUDA_VISIBLE_DEVICES=0 # Use the first GPU if available
|
|
9 |
ollama serve & # Use GPU 0 if available
|
10 |
|
11 |
# Pull the model if not already present
|
12 |
-
if ! ollama list | grep -q "
|
13 |
-
ollama pull
|
14 |
fi
|
15 |
|
16 |
# Wait for Ollama to start up (use a more robust check)
|
|
|
9 |
ollama serve & # Use GPU 0 if available
|
10 |
|
11 |
# Pull the model if not already present
|
12 |
+
if ! ollama list | grep -q "gemma2:2b"; then
|
13 |
+
ollama pull gemma2:2b
|
14 |
fi
|
15 |
|
16 |
# Wait for Ollama to start up (use a more robust check)
|