akukkapa commited on
Commit
40c76e9
·
verified ·
1 Parent(s): 5ecce5c

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +9 -2
app.py CHANGED
@@ -2,8 +2,15 @@ from transformers import Blip2Processor, Blip2ForConditionalGeneration, AutoMode
2
  import torch
3
  from PIL import Image, ImageDraw, ImageFont
4
  import gradio as gr
 
5
  import os
6
 
 
 
 
 
 
 
7
  # Initialize environment
8
  os.makedirs("./offload", exist_ok=True)
9
  HF_TOKEN = os.environ.get("HF_TOKEN")
@@ -28,7 +35,7 @@ phi_model = AutoModelForCausalLM.from_pretrained(
28
  trust_remote_code=True,
29
  device_map="auto",
30
  torch_dtype=torch.float16,
31
- load_in_4bit=True,
32
  token=HF_TOKEN
33
  ).eval()
34
  phi_tokenizer = AutoTokenizer.from_pretrained(
@@ -123,4 +130,4 @@ with gr.Blocks(title="AI Meme Generator") as demo:
123
  )
124
 
125
  if __name__ == "__main__":
126
- demo.launch()
 
2
  import torch
3
  from PIL import Image, ImageDraw, ImageFont
4
  import gradio as gr
5
+ import torch
6
  import os
7
 
8
+ # Check CUDA availability
9
+ if not torch.cuda.is_available():
10
+ os.environ["BITSANDBYTES_NOWELCOME"] = "1"
11
+ os.environ["CUDA_VISIBLE_DEVICES"] = ""
12
+ os.environ["LIBRARY_PATH"] = "/usr/local/cuda/lib64/stubs:$LIBRARY_PATH"
13
+
14
  # Initialize environment
15
  os.makedirs("./offload", exist_ok=True)
16
  HF_TOKEN = os.environ.get("HF_TOKEN")
 
35
  trust_remote_code=True,
36
  device_map="auto",
37
  torch_dtype=torch.float16,
38
+ load_in_4bit=torch.cuda.is_available(), # Only use 4bit if CUDA available
39
  token=HF_TOKEN
40
  ).eval()
41
  phi_tokenizer = AutoTokenizer.from_pretrained(
 
130
  )
131
 
132
  if __name__ == "__main__":
133
+ demo.launch(server_name="0.0.0.0", server_port=7860)