ericwithpy commited on
Commit
7e55c71
·
verified ·
1 Parent(s): 911b23c

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +8 -1
app.py CHANGED
@@ -4,7 +4,14 @@ from transformers import AutoModelForCausalLM, AutoTokenizer
4
  device = "cuda" if torch.cuda.is_available() else "cpu" # Automatically detect GPU or CPU
5
  model_name = "tanusrich/Mental_Health_Chatbot"
6
 
7
- model = AutoModelForCausalLM.from_pretrained(model_name)
 
 
 
 
 
 
 
8
 
9
  tokenizer = AutoTokenizer.from_pretrained(model_name)
10
  '''
 
4
  device = "cuda" if torch.cuda.is_available() else "cpu" # Automatically detect GPU or CPU
5
  model_name = "tanusrich/Mental_Health_Chatbot"
6
 
7
+ model = AutoModelForCausalLM.from_pretrained(
8
+ model_name,
9
+ torch_dtype=torch.float16, # Reduce memory usage
10
+ device_map="auto", # Automatically assigns to GPU if available
11
+ low_cpu_mem_usage=True,
12
+ max_memory={0: "4GiB", "cpu": "6GiB"}, # Optimize CPU memory
13
+ offload_folder=None
14
+ )
15
 
16
  tokenizer = AutoTokenizer.from_pretrained(model_name)
17
  '''