Panyun commited on
Commit
dda1778
·
verified ·
1 Parent(s): 0626db3

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +8 -6
app.py CHANGED
@@ -1,13 +1,15 @@
1
  import gradio as gr
2
  from transformers import AutoModelForCausalLM, AutoTokenizer
3
 
4
- # 加载模型(使用量化版本节省显存)
5
- model_name = "Qwen/Qwen2.5-7B-Instruct-GPTQ-Int8"
6
- tokenizer = AutoTokenizer.from_pretrained(model_name)
7
  model = AutoModelForCausalLM.from_pretrained(
8
- model_name,
9
- device_map="auto", # 自动分配GPU/CPU
10
- torch_dtype="auto"
 
 
11
  )
12
 
13
  # 定义生成函数
 
1
  import gradio as gr
2
  from transformers import AutoModelForCausalLM, AutoTokenizer
3
 
4
+ from huggingface_hub import login
5
+ login(token="你的Token") # 添加在模型加载前
6
+
7
  model = AutoModelForCausalLM.from_pretrained(
8
+ "Qwen/Qwen2.5-7B-Instruct-AWQ",
9
+ device_map="auto",
10
+ use_auth_token=True, # 启用认证
11
+ resume_download=True, # 断点续传
12
+ local_files_only=False
13
  )
14
 
15
  # 定义生成函数