taybeyond commited on
Commit
dde1432
·
verified ·
1 Parent(s): 6eceadc

Upload 2 files

Browse files
Files changed (2) hide show
  1. app.py +26 -0
  2. requirements.txt +5 -0
app.py ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import gradio as gr
2
+ from transformers import AutoProcessor, AutoModelForCausalLM
3
+ import torch
4
+ from PIL import Image
5
+
6
+ MODEL_ID = "Qwen/Qwen1.5-VL-Chat"
7
+
8
+ processor = AutoProcessor.from_pretrained(MODEL_ID, trust_remote_code=True)
9
+ model = AutoModelForCausalLM.from_pretrained(MODEL_ID, device_map="auto", trust_remote_code=True, torch_dtype=torch.bfloat16)
10
+
11
+ def qwen_vl_chat(image, question):
12
+ inputs = processor(text=question, images=image, return_tensors="pt").to(model.device)
13
+ outputs = model.generate(**inputs, max_new_tokens=256)
14
+ answer = processor.batch_decode(outputs, skip_special_tokens=True)[0]
15
+ return answer
16
+
17
+ demo = gr.Interface(
18
+ fn=qwen_vl_chat,
19
+ inputs=[gr.Image(type="pil"), gr.Textbox(label="请输入问题")],
20
+ outputs="text",
21
+ title="🧠 Qwen1.5-VL 图文问答 Demo",
22
+ description="上传一张图片,问一个问题,模型会给出答案。"
23
+ )
24
+
25
+ if __name__ == "__main__":
26
+ demo.launch()
requirements.txt ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ torch
2
+ transformers==4.38.2
3
+ gradio==4.14.0
4
+ accelerate
5
+ transformers_stream_generator