ThongCoding commited on
Commit
da4e0ac
·
verified ·
1 Parent(s): ff659ba

Upload 4 files

Browse files
Files changed (4) hide show
  1. .gitattributes +35 -35
  2. app.py +36 -25
  3. model/temp.txt +0 -0
  4. requirements.txt +2 -4
.gitattributes CHANGED
@@ -1,35 +1,35 @@
1
- *.7z filter=lfs diff=lfs merge=lfs -text
2
- *.arrow filter=lfs diff=lfs merge=lfs -text
3
- *.bin filter=lfs diff=lfs merge=lfs -text
4
- *.bz2 filter=lfs diff=lfs merge=lfs -text
5
- *.ckpt filter=lfs diff=lfs merge=lfs -text
6
- *.ftz filter=lfs diff=lfs merge=lfs -text
7
- *.gz filter=lfs diff=lfs merge=lfs -text
8
- *.h5 filter=lfs diff=lfs merge=lfs -text
9
- *.joblib filter=lfs diff=lfs merge=lfs -text
10
- *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
- *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
- *.model filter=lfs diff=lfs merge=lfs -text
13
- *.msgpack filter=lfs diff=lfs merge=lfs -text
14
- *.npy filter=lfs diff=lfs merge=lfs -text
15
- *.npz filter=lfs diff=lfs merge=lfs -text
16
- *.onnx filter=lfs diff=lfs merge=lfs -text
17
- *.ot filter=lfs diff=lfs merge=lfs -text
18
- *.parquet filter=lfs diff=lfs merge=lfs -text
19
- *.pb filter=lfs diff=lfs merge=lfs -text
20
- *.pickle filter=lfs diff=lfs merge=lfs -text
21
- *.pkl filter=lfs diff=lfs merge=lfs -text
22
- *.pt filter=lfs diff=lfs merge=lfs -text
23
- *.pth filter=lfs diff=lfs merge=lfs -text
24
- *.rar filter=lfs diff=lfs merge=lfs -text
25
- *.safetensors filter=lfs diff=lfs merge=lfs -text
26
- saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
- *.tar.* filter=lfs diff=lfs merge=lfs -text
28
- *.tar filter=lfs diff=lfs merge=lfs -text
29
- *.tflite filter=lfs diff=lfs merge=lfs -text
30
- *.tgz filter=lfs diff=lfs merge=lfs -text
31
- *.wasm filter=lfs diff=lfs merge=lfs -text
32
- *.xz filter=lfs diff=lfs merge=lfs -text
33
- *.zip filter=lfs diff=lfs merge=lfs -text
34
- *.zst filter=lfs diff=lfs merge=lfs -text
35
- *tfevents* filter=lfs diff=lfs merge=lfs -text
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
6
+ *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gz filter=lfs diff=lfs merge=lfs -text
8
+ *.h5 filter=lfs diff=lfs merge=lfs -text
9
+ *.joblib filter=lfs diff=lfs merge=lfs -text
10
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
+ *.model filter=lfs diff=lfs merge=lfs -text
13
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
14
+ *.npy filter=lfs diff=lfs merge=lfs -text
15
+ *.npz filter=lfs diff=lfs merge=lfs -text
16
+ *.onnx filter=lfs diff=lfs merge=lfs -text
17
+ *.ot filter=lfs diff=lfs merge=lfs -text
18
+ *.parquet filter=lfs diff=lfs merge=lfs -text
19
+ *.pb filter=lfs diff=lfs merge=lfs -text
20
+ *.pickle filter=lfs diff=lfs merge=lfs -text
21
+ *.pkl filter=lfs diff=lfs merge=lfs -text
22
+ *.pt filter=lfs diff=lfs merge=lfs -text
23
+ *.pth filter=lfs diff=lfs merge=lfs -text
24
+ *.rar filter=lfs diff=lfs merge=lfs -text
25
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
26
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
28
+ *.tar filter=lfs diff=lfs merge=lfs -text
29
+ *.tflite filter=lfs diff=lfs merge=lfs -text
30
+ *.tgz filter=lfs diff=lfs merge=lfs -text
31
+ *.wasm filter=lfs diff=lfs merge=lfs -text
32
+ *.xz filter=lfs diff=lfs merge=lfs -text
33
+ *.zip filter=lfs diff=lfs merge=lfs -text
34
+ *.zst filter=lfs diff=lfs merge=lfs -text
35
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
app.py CHANGED
@@ -1,25 +1,36 @@
1
- import os
2
- import gradio as gr
3
- from transformers import AutoModelForCausalLM, AutoTokenizer
4
-
5
- # Hàm để sử dụng mô hình trả lời câu hỏi
6
- def chatbot(input_text):
7
- # Đảm bảo bạn đã load mô hình và tokenizer
8
- model_name = "Viet-Mistral/Vistral-7B-Chat" # Hoặc tên mô hình bạn đang sử dụng
9
- hf_token = os.getenv("HF_AUTH_TOKEN")
10
- model = AutoModelForCausalLM.from_pretrained(model_name, token=hf_token)
11
- tokenizer = AutoTokenizer.from_pretrained(model_name, token=hf_token)
12
-
13
- # Xử đầu vào và tạo ra câu trả lời
14
- inputs = tokenizer(input_text, return_tensors="pt")
15
- outputs = model.generate(inputs["input_ids"])
16
- response = tokenizer.decode(outputs[0], skip_special_tokens=True)
17
-
18
- # Trả về câu trả lời
19
- return response
20
-
21
- # Tạo giao diện Gradio
22
- iface = gr.Interface(fn=chatbot, inputs="text", outputs="text", live=True)
23
-
24
- # Khởi động giao diện
25
- iface.launch()
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import os
2
+ import subprocess
3
+ import gradio as gr
4
+ from llama_cpp import Llama
5
+
6
+ MODEL_URL = "https://drive.google.com/uc?id=1aYzH-tUYl8LT-0KdO1NK9mA1w37CkM7s&export=download"
7
+ MODEL_PATH = "model/phogpt-4b-chat-q4_K_M.gguf"
8
+
9
+ # Tải model nếu chưa có
10
+ def download_model():
11
+ if not os.path.exists(MODEL_PATH):
12
+ os.makedirs("model", exist_ok=True)
13
+ print("⬇️ Tải model từ Google Drive...")
14
+ subprocess.run([
15
+ "wget", "-O", MODEL_PATH, MODEL_URL
16
+ ])
17
+ print("✅ Tải xong!")
18
+
19
+ download_model()
20
+
21
+ # Load model
22
+ llm = Llama(
23
+ model_path=MODEL_PATH,
24
+ n_ctx=2048,
25
+ n_threads=4,
26
+ n_gpu_layers=35,
27
+ verbose=True
28
+ )
29
+
30
+ def chat_fn(message, history):
31
+ prompt = f"<|user|>\n{message}\n<|assistant|>\n"
32
+ response = llm(prompt, max_tokens=512, stop=["<|user|>"], echo=False)
33
+ output = response["choices"][0]["text"].strip()
34
+ return output
35
+
36
+ gr.ChatInterface(chat_fn, title="🧠 PhoGPT-4B Chatbot (Tiếng Việt)").launch()
model/temp.txt ADDED
File without changes
requirements.txt CHANGED
@@ -1,4 +1,2 @@
1
- torch
2
- transformers
3
- gradio
4
- accelerate
 
1
+ gradio
2
+ llama-cpp-python>=0.2.38