jedeland commited on
Commit
66d8dc1
·
1 Parent(s): 4150564
Files changed (2) hide show
  1. app.py +13 -9
  2. requirements.txt +2 -1
app.py CHANGED
@@ -1,10 +1,17 @@
1
  import gradio as gr
2
 
3
  # Load model directly
4
- from transformers import AutoModel, AutoTokenizer
5
 
6
- model = AutoModel.from_pretrained("ID2223JR/gguf_model")
7
- tokenizer = AutoTokenizer.from_pretrained("ID2223JR/gguf_model")
 
 
 
 
 
 
 
8
 
9
  # Data storage
10
  ingredients_list = []
@@ -38,12 +45,9 @@ def submit_to_model():
38
  ingredients_list
39
  )
40
 
41
- # Tokenize and pass the prompt to the model
42
- inputs = tokenizer(prompt, return_tensors="pt")
43
- outputs = model.generate(**inputs, max_new_tokens=100)
44
-
45
- # Decode the model output
46
- response = tokenizer.decode(outputs[0], skip_special_tokens=True)
47
  return response
48
 
49
 
 
1
  import gradio as gr
2
 
3
  # Load model directly
4
+ # from transformers import AutoModel, AutoTokenizer
5
 
6
+ # model = AutoModel.from_pretrained("ID2223JR/gguf_model")
7
+ # tokenizer = AutoTokenizer.from_pretrained("ID2223JR/gguf_model")
8
+
9
+ from llama_cpp import Llama
10
+
11
+ llm = Llama.from_pretrained(
12
+ repo_id="ID2223JR/gguf_model",
13
+ filename="GGUF_FILE",
14
+ )
15
 
16
  # Data storage
17
  ingredients_list = []
 
45
  ingredients_list
46
  )
47
 
48
+ response = llm.create_chat_completion(
49
+ messages=prompt,
50
+ )
 
 
 
51
  return response
52
 
53
 
requirements.txt CHANGED
@@ -1,4 +1,5 @@
1
  gradio==5.1.0
2
  llama-cpp-python==0.2.24
3
  transformers==4.46.3
4
- torch==2.5.1
 
 
1
  gradio==5.1.0
2
  llama-cpp-python==0.2.24
3
  transformers==4.46.3
4
+ torch==2.5.1
5
+ hugingface-hub==0.25.2