Nguyen17 commited on
Commit
0050bbf
·
1 Parent(s): 7bedafb
Files changed (1) hide show
  1. app.py +50 -11
app.py CHANGED
@@ -188,12 +188,37 @@ def get_llm(cfg):
188
 
189
 
190
  def run(text, intensity):
191
- prompt = retriever.query(text).response
192
- prompt = tokenizer.bos_token + '[INST] ' + prompt + ' [/INST]'
193
- streamer = TextStreamer(tokenizer, skip_prompt=True)
194
- input_ids = tokenizer([prompt], return_tensors='pt').to(cfg.environment.device)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
195
 
196
- _ = language_model.generate(
197
  **input_ids,
198
  streamer=streamer,
199
  pad_token_id=tokenizer.pad_token_id,
@@ -202,19 +227,33 @@ def run(text, intensity):
202
  temperature=cfg.generation.temperature
203
  )
204
 
205
- # print(20*'---')
 
 
 
 
 
 
 
 
206
 
 
 
207
 
208
 
209
- res="Chatbot Data Mining 2024 \n \n \n"
210
- max_length=intensity
211
 
212
- return _
 
 
 
 
 
 
213
 
214
 
215
- def vistral_chat(cfg, retriever, tokenizer, language_model):
216
  demo = gr.Interface(fn=run,
217
- inputs=[gr.Textbox(label="Nhập vào nội dung input",value="Con đường xưa em đi"),gr.Slider(label="Độ dài output muốn tạo ra", value=20, minimum=10, maximum=100, step=2)],
218
  outputs=gr.Textbox(label="Output"), # <-- Number of output components: 1
219
  )
220
 
 
188
 
189
 
190
  def run(text, intensity):
191
+ # Configure logging
192
+ logging.basicConfig(level=logging.INFO,
193
+ format='%(asctime)s - %(name)s - %(levelname)s - %(message)s')
194
+ logger = logging.getLogger(__name__)
195
+ sample_outputs = "start: "
196
+
197
+ try:
198
+ # Log the start of the process
199
+ logger.info("Starting the process with config file: %s", config_path)
200
+
201
+ # Load configuration from the file
202
+ config = load_config(config_path)
203
+
204
+ # Load necessary components
205
+ prompt_template = get_prompt_template()
206
+
207
+ # Replace OpenAI embed model and llm with custom ones
208
+ reset_settings(config)
209
+
210
+ # Get retriever
211
+ retriever = get_retriever(config, prompt_template)
212
+
213
+ # Load tokenizer and language model
214
+ tokenizer = load_tokenizer(config)
215
+ language_model = get_llm(config)
216
+ prompt = retriever.query(text).response
217
+ prompt = tokenizer.bos_token + '[INST] ' + prompt + ' [/INST]'
218
+ streamer = TextStreamer(tokenizer, skip_prompt=True)
219
+ input_ids = tokenizer([prompt], return_tensors='pt').to(cfg.environment.device)
220
 
221
+ sample_outputs = language_model.generate(
222
  **input_ids,
223
  streamer=streamer,
224
  pad_token_id=tokenizer.pad_token_id,
 
227
  temperature=cfg.generation.temperature
228
  )
229
 
230
+
231
+ # Start the command line interface
232
+ # vistral_chat(config, retriever, tokenizer, language_model)
233
+
234
+ # Log successful completion
235
+ logger.info("Process completed successfully.")
236
+
237
+ except FileNotFoundError as e:
238
+ logger.error("Configuration file not found: %s", e)
239
 
240
+ except Exception as e:
241
+ logger.exception("An error occurred: %s", e)
242
 
243
 
 
 
244
 
245
+
246
+ # print(20*'---')
247
+
248
+ # res="Chatbot Data Mining 2024 \n \n \n"
249
+ # max_length=intensity
250
+
251
+ return sample_outputs
252
 
253
 
254
+ def vistral_chat():
255
  demo = gr.Interface(fn=run,
256
+ inputs=[gr.Textbox(label="Nhập vào nội dung input",value="Con đường xưa em đi"),gr.Slider(label="Độ dài output muốn tạo ra", value=20, minimum=10, maximum=100, step=2),],
257
  outputs=gr.Textbox(label="Output"), # <-- Number of output components: 1
258
  )
259