Nguyen17 commited on
Commit
b85e8d6
·
1 Parent(s): 0050bbf
Files changed (1) hide show
  1. app.py +10 -69
app.py CHANGED
@@ -188,37 +188,13 @@ def get_llm(cfg):
188
 
189
 
190
  def run(text, intensity):
191
- # Configure logging
192
- logging.basicConfig(level=logging.INFO,
193
- format='%(asctime)s - %(name)s - %(levelname)s - %(message)s')
194
- logger = logging.getLogger(__name__)
195
- sample_outputs = "start: "
196
-
197
- try:
198
  # Log the start of the process
199
- logger.info("Starting the process with config file: %s", config_path)
200
-
201
- # Load configuration from the file
202
- config = load_config(config_path)
203
-
204
- # Load necessary components
205
- prompt_template = get_prompt_template()
206
-
207
- # Replace OpenAI embed model and llm with custom ones
208
- reset_settings(config)
209
-
210
- # Get retriever
211
- retriever = get_retriever(config, prompt_template)
212
-
213
- # Load tokenizer and language model
214
- tokenizer = load_tokenizer(config)
215
- language_model = get_llm(config)
216
- prompt = retriever.query(text).response
217
- prompt = tokenizer.bos_token + '[INST] ' + prompt + ' [/INST]'
218
- streamer = TextStreamer(tokenizer, skip_prompt=True)
219
- input_ids = tokenizer([prompt], return_tensors='pt').to(cfg.environment.device)
220
 
221
- sample_outputs = language_model.generate(
222
  **input_ids,
223
  streamer=streamer,
224
  pad_token_id=tokenizer.pad_token_id,
@@ -227,27 +203,6 @@ def run(text, intensity):
227
  temperature=cfg.generation.temperature
228
  )
229
 
230
-
231
- # Start the command line interface
232
- # vistral_chat(config, retriever, tokenizer, language_model)
233
-
234
- # Log successful completion
235
- logger.info("Process completed successfully.")
236
-
237
- except FileNotFoundError as e:
238
- logger.error("Configuration file not found: %s", e)
239
-
240
- except Exception as e:
241
- logger.exception("An error occurred: %s", e)
242
-
243
-
244
-
245
-
246
- # print(20*'---')
247
-
248
- # res="Chatbot Data Mining 2024 \n \n \n"
249
- # max_length=intensity
250
-
251
  return sample_outputs
252
 
253
 
@@ -261,23 +216,6 @@ def vistral_chat():
261
 
262
 
263
 
264
- # while True:
265
- # user_query = input('👨‍🦰 ')
266
- # prompt = retriever.query(user_query).response
267
- # prompt = tokenizer.bos_token + '[INST] ' + prompt + ' [/INST]'
268
- # streamer = TextStreamer(tokenizer, skip_prompt=True)
269
- # input_ids = tokenizer([prompt], return_tensors='pt').to(cfg.environment.device)
270
-
271
- # _ = language_model.generate(
272
- # **input_ids,
273
- # streamer=streamer,
274
- # pad_token_id=tokenizer.pad_token_id,
275
- # max_new_tokens=cfg.generation.max_new_tokens,
276
- # do_sample=cfg.generation.do_sample,
277
- # temperature=cfg.generation.temperature
278
- # )
279
-
280
- # print(20*'---')
281
 
282
 
283
  def main1(config_path):
@@ -285,7 +223,10 @@ def main1(config_path):
285
  logging.basicConfig(level=logging.INFO,
286
  format='%(asctime)s - %(name)s - %(levelname)s - %(message)s')
287
  logger = logging.getLogger(__name__)
288
-
 
 
 
289
  try:
290
  # Log the start of the process
291
  logger.info("Starting the process with config file: %s", config_path)
@@ -307,7 +248,7 @@ def main1(config_path):
307
  language_model = get_llm(config)
308
 
309
  # Start the command line interface
310
- vistral_chat(config, retriever, tokenizer, language_model)
311
 
312
  # Log successful completion
313
  logger.info("Process completed successfully.")
 
188
 
189
 
190
  def run(text, intensity):
 
 
 
 
 
 
 
191
  # Log the start of the process
192
+ prompt = retriever.query(text).response
193
+ prompt = tokenizer.bos_token + '[INST] ' + prompt + ' [/INST]'
194
+ streamer = TextStreamer(tokenizer, skip_prompt=True)
195
+ input_ids = tokenizer([prompt], return_tensors='pt').to(cfg.environment.device)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
196
 
197
+ sample_outputs = language_model.generate(
198
  **input_ids,
199
  streamer=streamer,
200
  pad_token_id=tokenizer.pad_token_id,
 
203
  temperature=cfg.generation.temperature
204
  )
205
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
206
  return sample_outputs
207
 
208
 
 
216
 
217
 
218
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
219
 
220
 
221
  def main1(config_path):
 
223
  logging.basicConfig(level=logging.INFO,
224
  format='%(asctime)s - %(name)s - %(levelname)s - %(message)s')
225
  logger = logging.getLogger(__name__)
226
+ global config = None
227
+ global retriever=None
228
+ global tokenizer=None
229
+ global language_model= None
230
  try:
231
  # Log the start of the process
232
  logger.info("Starting the process with config file: %s", config_path)
 
248
  language_model = get_llm(config)
249
 
250
  # Start the command line interface
251
+ vistral_chat()
252
 
253
  # Log successful completion
254
  logger.info("Process completed successfully.")