Nac31 commited on
Commit
905ea86
·
1 Parent(s): 50a2b44

Update app

Browse files
Files changed (1) hide show
  1. app.py +10 -3
app.py CHANGED
@@ -5,6 +5,7 @@ import os
5
  from dotenv import load_dotenv
6
  from huggingface_hub import login
7
  from transformers import BitsAndBytesConfig
 
8
 
9
  load_dotenv()
10
 
@@ -32,13 +33,19 @@ model = AutoModelForCausalLM.from_pretrained(
32
  )
33
  pipe = pipeline("text-generation", model=model, tokenizer=tokenizer)
34
 
35
- def generate_response(message, temperature=0.7, max_new_tokens=500):
36
  try:
37
  response = pipe(
38
  message,
39
  temperature=temperature,
40
  max_new_tokens=max_new_tokens,
41
- do_sample=True
 
 
 
 
 
 
42
  )
43
  return response[0]['generated_text']
44
  except Exception as e:
@@ -50,7 +57,7 @@ demo = gr.Interface(
50
  inputs=[
51
  gr.Textbox(label="Votre message", placeholder="Entrez votre message ici..."),
52
  gr.Slider(minimum=0.1, maximum=1.0, value=0.7, label="Température"),
53
- gr.Slider(minimum=10, maximum=2000, value=500, step=10, label="Nombre de tokens")
54
  ],
55
  outputs=gr.Textbox(label="Réponse"),
56
  title="Chat avec Sacha-Mistral",
 
5
  from dotenv import load_dotenv
6
  from huggingface_hub import login
7
  from transformers import BitsAndBytesConfig
8
+ from functools import lru_cache
9
 
10
  load_dotenv()
11
 
 
33
  )
34
  pipe = pipeline("text-generation", model=model, tokenizer=tokenizer)
35
 
36
+ def generate_response(message, temperature=0.7, max_new_tokens=256):
37
  try:
38
  response = pipe(
39
  message,
40
  temperature=temperature,
41
  max_new_tokens=max_new_tokens,
42
+ do_sample=True,
43
+ top_k=50,
44
+ top_p=0.9,
45
+ repetition_penalty=1.2,
46
+ pad_token_id=tokenizer.pad_token_id,
47
+ eos_token_id=tokenizer.eos_token_id,
48
+ batch_size=1
49
  )
50
  return response[0]['generated_text']
51
  except Exception as e:
 
57
  inputs=[
58
  gr.Textbox(label="Votre message", placeholder="Entrez votre message ici..."),
59
  gr.Slider(minimum=0.1, maximum=1.0, value=0.7, label="Température"),
60
+ gr.Slider(minimum=10, maximum=3000, value=256, step=10, label="Nombre de tokens")
61
  ],
62
  outputs=gr.Textbox(label="Réponse"),
63
  title="Chat avec Sacha-Mistral",