pmelnechuk commited on
Commit
9589f70
verified
1 Parent(s): 783a32a

Prueba copiando falcon3

Browse files
Files changed (1) hide show
  1. src/model_load.py +1 -4
src/model_load.py CHANGED
@@ -13,10 +13,7 @@ def load_model():
13
  # Cargar tokenizer y modelo de Hugging Face
14
  tokenizer = AutoTokenizer.from_pretrained(model_name)
15
  model = AutoModelForCausalLM.from_pretrained(model_name,
16
- device_map="auto",
17
- quantization_config=BitsAndBytesConfig(load_in_8bit=True),
18
- torch_dtype="auto",
19
- max_memory=max_memory)
20
 
21
  # Crear pipeline de generaci贸n de texto
22
  text_generation_pipeline = pipeline(
 
13
  # Cargar tokenizer y modelo de Hugging Face
14
  tokenizer = AutoTokenizer.from_pretrained(model_name)
15
  model = AutoModelForCausalLM.from_pretrained(model_name,
16
+ torch_dtype=torch.bfloat16).to("cuda")
 
 
 
17
 
18
  # Crear pipeline de generaci贸n de texto
19
  text_generation_pipeline = pipeline(