IST199655 commited on
Commit
64cfbfa
·
1 Parent(s): 3d5b038
Files changed (1) hide show
  1. app.py +2 -2
app.py CHANGED
@@ -5,14 +5,14 @@ from huggingface_hub import InferenceClient
5
  Copied from inference in colab notebook
6
  """
7
 
8
- from transformers import LlamaForCausalLM, LlamaTokenizer
9
  import torch
10
 
11
  # Load model and tokenizer globally to avoid reloading for every request
12
  model_path = "llama_lora_model_1"
13
 
14
  # Load tokenizer
15
- tokenizer = LlamaTokenizer.from_pretrained(model_path)
16
 
17
  # Load model
18
  model = LlamaForCausalLM.from_pretrained(
 
5
  Copied from inference in colab notebook
6
  """
7
 
8
+ from transformers import LlamaForCausalLM, AutoTokenizer
9
  import torch
10
 
11
  # Load model and tokenizer globally to avoid reloading for every request
12
  model_path = "llama_lora_model_1"
13
 
14
  # Load tokenizer
15
+ tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=True, legacy=False)
16
 
17
  # Load model
18
  model = LlamaForCausalLM.from_pretrained(