IST199655 commited on
Commit
16ca936
·
1 Parent(s): f2d1f01

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +2 -3
app.py CHANGED
@@ -5,8 +5,7 @@ from huggingface_hub import InferenceClient
5
  Copied from inference in colab notebook
6
  """
7
 
8
- from transformers import AutoTokenizer , AutoModelForCausalLM , TextIteratorStreamer
9
- import torch
10
  from threading import Thread
11
 
12
  # Load model and tokenizer globally to avoid reloading for every request
@@ -17,7 +16,7 @@ model_path = "Mat17892/t5small_enfr_opus"
17
  tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=True, legacy=False)
18
 
19
  # Load the base model (e.g., LLaMA)
20
- base_model = AutoModelForCausalLM.from_pretrained(base_model)
21
 
22
  # Load LoRA adapter
23
  from peft import PeftModel
 
5
  Copied from inference in colab notebook
6
  """
7
 
8
+ from transformers import AutoTokenizer , AutoModelForSeq2SeqLM , TextIteratorStreamer
 
9
  from threading import Thread
10
 
11
  # Load model and tokenizer globally to avoid reloading for every request
 
16
  tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=True, legacy=False)
17
 
18
  # Load the base model (e.g., LLaMA)
19
+ base_model = AutoModelForSeq2SeqLM.from_pretrained(base_model)
20
 
21
  # Load LoRA adapter
22
  from peft import PeftModel