FlawedLLM commited on
Commit
e74adc0
·
verified ·
1 Parent(s): eaaf231

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +15 -8
app.py CHANGED
@@ -1,7 +1,7 @@
1
  import re
2
  import spaces
3
  import gradio as gr
4
- from transformers import AutoTokenizer, AutoModel
5
  import torch
6
  # from peft import PeftModel, PeftConfig
7
 
@@ -36,13 +36,20 @@ import torch
36
 
37
  # model = AutoModel.from_pretrained("FlawedLLM/Bhashini", load_in_4bit=True, device_map='auto')
38
  # I highly do NOT suggest - use Unsloth if possible
39
- from peft import AutoPeftModelForCausalLM
40
- from transformers import AutoTokenizer
41
- model = AutoPeftModelForCausalLM.from_pretrained(
42
- "FlawedLLM/Bhashini", # YOUR MODEL YOU USED FOR TRAINING
43
- load_in_4bit = True,
44
- )
45
- tokenizer = AutoTokenizer.from_pretrained("FlawedLLM/Bhashini")
 
 
 
 
 
 
 
46
 
47
  @spaces.GPU(duration=300)
48
  def chunk_it(input_command):
 
1
  import re
2
  import spaces
3
  import gradio as gr
4
+ from transformers import AutoTokenizer, AutoModel, BitsAndBytesConfig
5
  import torch
6
  # from peft import PeftModel, PeftConfig
7
 
 
36
 
37
  # model = AutoModel.from_pretrained("FlawedLLM/Bhashini", load_in_4bit=True, device_map='auto')
38
  # I highly do NOT suggest - use Unsloth if possible
39
+ # from peft import AutoPeftModelForCausalLM
40
+ # from transformers import AutoTokenizer
41
+ # model = AutoPeftModelForCausalLM.from_pretrained(
42
+ # "FlawedLLM/Bhashini", # YOUR MODEL YOU USED FOR TRAINING
43
+ # load_in_4bit = True,
44
+ # )
45
+ # tokenizer = AutoTokenizer.from_pretrained("FlawedLLM/Bhashini")
46
+ from unsloth import FastLanguageModel
47
+ model, tokenizer = FastLanguageModel.from_pretrained(
48
+ model_name = "BhashiniLLama", # YOUR MODEL YOU USED FOR TRAINING
49
+ max_seq_length = max_seq_length,
50
+ dtype = dtype,
51
+ load_in_4bit = load_in_4bit,)
52
+ FastLanguageModel.for_inference(model)
53
 
54
  @spaces.GPU(duration=300)
55
  def chunk_it(input_command):