ethanlshen commited on
Commit
37a92ca
·
verified ·
1 Parent(s): 2b84629

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +2 -1
app.py CHANGED
@@ -46,7 +46,7 @@ n_token_sample = params["n_token_sample"]
46
  i_weights = params["i_weights"]
47
  i_length = params["i_length"]
48
  # Load main model
49
- model = load_models().to('cuda')
50
  tokenizer = Tokenizer(f'{weight_path}/tokenizer.model')
51
  # Create ngram models
52
  ngrams = make_models("ckpts-200k", bigram=True, trigram=True, fourgram=True, fivegram=True, sixgram=True, sevengram=False)
@@ -66,6 +66,7 @@ def decode(tokenizer, encoding):
66
 
67
  @spaces.GPU
68
  def update_options(input, num_tokens):
 
69
  tokenized_prompts = tokenizer.encode([input], True, False)
70
  alive_gens, _ = model.sup_generate(prompt_tokens=tokenized_prompts,
71
  smoothing="geom",
 
46
  i_weights = params["i_weights"]
47
  i_length = params["i_length"]
48
  # Load main model
49
+ model = load_models()
50
  tokenizer = Tokenizer(f'{weight_path}/tokenizer.model')
51
  # Create ngram models
52
  ngrams = make_models("ckpts-200k", bigram=True, trigram=True, fourgram=True, fivegram=True, sixgram=True, sevengram=False)
 
66
 
67
  @spaces.GPU
68
  def update_options(input, num_tokens):
69
+ model.to("cuda")
70
  tokenized_prompts = tokenizer.encode([input], True, False)
71
  alive_gens, _ = model.sup_generate(prompt_tokens=tokenized_prompts,
72
  smoothing="geom",