prajdabre commited on
Commit
bd34b3e
·
1 Parent(s): cdd0e52

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +15 -2
app.py CHANGED
@@ -19,10 +19,23 @@ def generate(input, slang, tlang):
19
  tlang = CREOLE[tlang]
20
  inp = tokenizer(input.strip() + " </s> <2" + slang + ">",
21
  add_special_tokens=False, return_tensors="pt", padding=True).input_ids
22
- model_output = model.generate(inp, use_cache=True, num_beams=1, max_length=int(2*len(inp[0])), min_length=1, early_stopping=True, pad_token_id=pad_id,
 
23
  bos_token_id=bos_id, eos_token_id=eos_id, decoder_start_token_id=tokenizer._convert_token_to_id_with_added_voc("<2"+tlang+">"))
24
- decoded_output = tokenizer.decode(
25
  model_output[0], skip_special_tokens=True, clean_up_tokenization_spaces=False)
 
 
 
 
 
 
 
 
 
 
 
 
26
 
27
  return decoded_output
28
 
 
19
  tlang = CREOLE[tlang]
20
  inp = tokenizer(input.strip() + " </s> <2" + slang + ">",
21
  add_special_tokens=False, return_tensors="pt", padding=True).input_ids
22
+ if (slang != "eng" and tlang == "eng") or (slang == "eng" and tlang != "eng"):
23
+ model_output = model.generate(inp, use_cache=True, num_beams=1, max_length=int(2*len(inp[0])), min_length=1, early_stopping=True, pad_token_id=pad_id,
24
  bos_token_id=bos_id, eos_token_id=eos_id, decoder_start_token_id=tokenizer._convert_token_to_id_with_added_voc("<2"+tlang+">"))
25
+ decoded_output = tokenizer.decode(
26
  model_output[0], skip_special_tokens=True, clean_up_tokenization_spaces=False)
27
+ elif slang != tlang:
28
+ model_output = model.generate(inp, use_cache=True, num_beams=1, max_length=int(2*len(inp[0])), min_length=1, early_stopping=True, pad_token_id=pad_id,
29
+ bos_token_id=bos_id, eos_token_id=eos_id, decoder_start_token_id=tokenizer._convert_token_to_id_with_added_voc("<2eng>"))
30
+ decoded_output = tokenizer.decode(
31
+ model_output[0], skip_special_tokens=True, clean_up_tokenization_spaces=False)
32
+ inp = tokenizer(decoded_output + " </s> <2eng>",
33
+ add_special_tokens=False, return_tensors="pt", padding=True).input_ids
34
+ model_output = model.generate(inp, use_cache=True, num_beams=1, max_length=int(2*len(inp[0])), min_length=1, early_stopping=True, pad_token_id=pad_id,
35
+ bos_token_id=bos_id, eos_token_id=eos_id, decoder_start_token_id=tokenizer._convert_token_to_id_with_added_voc("<2"+tlang+">"))
36
+ decoded_output = tokenizer.decode(
37
+ model_output[0], skip_special_tokens=True, clean_up_tokenization_spaces=False)
38
+
39
 
40
  return decoded_output
41