Spaces:
Sleeping
Sleeping
Commit
·
2b75dc0
1
Parent(s):
ddf3c8c
Bug fix
Browse files
app.py
CHANGED
@@ -21,7 +21,7 @@ model = AutoModelForCausalLM.from_pretrained(model_id,
|
|
21 |
# model.to("cpu")
|
22 |
|
23 |
def generate_code(prompt, style="Clean & Pythonic"):
|
24 |
-
spinner.update(visible=True)
|
25 |
if style == "Verbose like a 15th-century manuscript":
|
26 |
prompt = "In a manner most detailed, write code that... " + prompt
|
27 |
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
|
@@ -35,7 +35,7 @@ def generate_code(prompt, style="Clean & Pythonic"):
|
|
35 |
num_return_sequences=1,
|
36 |
eos_token_id=tokenizer.eos_token_id
|
37 |
)
|
38 |
-
spinner.update(visible=False)
|
39 |
return tokenizer.decode(outputs[0], skip_special_tokens=True)
|
40 |
|
41 |
# demo = gr.Interface(
|
|
|
21 |
# model.to("cpu")
|
22 |
|
23 |
def generate_code(prompt, style="Clean & Pythonic"):
|
24 |
+
# spinner.update(visible=True)
|
25 |
if style == "Verbose like a 15th-century manuscript":
|
26 |
prompt = "In a manner most detailed, write code that... " + prompt
|
27 |
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
|
|
|
35 |
num_return_sequences=1,
|
36 |
eos_token_id=tokenizer.eos_token_id
|
37 |
)
|
38 |
+
# spinner.update(visible=False)
|
39 |
return tokenizer.decode(outputs[0], skip_special_tokens=True)
|
40 |
|
41 |
# demo = gr.Interface(
|