Update app.py
Browse files
app.py
CHANGED
@@ -24,6 +24,12 @@ def stop_generation():
|
|
24 |
return "Generation stopped."
|
25 |
|
26 |
def CTXGen(X0, X1, X2, τ, g_num, model_name):
|
|
|
|
|
|
|
|
|
|
|
|
|
27 |
global is_stopped
|
28 |
is_stopped = False
|
29 |
|
@@ -35,7 +41,7 @@ def CTXGen(X0, X1, X2, τ, g_num, model_name):
|
|
35 |
train_seq = train_seqs['Seq'].tolist()
|
36 |
model = torch.load(save_path, map_location=torch.device('cpu'))
|
37 |
model = model.to(device)
|
38 |
-
|
39 |
|
40 |
X3 = "X" * len(X0)
|
41 |
msa_data = pd.read_csv('conoData_C0.csv')
|
@@ -96,6 +102,7 @@ def CTXGen(X0, X1, X2, τ, g_num, model_name):
|
|
96 |
break
|
97 |
|
98 |
seq = [f"{X1}|{X2}|{X3}|{X4}|{X5}|{X6}"]
|
|
|
99 |
vocab_mlm.token_to_idx["X"] = 4
|
100 |
|
101 |
padded_seq, _, _, _ = get_paded_token_idx_gen(vocab_mlm, seq, new_seq)
|
|
|
24 |
return "Generation stopped."
|
25 |
|
26 |
def CTXGen(X0, X1, X2, τ, g_num, model_name):
|
27 |
+
print(X0)
|
28 |
+
print(X1)
|
29 |
+
print(X2)
|
30 |
+
print(τ)
|
31 |
+
print(g_num)
|
32 |
+
print(model_name)
|
33 |
global is_stopped
|
34 |
is_stopped = False
|
35 |
|
|
|
41 |
train_seq = train_seqs['Seq'].tolist()
|
42 |
model = torch.load(save_path, map_location=torch.device('cpu'))
|
43 |
model = model.to(device)
|
44 |
+
print(model)
|
45 |
|
46 |
X3 = "X" * len(X0)
|
47 |
msa_data = pd.read_csv('conoData_C0.csv')
|
|
|
102 |
break
|
103 |
|
104 |
seq = [f"{X1}|{X2}|{X3}|{X4}|{X5}|{X6}"]
|
105 |
+
print(seq)
|
106 |
vocab_mlm.token_to_idx["X"] = 4
|
107 |
|
108 |
padded_seq, _, _, _ = get_paded_token_idx_gen(vocab_mlm, seq, new_seq)
|