jgyasu's picture
Add entire pipeline
060ac52
# This is the official config file.
PECCAVI_TEXT:
Entailment:
task: "text-classification"
model: "ynie/roberta-large-snli_mnli_fever_anli_R1_R2_R3-nli"
Masking:
task: "fill-mask"
tokenizer: "bert-base-uncased"
model: "bert-base-uncased"
# tokenizer: "bert-large-cased-whole-word-masking"
# model: "bert-large-cased-whole-word-masking"
Vocabulary:
tokenizer: "bert-base-uncased"
model: "bert-base-uncased"
# permissible_ratio: 0.5
# tokenizer: "bert-large-cased-whole-word-masking"
# model: "bert-large-cased-whole-word-masking"
permissible_ratio: 1.0
Sampling:
tokenizer: "bert-base-uncased"
model: "bert-base-uncased"
# tokenizer: "bert-large-cased-whole-word-masking"
# model: "bert-large-cased-whole-word-masking"
Metrics:
EuclideanDistance: "sentence-transformers/all-MiniLM-L6-v2"
Distortion: "gpt2"
Detector:
tokenizer: "bert-base-uncased"
model: "bert-base-uncased"
# tokenizer: "bert-large-cased-whole-word-masking"
# model: "bert-large-cased-whole-word-masking"
Paraphrase:
tokenizer: "humarin/chatgpt_paraphraser_on_T5_base"
model: "humarin/chatgpt_paraphraser_on_T5_base"
num_beams: 10
num_beam_groups: 10
num_return_sequences: 10
repetition_penalty: 10.0
diversity_penalty: 3.0
no_repeat_ngram_size: 2
temperature: 0.7
max_length: 64