DiffSketcher / conf /x /clipascene.yaml
hjc-owo
init repo
966ae59
raw
history blame contribute delete
1.61 kB
method: 'clipascene'
im_name: ""
image_size: 224
u2net_path: "./checkpoint/u2net/u2net.pth"
background_layer: 2 # 2, 8, 11
background_div: 0.35 # 0.35, 0.5, 0.85
background_num_iter: 1501
foreground_layer: 2 # 2, 8, 11
foreground_div: 0.4 # 0.4, 0.5, 0.9
foreground_num_iter: 600 # 1000 if foreground_layer >= 8 else 600
# general
target: null
output_dir: null
path_svg: "none"
mask_object: 0
resize_obj: 0
fix_scale: 0
display_logs: 0
display: 0
test_name: "test"
# training
num_iter: 2001
num_stages: 1
lr_scheduler: 0
lr: 0.0001
color_lr: 0.01
width_lr: 0.0001
color_vars_threshold: 0.0
batch_size: 1
save_step: 100
eval_step: 20
loss_mask: "none"
dilated_mask: 0
mask_cls: None
mask_attention: 0
# strokes params
num_paths: 64
width: 1.5
control_points_per_seg: 4
num_segments: 1
attention_init: 1
saliency_model: "clip"
saliency_clip_model: "ViT-B/32"
xdog_intersec: 1
mask_object_attention: 0
softmax_temp: 0.3
mlp_train: 1
width_optim: 0
mlp_width_weights_path: "none"
mlp_points_weights_path: "none"
switch_loss: 0
gumbel_temp: 0.2
width_loss_weight: 0
width_loss_type: "L1"
optimize_points: 1
load_points_opt_weights: 0
gradnorm: 0
width_weights_lst: ""
ratio_loss: 0
# loss
percep_loss: "none"
perceptual_weight: 0
train_with_clip: 0
clip_weight: 0
start_clip: 0
num_aug_clip: 4
include_target_in_aug: 0
augment_both: 1
augemntations: "affine"
noise_thresh: 0.5
aug_scale_min: 0.7
force_sparse: 0
clip_conv_loss: 1
clip_mask_loss: 0
clip_conv_loss_type: "L2"
clip_conv_layer_weights: "0,0,1.0,1.0,0"
clip_model_name: "ViT-B/32"
clip_fc_loss_weight: 0
clip_text_guide: 0
text_target: None