Spaces:
Running
Running
File size: 1,605 Bytes
966ae59 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 |
method: 'clipascene'
im_name: ""
image_size: 224
u2net_path: "./checkpoint/u2net/u2net.pth"
background_layer: 2 # 2, 8, 11
background_div: 0.35 # 0.35, 0.5, 0.85
background_num_iter: 1501
foreground_layer: 2 # 2, 8, 11
foreground_div: 0.4 # 0.4, 0.5, 0.9
foreground_num_iter: 600 # 1000 if foreground_layer >= 8 else 600
# general
target: null
output_dir: null
path_svg: "none"
mask_object: 0
resize_obj: 0
fix_scale: 0
display_logs: 0
display: 0
test_name: "test"
# training
num_iter: 2001
num_stages: 1
lr_scheduler: 0
lr: 0.0001
color_lr: 0.01
width_lr: 0.0001
color_vars_threshold: 0.0
batch_size: 1
save_step: 100
eval_step: 20
loss_mask: "none"
dilated_mask: 0
mask_cls: None
mask_attention: 0
# strokes params
num_paths: 64
width: 1.5
control_points_per_seg: 4
num_segments: 1
attention_init: 1
saliency_model: "clip"
saliency_clip_model: "ViT-B/32"
xdog_intersec: 1
mask_object_attention: 0
softmax_temp: 0.3
mlp_train: 1
width_optim: 0
mlp_width_weights_path: "none"
mlp_points_weights_path: "none"
switch_loss: 0
gumbel_temp: 0.2
width_loss_weight: 0
width_loss_type: "L1"
optimize_points: 1
load_points_opt_weights: 0
gradnorm: 0
width_weights_lst: ""
ratio_loss: 0
# loss
percep_loss: "none"
perceptual_weight: 0
train_with_clip: 0
clip_weight: 0
start_clip: 0
num_aug_clip: 4
include_target_in_aug: 0
augment_both: 1
augemntations: "affine"
noise_thresh: 0.5
aug_scale_min: 0.7
force_sparse: 0
clip_conv_loss: 1
clip_mask_loss: 0
clip_conv_loss_type: "L2"
clip_conv_layer_weights: "0,0,1.0,1.0,0"
clip_model_name: "ViT-B/32"
clip_fc_loss_weight: 0
clip_text_guide: 0
text_target: None
|