-
Notifications
You must be signed in to change notification settings - Fork 24
/
Copy pathconfig_synthetic.yml
58 lines (39 loc) · 1.31 KB
/
config_synthetic.yml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
comment = "Write some meaningful comments that can be used in future to identify the intents of running this experiment."
[train]
model_path = temp_{}
# Whether to load a pretrained model or not
preload_model = True
# path to the pre-trained model
pretrain_model_path = "trained_models/mix_len_cr_percent_equal_batch_3_13_prop_100_hdsz_2048_batch_2000_optim_adam_lr_0.001_wd_0.0_enocoderdrop_0.0_drop_0.2_step_mix_mode_12.pth"
# Proportion of the dataset to be used while training, use 100
proportion = 100
# Number of epochs to run during training
num_epochs = 400
# batch size, based on the GPU memory
batch_size = 100
# hidden size of RNN
hidden_size = 2048
# Output feature size from CNN
input_size = 2048
# Number of batches to be collected before the network update
num_traj = 1
# Canvas shape, keep it 64
canvas_shape = 64
# Learning rate
lr = 0.001
# Optimizer: RL training -> "sgd" or supervised training -> "adam"
optim = adam
# Epsilon for the RL training, not applicable in Supervised training
epsilon = 1
# l2 Weight decay
weight_decay = 0.0
# dropout for Decoder network
dropout = 0.2
# Encoder dropout
encoder_drop = 0.2
# Whether to schedule the learning rate or not
lr_sch = True
# Number of epochs to wait before decaying the learning rate.
patience = 8
# Mode of training, 1: supervised, 2: RL
mode = 1