forked from FongMunHong/FlowER
-
Notifications
You must be signed in to change notification settings - Fork 0
Expand file tree
/
Copy pathsettings.py
More file actions
87 lines (71 loc) · 2.45 KB
/
settings.py
File metadata and controls
87 lines (71 loc) · 2.45 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
import os
DATA_NAME = os.environ.get("DATA_NAME", "USPTO")
EXP_NAME = os.environ.get("EXP_NAME", "")
SCALE = int(os.environ.get("SCALE", 4)) # train & val
# SCALE = 1 # test
SAMPLE_SIZE = 64 // SCALE
NUM_GPU = int(os.environ.get("NUM_GPUS_PER_NODE", 1))
TRAIN_BATCH_SIZE = int(os.environ.get("TRAIN_BATCH_SIZE", 4096))
VAL_BATCH_SIZE = int(os.environ.get("VAL_BATCH_SIZE", 4096))
TEST_BATCH_SIZE = int(os.environ.get("TEST_BATCH_SIZE", 512 * NUM_GPU * SCALE))
NUM_NODES = int(os.environ.get("NUM_NODES", 1))
ACCUMULATION_COUNT = int(os.environ.get("ACCUMULATION_COUNT", 1))
NUM_WORKERS = int(os.environ.get("NUM_WORKERS", 16))
MODEL_NAME = os.environ.get("MODEL_NAME")
class Args:
# train #
model_name = MODEL_NAME
exp_name = EXP_NAME
train_path = os.environ.get("TRAIN_FILE")
val_path = os.environ.get("VAL_FILE")
test_path = os.environ.get("TEST_FILE")
model_path = os.environ.get("MODEL_PATH")
result_path = os.environ.get("RESULT_PATH")
data_name = f"{DATA_NAME}"
log_file = f"FlowER"
load_from = ""
# resume = True
# load_from = f"{model_path}{MODEL_NAME}"
backend = "nccl"
num_workers = NUM_WORKERS
emb_dim = int(os.environ.get("EMB_DIM"))
enc_num_layers = 12
post_processing_layers = 1
enc_heads = 32
enc_filter_size = 2048
dropout = 0.0
attn_dropout = 0.0
rel_pos = "emb_only"
shared_attention_layer = 0
sigma = float(os.environ.get("SIGMA"))
train_batch_size = (TRAIN_BATCH_SIZE / ACCUMULATION_COUNT / NUM_GPU / NUM_NODES)
val_batch_size = (VAL_BATCH_SIZE / ACCUMULATION_COUNT / NUM_GPU / NUM_NODES)
test_batch_size = TEST_BATCH_SIZE
batch_type = "tokens_sum"
lr = 0.0001
beta1 = 0.9
beta2 = 0.998
eps = 1e-9
weight_decay = 1e-2
warmup_steps = 30000
clip_norm = 200
epoch = int(os.environ.get("EPOCH", 100))
max_steps = 3000000
accumulation_count = ACCUMULATION_COUNT
save_iter = int(os.environ.get("SAVE_ITER", 30000))
log_iter = int(os.environ.get("LOG_ITER", 100))
eval_iter = int(os.environ.get("EVAL_ITER", 30000))
sample_size = SAMPLE_SIZE
rbf_low = 0
rbf_high = float(os.environ.get("RBF_HIGH"))
rbf_gap = float(os.environ.get("RBF_GAP"))
# validation #
# do_validate = True
# steps2validate = ["1050000", "1320000", "1500000", "930000", "1020000"]
# inference #
do_validate = False
# beam-search #
beam_size = 5
nbest = 3
max_depth = 15
chunk_size = 50