This repository was archived by the owner on Sep 9, 2022. It is now read-only.
-
Notifications
You must be signed in to change notification settings - Fork 5
Expand file tree
/
Copy pathconfig.py
More file actions
91 lines (77 loc) · 2.88 KB
/
config.py
File metadata and controls
91 lines (77 loc) · 2.88 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
###
# DATA
###
texts_lang = "en"
seg_start = "====="
fake_sent = "fake sent 123: bla one bla day bla whatever."
# pre-trained word embs
vecs_dim = 300
MODEL_PATH="/home/nlp-text/static/aganesh002/text-segmentation"
vocab_path_en = "{}/data/embeddings/en.vocab".format(MODEL_PATH)
vecs_path_en = "{}/data/embeddings/en.vectors".format(MODEL_PATH)
vocab_path_lang = "{}/data/embeddings/hr.vocab".format(MODEL_PATH)
vecs_path_lang = "{}/data/embeddings/hr.vectors".format(MODEL_PATH)
###
# MODEL
###
MODEL_TYPE = "cats" # 'cats' or 'tlt'
MODEL_HOME = "{}/data/models/cats_pretrained".format(MODEL_PATH) # for TLT, use "data/models/tlt_pretrained"
###
# ARCHITECTURE AND TRAINING
###
# general
batch_size = 20
sent_window = 16
sent_stride = 8
perc_blocks_train = 0.35
max_sent_len = 50
positional_embs_size = 10
# transformers
TOK_TRANS_PARAMS = {"num_hidden_layers" : 6,
"hidden_size" : vecs_dim + 2*positional_embs_size,
"num_heads" : 4, "filter_size" : 1024,
"relu_dropout" : 0.1,
"attention_dropout" : 0.1,
"layer_postprocess_dropout" : 0.1,
"allow_ffn_pad" : True
}
SENT_TRANS_PARAMS = {"num_hidden_layers" : 6,
"hidden_size" : vecs_dim + 2*positional_embs_size,
"num_heads" : 4,
"filter_size" : 1024,
"relu_dropout" : 0.1,
"attention_dropout" : 0.1,
"layer_postprocess_dropout" : 0.1,
"allow_ffn_pad" : True
}
TOK_TRANS_PARAMS_PREDICT = {"num_hidden_layers" : 6,
"hidden_size" : vecs_dim + 2*positional_embs_size,
"num_heads" : 4,
"filter_size" : 1024,
"relu_dropout" : 0,
"attention_dropout" : 0,
"layer_postprocess_dropout" : 0,
"allow_ffn_pad" : True
}
SENT_TRANS_PARAMS_PREDICT_CATS = {"num_hidden_layers" : 4,
"hidden_size" : vecs_dim + 2*positional_embs_size,
"num_heads" : 2,
"filter_size" : 1024,
"relu_dropout" : 0,
"attention_dropout" : 0,
"layer_postprocess_dropout" : 0,
"allow_ffn_pad" : True
}
SENT_TRANS_PARAMS_PREDICT_TLT = {"num_hidden_layers" : 6,
"hidden_size" : vecs_dim + 2*positional_embs_size,
"num_heads" : 4,
"filter_size" : 1024,
"relu_dropout" : 0,
"attention_dropout" : 0,
"layer_postprocess_dropout" : 0,
"allow_ffn_pad" : True
}
# training
tfrec_train = ""
EPOCHS = 100
SAVE_CHECKPOINT_STEPS = 500