forked from lorenlugosch/end-to-end-SLU
-
Notifications
You must be signed in to change notification settings - Fork 0
/
unfreeze_word_layers.cfg
61 lines (54 loc) · 1.47 KB
/
unfreeze_word_layers.cfg
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
[experiment]
seed=1234
folder=experiments/unfreeze_word_layers
[phoneme_module]
use_sincnet=True
fs=16000
cnn_N_filt=80,60,60
cnn_len_filt=401,5,5
cnn_stride=80,1,1
cnn_max_pool_len=2,1,1
cnn_use_laynorm_inp=True
cnn_use_batchnorm_inp=False
cnn_use_laynorm=True,True,True
cnn_use_batchnorm=False,False,False
cnn_act=leaky_relu,leaky_relu,leaky_relu
cnn_drop=0.0,0.0,0.0
phone_rnn_num_hidden=128,128
phone_downsample_len=2,2
phone_downsample_type=avg,avg
phone_rnn_drop=0.5,0.5
phone_rnn_bidirectional=True
[word_module]
word_rnn_num_hidden=128,128
word_downsample_len=2,2
word_downsample_type=avg,avg
word_rnn_drop=0.5,0.5
word_rnn_bidirectional=True
vocabulary_size=10000
[intent_module]
intent_rnn_num_hidden=128
intent_downsample_len=1
intent_downsample_type=none
intent_rnn_drop=0.5
intent_rnn_bidirectional=True
[pretraining]
asr_path=/data/speechdb/librispeech
pretraining_type=2
; 0 - no pre-training, 1 - phoneme loss, 2 - word loss + phoneme loss
pretraining_lr=0.001
pretraining_batch_size=64
pretraining_num_epochs=10
pretraining_length_mean=2.25
pretraining_length_var=1
[training]
slu_path=/data/speechdb/fluent_speech_commands_dataset
unfreezing_type=1
; 0 - no unfreezing, 1 - unfreeze word layers, 2 - unfreeze word layers and phoneme layers
training_lr=0.001
training_batch_size=64
training_num_epochs=20
dataset_subset_percentage=1
train_wording_path=None
; path to .txt file containing phrases to be included during training; if None, uses all phrases
test_wording_path=None