forked from kaldi-asr/kaldi
-
Notifications
You must be signed in to change notification settings - Fork 3
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
- Loading branch information
1 parent
2668098
commit 63a64e7
Showing
90 changed files
with
8,119 additions
and
498 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
|
@@ -20,22 +20,25 @@ set -e -o pipefail | |
|
||
# First the options that are passed through to run_ivector_common.sh | ||
# (some of which are also used in this script directly). | ||
stage=17 | ||
stage=0 | ||
nj=30 | ||
dropout_schedule='0,[email protected],[email protected],0' | ||
train_set=train_cleaned | ||
gmm=tri5_cleaned # the gmm for the target data | ||
langdir=data/langp/tri5_ali | ||
num_threads_ubm=12 | ||
nnet3_affix=_cleaned # cleanup affix for nnet3 and chain dirs, e.g. _cleaned | ||
num_epochs=4 | ||
extractor= # Use a pre-trained i-vector extractor | ||
|
||
# The rest are configs specific to this script. Most of the parameters | ||
# are just hardcoded at this level, in the commands below. | ||
train_stage=-10 | ||
tree_affix= # affix for tree directory, e.g. "a" or "b", in case we change the configuration. | ||
tdnn_affix="_bab7" #affix for TDNN directory, e.g. "a" or "b", in case we change the configuration. | ||
common_egs_dir=exp/chain_cleaned/tdnn_lstm_sp/egs # you can set this to use previously dumped egs. | ||
common_egs_dir= # you can set this to use previously dumped egs. | ||
chunk_width=150,120,90,75 | ||
chunk_left_context=40 | ||
|
||
# End configuration section. | ||
echo "$0 $@" # Print the command line for logging | ||
|
@@ -58,7 +61,8 @@ local/chain/run_ivector_common.sh --stage $stage \ | |
--train-set $train_set \ | ||
--gmm $gmm \ | ||
--num-threads-ubm $num_threads_ubm \ | ||
--nnet3-affix "$nnet3_affix" | ||
--nnet3-affix "$nnet3_affix" \ | ||
--extractor "$extractor" | ||
|
||
|
||
gmm_dir=exp/$gmm | ||
|
@@ -194,11 +198,15 @@ if [ $stage -le 18 ]; then | |
--chain.apply-deriv-weights false \ | ||
--chain.lm-opts="--num-extra-lm-states=2000" \ | ||
--egs.dir "$common_egs_dir" \ | ||
--egs.chunk-left-context $chunk_left_context \ | ||
--egs.chunk-right-context 0 \ | ||
--egs.chunk-left-context-initial 0 \ | ||
--egs.chunk-right-context-final 0 \ | ||
--egs.opts "--frames-overlap-per-eg 0" \ | ||
--egs.chunk-width $chunk_width \ | ||
--trainer.num-chunk-per-minibatch 128 \ | ||
--trainer.num-chunk-per-minibatch 128,64 \ | ||
--trainer.frames-per-iter 1500000 \ | ||
--trainer.num-epochs 4 \ | ||
--trainer.num-epochs $num_epochs \ | ||
--trainer.optimization.num-jobs-initial 2 \ | ||
--trainer.optimization.num-jobs-final 12 \ | ||
--trainer.dropout-schedule $dropout_schedule \ | ||
|
Oops, something went wrong.