Skip to content

add reverberated TDNN+LSTM recipe on AMI; a fix to aspire recipe#1314

Merged
vijayaditya merged 3 commits intokaldi-asr:masterfrom
tomkocse:ami_rvb2
Jan 7, 2017
Merged

add reverberated TDNN+LSTM recipe on AMI; a fix to aspire recipe#1314
vijayaditya merged 3 commits intokaldi-asr:masterfrom
tomkocse:ami_rvb2

Conversation

@tomkocse
Copy link
Contributor

@tomkocse tomkocse commented Jan 4, 2017

No description provided.

@danpovey
Copy link
Contributor

danpovey commented Jan 4, 2017 via email

@vijayaditya
Copy link
Contributor

vijayaditya commented Jan 4, 2017 via email

Copy link
Contributor

@vijayaditya vijayaditya left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Could you also softlink run_tdnn_lstm.sh in the chain/ directory to tdnn_lstm_1i.sh script.

# local/chain/multi_condition/run_tdnn_lstm.sh --mic mdm8 --use-ihm-ali true --train-set train_cleaned --gmm tri3_cleaned
# cleanup + chain TDNN+LSTM model, MDM original + IHM reverberated data, alignments from IHM data
# *** best system ***
%WER 31.8 | 14488 94497 | 71.8 15.4 12.8 3.5 31.8 62.7 | 0.698 | exp/mdm8/chain_cleaned_rvb/tdnn_lstm1i_sp_rvb_bi_ihmali/decode_dev/ascore_10/dev_hires_o4.ctm.filt.sys
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@tomkocse Very impressive results. Please add the version number of your script which generates these results (eg. run_tdnn_lstm_1a.sh ) as we might replace run_tdnn_lstm.sh with a better tdnn_lstm architecture in the future.

Move this script to tuning directory as run_tdnn_lstm_1b.sh and softlink it from run_tdnn_lstm.sh

%WER 40.9 | 13807 89961 | 62.4 20.0 17.6 3.3 40.9 65.7 | 0.612 | exp/sdm1/chain_cleaned/tdnn_lstm1i_sp_bi_ihmali_ld5/decode_eval/ascore_10/eval_hires_o4.ctm.filt.sys


# local/chain/multi_condition/run_tdnn_lstm.sh --mic sdm1 --use-ihm-ali true --train-set train_cleaned --gmm tri3_cleaned
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

same comment as above; start versioning your run*.sh scripts.

@@ -0,0 +1,334 @@
#!/bin/bash

# This is a chain-training script with TDNN+LSTM neural networks.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Not checking this in detail as I assume it is similar to multi_condition/run_tdnn.sh except for the xconfig and some lstm specific parameters.


# check steps/libs/nnet3/xconfig/lstm.py for the other options and defaults
lstmp-layer name=lstm1 cell-dim=1024 recurrent-projection-dim=256 non-recurrent-projection-dim=256 delay=-3
relu-renorm-layer name=tdnn4 input=Append(-3,0,3) dim=1024
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Try increasing the parameters further as you have more data than the SDM/MDM only case. Though we might not ultimately recommend a large parameter model it is better to know the least possible WER with this architecture and simulated multi_condition data. You can try increasing the relu-norm-layer dim in increments of 256 or 128.

--egs.chunk-right-context $chunk_right_context \
--trainer.num-chunk-per-minibatch 64 \
--trainer.frames-per-iter 1500000 \
--trainer.num-epochs 4 \
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Check if you can increase the number of epochs.

data/train data/train_temp_for_lats
utils/data/combine_short_segments.sh \
data/train_temp_for_lats $min_seg_len data/train_min${min_seg_len}
steps/compute_cmvn_stats.sh data/train_min${min_seg_len} exp/make_mfcc/train_min${min_seg_len} mfcc || exit 1;
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I recommend using the new style of not specifying logdirs.

@danpovey
Copy link
Contributor

danpovey commented Jan 4, 2017 via email

@vijayaditya
Copy link
Contributor

@danpovey Thanks for pointing out I didn't observe that the num-epochs was not reduced. @tomkocse Could you please generate the log-probability plots for the training run with multi_condition data. You might know if your training has converged.

@vijayaditya vijayaditya merged commit a4b2091 into kaldi-asr:master Jan 7, 2017
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants