Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
368 commits
Select commit Hold shift + click to select a range
638566b
[egs] Small fix to madcat_zh recipe (#2681)
ChunChiehChang Sep 4, 2018
7531b6b
[egs] Fix issue in egs/csj/s5/local/csj_data_prep.sh (failure if run …
tekezo Sep 5, 2018
c40205f
[build] Allow configure script to handle package-based OpenBLAS (#2618)
btiplitz Sep 5, 2018
1cd9d41
[egs] updating local/make_voxceleb1.pl so that it works with newer ve…
david-ryan-snyder Sep 6, 2018
72ff0cd
[egs,scripts] Remove unused --nj option from some scripts (#2679)
david-ryan-snyder Sep 6, 2018
0859b4f
[egs] Fix to tedlium v3 run.sh (rnnlm rescoring) (#2686)
chaoweihuang Sep 6, 2018
c81f517
[scripts,egs] Tamil OCR with training data from yomdle and testing da…
aarora8 Sep 6, 2018
312c40e
[egs] mini_librispeech: allow relative pathnames in download_and_unta…
desh2608 Sep 7, 2018
7741f7c
[egs] Updating SITW recipe to account for changes to VoxCeleb1 (#2690)
david-ryan-snyder Sep 7, 2018
c8e46d1
[src] Fix nnet1 proj-lstm bug where gradient clipping not used; thx:@…
danpovey Sep 11, 2018
dcf2466
[egs] Update aishell2 recipe to allow online decoding (no pitch for i…
naxingyu Sep 12, 2018
a06a7df
[src] Make cublas and cusparse use per-thread streams. (#2692)
galv Sep 12, 2018
c99a860
[src] improve handling of low-rank covariance in ivector-compute-lda …
david-ryan-snyder Sep 12, 2018
17b8f6d
[egs] Changes to IAM handwriting-recognition recipe, including BPE en…
aarora8 Sep 12, 2018
697723a
[scripts] Make sure pitch is not included in i-vector feats, in onlin…
naxingyu Sep 13, 2018
076cc46
[src] fix help message in post-to-smat (#2703)
leibny Sep 13, 2018
6c9c00d
[scripts] Fix to steps/cleanup/debug_lexicon.sh (#2704)
danpovey Sep 13, 2018
62a6e5e
[egs] Cosmetic and file-mode fixes in HKUST recipe (#2708)
danpovey Sep 14, 2018
0cf2e23
[scripts] nnet1: remove the log-print of args in 'make_nnet_proto.py'…
KarelVesely84 Sep 14, 2018
2cfcfda
[egs] update README in AISHELL-2 (#2710)
Sep 15, 2018
4f108fc
[src] Make constructor of CuDevice private (#2711)
galv Sep 16, 2018
ab56793
[egs] fix sorting issue in aishell v1 (#2705)
naxingyu Sep 16, 2018
88f23ce
[egs] Add soft links for CNN+TDNN scripts (#2715)
danpovey Sep 17, 2018
dfb554f
[build] Add missing packages in extras/check_dependencies.sh (#2719)
hhadian Sep 18, 2018
1f1ed0f
[egs] madcat arabic: clean scripts, tuning, use 6-gram LM (#2718)
aarora8 Sep 19, 2018
9b9196b
[egs] Update WSJ run.sh: comment out outdated things, add run_tdnn.sh…
danpovey Sep 19, 2018
69cd717
[scripts,src] Fix potential issue in scripts; minor fixes. (#2724)
danpovey Sep 19, 2018
4566a3e
[egs] add example script for RNNLM lattice rescoring for WSJ recipe (…
hainan-xv Sep 20, 2018
ebbae44
[egs] add rnnlm example on tedlium+lm1b; add rnnlm rescoring results …
keli78 Sep 20, 2018
acda0f4
[scripts] Small fix to utils/data/convert_data_dir_to_whole.sh (RE ba…
sih4sing5hong5 Sep 25, 2018
8cf8545
[src] fix memory bug in kaldi::~LatticeFasterDecoderTpl(), (#2737)
KarelVesely84 Sep 25, 2018
0b0e512
[egs] Remove per-utt option from nnet3/align scripts (#2717)
vimalmanohar Sep 26, 2018
396c779
[egs] Small Librispeech example fix, thanks: Yasasa Tennakoon. (#2738)
danpovey Sep 26, 2018
1d079fa
[egs] Aishell2 recipe: turn off jieba's new word discovery in word se…
dophist Sep 26, 2018
f1f9a48
[egs] Add missing file local/join_suffix.py in TEDLIUM s5_r3; thx:ana…
huangruizhe Sep 26, 2018
6d919f0
[egs,scripts] Add Tunisian Arabic (MSA) recipe; cosmetic fixes to pbs…
xiaohui-zhang Sep 26, 2018
19a1811
[scripts] Fix missing import in utils/langs/grammar/augment_words_txt…
francoishernandez Sep 28, 2018
02d462b
[scripts] Fix build_const_arpa_lm.sh w.r.t. where <s> appears inside…
ChunChiehChang Oct 1, 2018
1be99b6
[scripts] Slight improvements to decode_score_fusion.sh usability (#2…
teinhonglo Oct 2, 2018
3d3d02e
[build] update configure to support cuda 10 (#2747)
leomao Oct 2, 2018
168789d
[scripts] Fix bug in utils/data/resample_data_dir.sh (#2749)
hhadian Oct 2, 2018
d99aa70
[scripts] Fix bug in cleanup after steps/cleanup/clean_and_segment_da…
teinhonglo Oct 2, 2018
23ec7a1
[egs] several updates of the tunisian_msa recipe (#2752)
xiaohui-zhang Oct 3, 2018
9c813ca
[egs] Small fix to Tunisian MSA TDNN script (RE train_stage) (#2757)
francoishernandez Oct 3, 2018
c78875a
[src,scripts] Batched nnet3 computation (#2726)
danpovey Oct 4, 2018
8b8604c
[build] Add python3 compatibility to install scripts (#2748)
leomao Oct 4, 2018
2d163f4
[scripts] tfrnnlm: Modify TensorFlow flag format for compatibility wi…
ervaibhavkumar Oct 4, 2018
e3150a2
[egs] fix old style perl regex in egs/chime1/s5/local/chime1_prepare_…
kobayakawa-t-ko Oct 5, 2018
ad2bb82
[scripts] Fix bug in steps/cleanup/debug_lexicon.sh (#2763)
xiaohui-zhang Oct 5, 2018
c0e34f8
[egs] Add example for Yomdle Farsi OCR (#2702)
ChunChiehChang Oct 5, 2018
4c8d446
[scripts] debug_lexicon.sh: Fix bug introduced in #2763. (#2764)
danpovey Oct 6, 2018
f279dd9
[egs] add missing online cmvn config in aishell2 (#2767)
dophist Oct 7, 2018
f91dde6
[egs] Add CNN-TDNN-F script for Librispeech (#2744)
yangxueruivs Oct 7, 2018
e54462e
[src] Some minor cleanup/fixes regarding CUDA memory allocation; othe…
danpovey Oct 7, 2018
a702621
[scripts] Update reverberate_data_dir.py so that it works with python…
david-ryan-snyder Oct 8, 2018
00c7382
[egs] Chime5: fix total number of words for WER calculation (#2772)
siddalmia Oct 9, 2018
50411bd
[egs] RNNLMs on Tedlium w/ Google 1Bword: Increase epochs, update res…
keli78 Oct 9, 2018
735e2a5
[scripts,egs] Added phonetisaurus-based g2p scripts (#2730)
huangruizhe Oct 9, 2018
43ec82e
[egs] madcat arabic: clean scripts, tuning, rescoring, text localizat…
aarora8 Oct 9, 2018
535bb2c
[scripts] Enhancements & minor bugfix to segmentation postprocessing …
stanleyguan Oct 11, 2018
8d1eb93
[src] Update gmm-decode-simple to accept ConstFst (#2787)
desh2608 Oct 12, 2018
087c21f
[scripts] Update documentation of train_raw_dnn.py (#2785)
vimalmanohar Oct 13, 2018
a10e56e
[src] nnet3: extend what descriptors can be parsed. (#2780)
danpovey Oct 13, 2018
8b08e00
[src] Small fix to 'fstrand' (make sure args are parsed) (#2777)
huangruizhe Oct 13, 2018
dc6a496
[src,scripts] Minor, mostly cosmetic updates (#2788)
danpovey Oct 13, 2018
59d1544
[src,scripts] Add script to compare alignment directories. (#2765)
danpovey Oct 13, 2018
7e9d8cb
[scripts] Small fixes to script usage messages, etc. (#2789)
teinhonglo Oct 13, 2018
fcf9b97
[egs] Update ami_download.sh after changes on Edinburgh website. (#2769)
puneetbawa Oct 15, 2018
024d130
[scripts] Update compare_alignments.sh to allow different lang dirs. …
danpovey Oct 17, 2018
f5e8171
[scripts] Change make_rttm.py so output is in determinstic order (#2794)
oplatek Oct 18, 2018
2668098
[egs] Fixes to yomdle_zh RE encoding direction, etc. (#2791)
ChunChiehChang Oct 19, 2018
8e30fdd
[src] Add support for context independent phones in gmm-init-biphone …
hhadian Oct 20, 2018
d68c00a
[egs] Simplifying multi-condition version of AMI recipe (#2800)
vimalmanohar Oct 27, 2018
8b4a1bf
[build] Fix openblas build for aarch64 (#2806)
ryanleary Oct 30, 2018
dafeb84
[build] Make CUDA_ARCH configurable at configure-script level (#2807)
ryanleary Nov 1, 2018
0300493
[src] Print maximum memory stats in CUDA allocator (#2799)
btiplitz Nov 1, 2018
faa1f3c
[src,scripts] Various minor code cleanups (#2809)
yzmyyff Nov 1, 2018
f70ec64
[scripts] Fix handling of UTF-8 in filenames, in wer_per_spk_details.…
agurianov Nov 5, 2018
4bdb05a
[egs] Update AMI chain recipes (#2817)
vimalmanohar Nov 5, 2018
cc745dd
[egs] Improvements to multi_en tdnn-opgru/lstm recipes (#2824)
xiaohui-zhang Nov 6, 2018
ec7e592
[scripts] Fix initial prob of silence when lexicon has silprobs. Thx…
danpovey Nov 7, 2018
8ea2a7e
[scripts,src] Fix to multitask nnet3 training (#2818); cosmetic code …
danpovey Nov 8, 2018
5b23ace
[scripts] Create shared versions of get_ctm_conf.sh, add get_ctm_conf…
xiaohui-zhang Nov 9, 2018
d66bee3
[src] Use cuda streams in matrix library (#2821)
ryanleary Nov 12, 2018
cd6d5a5
[egs] Add online-decoding recipe to aishell1 (#2829)
naxingyu Nov 13, 2018
3ae133c
[egs] Add DIHARD 2018 diarization recipe. (#2822)
HuangZiliAndy Nov 13, 2018
2c87b3a
[egs] add nnet3 online result for aishell1 (#2836)
naxingyu Nov 14, 2018
33ada45
[scripts] RNNLM scripts: don't die when features.txt is not present (…
slckl Nov 14, 2018
d337921
[src] Optimize cuda allocator for multi-threaded case (#2820)
ryanleary Nov 15, 2018
741b75b
[build] Add cub library for cuda projects (#2819)
ryanleary Nov 15, 2018
d9c2eb3
[src] Make Cuda allocator statistics visible to program (#2835)
btiplitz Nov 15, 2018
69a09b3
[src] Fix bug affecting scale in GeneralDropoutComponent (non-continu…
GaofengCheng Nov 16, 2018
219e60c
[build] FIX #2842: properly check $use_cuda against false. (#2843)
chuanma Nov 16, 2018
da76c6b
[doc] Add note about OOVs to data-prep. (#2844)
danpovey Nov 17, 2018
c738017
[scripts] Allow segmentation with nnet3 chain models (#2845)
nshmyrev Nov 17, 2018
f02b324
[build] Remove -lcuda from cuda makefiles which breaks operation when…
danpovey Nov 19, 2018
c026f72
[scripts] Fix error in analyze_lats.sh for long lattices (replace awk…
jinyiyang-jhu Nov 19, 2018
a121244
[egs] add rnnlm recipe for librispeech (#2830)
keli78 Nov 19, 2018
0cf1382
[build] change configure version from 9 to 10 (#2853) (#2855)
keli78 Nov 19, 2018
73fad9f
[src] fixed compilation errors when built with --DOUBLE_PRECISION=1 (…
huangruizhe Nov 20, 2018
f89709c
[build] Clarify instructions if cub is not found (#2858)
danpovey Nov 20, 2018
2489bb9
[egs] Limit MFCC feature extraction job number in Dihard recipe (#2865)
HuangZiliAndy Nov 21, 2018
60db654
[egs] Added Bentham handwriting recognition recipe (#2846)
desh2608 Nov 22, 2018
172e886
[src] Share roots of different tones of phones aishell (#2859)
naxingyu Nov 25, 2018
8755661
[egs] Fix path to sequitur in commonvoice egs (#2868)
vvasily Nov 25, 2018
64d5cf2
[egs] Update reverb recipe (#2753)
Szu-JuiChen Nov 26, 2018
d5bdb88
[scripts] Fix error while analyzing lattice (parsing bugs) (#2873)
jinyiyang-jhu Nov 27, 2018
8a87f53
[src] Fix memory leak in OnlineCacheFeature; thanks @Worldexe (#2872)
danpovey Nov 27, 2018
70aba21
[egs] TIMIT: fix mac compatibility of sed command (#2874)
nithinraok Nov 27, 2018
93a0a46
[egs] mini_librispeech: fixing some bugs and limiting repeated downlo…
jtrmal Nov 27, 2018
132eb42
[src,scripts,egs] Speedups to GRU-based networks (special components)…
LvHang Nov 27, 2018
84435ff
[src] Fix infinite recursion with -DDOUBLE_PRECISION=1. Thx: @hwiorn …
huangruizhe Nov 27, 2018
b196b7f
Revert "[src] Fix infinite recursion with -DDOUBLE_PRECISION=1. Thx: …
danpovey Nov 27, 2018
286e912
Revert "Revert "[src] Fix infinite recursion with -DDOUBLE_PRECISION=…
danpovey Nov 27, 2018
21ffb84
Revert "[src] Fix memory leak in OnlineCacheFeature; thanks @Worldexe…
danpovey Nov 27, 2018
643623d
[src] Remove unused code that caused Windows compile failure. Thx:@b…
danpovey Nov 27, 2018
d82fcc6
[src] Really fix memory leak in online decoding; thx:@worldexe (#2883)
danpovey Nov 28, 2018
486de21
[src] Fix Windows cuda build failure (use C++11 standard include) (#2…
btiplitz Nov 28, 2018
6d233d0
[src] Add #include that caused build failure on Windows (#2886)
btiplitz Nov 28, 2018
8daa02e
[scripts] Fix max duration check in sad_to_segments.py (#2889)
dogancan Nov 29, 2018
4b14967
[scripts] Fix speech duration calculation in sad_to_segments.py (#2891)
dogancan Nov 29, 2018
9bf0b6d
[src] Fix Windows build problem (timer.h) (#2888)
huangruizhe Nov 29, 2018
0e5d755
[egs] add HUB4 spanish tdnn-f and cnn-tdnn script (#2895)
entn-at Nov 30, 2018
6209b99
[egs] Fix Aishell2 dict prepare bug; should not affect results (#2890)
naxingyu Dec 1, 2018
40a9dc6
[egs] Self-contained example for KWS for mini_librispeech (#2887)
jtrmal Dec 1, 2018
ff514e3
[egs,scripts] Fix bugs in Dihard 2018 (#2897)
HuangZiliAndy Dec 3, 2018
a2a0564
[scripts] Check last character of files to match with newline (#2898)
jinyiyang-jhu Dec 4, 2018
37091d6
[egs] Update Librispeech RNNLM results; use correct training data (#2…
keli78 Dec 6, 2018
b50a4cf
[scripts] RNNLM: old iteration model cleanup; save space (#2885)
slckl Dec 7, 2018
a464bd7
[scripts] Make prepare_lang.sh cleanup beforehand (prevents certain f…
danpovey Dec 11, 2018
c41cbb1
[scripts] Expose dim-range-node at xconfig level (#2903)
yangxueruivs Dec 11, 2018
aa0ac7b
[scripts] Fix bug related to multi-task in train_raw_rnn.py (#2907)
danpovey Dec 12, 2018
3e50be9
[scripts] Cosmetic fix/clarification to utils/prepare_lang.sh (#2912)
danpovey Dec 12, 2018
791cd82
[scripts,egs] Added a new lexicon learning (adaptation) recipe for te…
xiaohui-zhang Dec 14, 2018
b126161
[egs] TDNN+LSTM example scripts, with RNNLM, for Librispeech (#2857)
GaofengCheng Dec 15, 2018
78f0127
[src] cosmetic fix in nnet1 code (#2921)
csukuangfj Dec 17, 2018
44980dd
[src] Fix incorrect invocation of mutex in nnet-batch-compute code (#…
danpovey Dec 21, 2018
a46f554
[egs,minor] Fix typo in comment in voxceleb script (#2926)
corollari Dec 23, 2018
2edb074
[src,egs] Mostly cosmetic changes; add some missing includes (#2936)
yzmyyff Dec 24, 2018
9b320ad
[egs] Fix path of rescoring binaries used in tfrnnlm scripts (#2941)
virenderkadyan Dec 27, 2018
3b0162b
[src] Fix bug in nnet3-latgen-faster-batch for determinize=false (#2945)
danpovey Dec 28, 2018
b984543
[egs] Add example for rimes handwriting database; Madcat arabic scrip…
aarora8 Dec 28, 2018
46826d9
[egs] Add scripts for yomdle korean (#2942)
aarora8 Dec 28, 2018
3e77220
[build] Refactor/cleanup build system, easier build on ubuntu 18.04. …
danpovey Dec 31, 2018
5a720ac
[scripts,egs] Changes for Python 2/3 compatibility (#2925)
desh2608 Dec 31, 2018
ca32c4e
[egs] Add more modern DNN recipe for fisher_callhome_spanish (#2951)
GoVivace Dec 31, 2018
1ea2ba7
[scripts] switch from bc to perl to reduce dependencies (diarization …
mmaciej2 Jan 1, 2019
969869c
[scripts] Further fix for Python 2/3 compatibility (#2957)
desh2608 Jan 2, 2019
97d14df
[egs] Remove no-longer-existing option in tedlium_r3 recipe (#2959)
huangruizhe Jan 3, 2019
0134893
[build] Handle dependencies for .cu files in addition to .cc files (#…
ryanleary Jan 3, 2019
5ae6e9e
[src] remove duplicate test mode option from class GeneralDropoutComp…
chihyang Jan 4, 2019
fef6909
[egs] Fix minor bugs in WSJ's flat-start/e2e recipe (#2968)
hhadian Jan 5, 2019
b9d4b76
[egs] Fix to BSD compatibility of TIMIT data prep (#2966)
danpovey Jan 5, 2019
6b17571
[scripts] Fix RNNLM training script problem (chunk_length was ignored…
hainan-xv Jan 5, 2019
32b8cf1
[src] Fix bug in lattice-1best.cc RE removing insertion penalty (#2970)
freewym Jan 6, 2019
1079922
[src] Compute a separate avg (start, end) interval for each sausage w…
dogancan Jan 6, 2019
205dbd8
[build] Move nvcc verbose flag to proper location (#2962)
ryanleary Jan 7, 2019
1cac236
[egs] Fix mini_librispeech download_lm.sh crash; thx:chris.keith.john…
danpovey Jan 7, 2019
37d6950
[egs] minor fixes related to python2 vs python3 differences (#2977)
david-ryan-snyder Jan 8, 2019
a6aa269
[src] Small fix in test code, avoid spurious failure (#2978)
danpovey Jan 9, 2019
2864465
[egs] Fix CSJ data-prep; minor path fix for USB version of data (#2979)
feddybear Jan 9, 2019
f02d2a3
[egs] Add paper ref to README.txt in reverb example (#2982)
sas91 Jan 10, 2019
ff6ddf7
[egs] Minor fixes to sitw recipe (fix problem introdueced in #2925) (…
david-ryan-snyder Jan 11, 2019
9b6fbdd
[scripts] Fix bug introduced in #2957, RE integer division (#2986)
aarora8 Jan 11, 2019
c017268
[egs] Update WSJ flat-start chain recipes to use TDNN-F not TDNN+LSTM…
hhadian Jan 12, 2019
c631fcb
[scripts] Fix typo introduced in #2925 (#2989)
desh2608 Jan 13, 2019
9f981d0
[build] Modify Makefile and travis script to fix Travis failures (#2987)
galv Jan 14, 2019
ae573c9
[src] Simplification and efficiency improvement in ivector-plda-scori…
david-ryan-snyder Jan 16, 2019
50af3fc
[egs] Update madcat Arabic and Chinese egs, IAM (#2964)
aarora8 Jan 16, 2019
f90a98c
[src] Fix overflow bug in convolution code (#2992)
ChunChiehChang Jan 16, 2019
fd0aca9
[src] Fix nan issue in ctm times introduced in #2972, thx: @vesis84 (…
vimalmanohar Jan 16, 2019
e8d1287
[src] Fix 'sausage-time' issue which occurs with disabled MBR decodin…
KarelVesely84 Jan 18, 2019
99dc4d8
[egs] Add scripts for yomdle Russian (OCR task) (#2953)
aarora8 Jan 21, 2019
7e529ed
[egs] Simplify lexicon preparation in Fisher callhome Spanish (#2999)
GoVivace Jan 21, 2019
25f09e8
[egs] Update GALE Arabic recipe (#2934)
aarora8 Jan 22, 2019
4338004
[egs] Remove outdated NN results from Gale Arabic recipe (#3002)
aarora8 Jan 22, 2019
05d9a3d
[egs] Add RESULTS file for the tedlium s5_r3 (release 3) setup (#3003)
huangruizhe Jan 23, 2019
1dcdf80
[src] Fixes to grammar-fst code to handle LM-disambig symbols properl…
danpovey Jan 26, 2019
6f56512
[src] Cosmetic change to mel computation (fix option string) (#3011)
boeddeker Jan 30, 2019
56cfb95
[src] Fix Visual Studio error due to alternate syntactic form of nore…
daanzu Feb 1, 2019
9e35898
[egs] Fix location of sequitur installation (#3017)
jybaek Feb 1, 2019
a51bd96
[src] Fix w/ ifdef Visual Studio error from alternate syntactic form …
daanzu Feb 3, 2019
41ea8cf
[egs] Some fixes to getting data in heroico recipe (#3021)
danpovey Feb 3, 2019
fb514dc
[egs] BABEL script fix: avoid make_L_align.sh generating invalid file…
jtrmal Feb 4, 2019
afc5e78
[src] Fix to older online decoding code in online/ (OnlineFeInput; wa…
jdieguez Feb 6, 2019
226cbf7
[script] Fix unset bash variable in make_mfcc.sh (#3030)
oplatek Feb 8, 2019
6fc4c60
[scripts] Extend limit_num_gpus.sh to support --num-gpus 0. (#3027)
oplatek Feb 8, 2019
2f92bd9
[scripts] fix bug in utils/add_lex_disambig.pl when sil-probs and pro…
Teddyang Feb 15, 2019
403c5ee
[egs] Fix path in Tedlium r3 rnnlm training script (#3039)
francoishernandez Feb 18, 2019
abfbc56
[src] Thread-safety for GrammarFst (thx:armando.muscariello@gmail.com…
danpovey Feb 20, 2019
f09d48a
[scripts] Cosmetic fix to get_degs.sh (#3045)
Teddyang Feb 21, 2019
b0fc09d
[egs] Small bug fixes for IAM and UW3 recipes (#3048)
ChunChiehChang Feb 21, 2019
4494a85
[scripts] Nnet3 segmentation: fix default params (#3051)
danpovey Feb 26, 2019
bf33f1f
[scripts] Allow perturb_data_dir_speed.sh to work with utt2lang (#3055)
igrinis Feb 26, 2019
5f05d59
[scripts] Make beam in monophone training configurable (#3057)
xiaohui-zhang Feb 27, 2019
c0a555e
[scripts] Allow reverberate_data_dir.py to support unicode filenames …
rezame Feb 27, 2019
2e26464
[scripts] Make some cleanup scripts work with python3 (#3054)
vimalmanohar Mar 1, 2019
d21be2d
[scripts] bug fix to nnet2->3 conversion, fixes #886 (#3071)
jfainberg Mar 4, 2019
8fa9648
[src] Make copies occur in per-thread default stream (for GPUs) (#3068)
luitjens Mar 4, 2019
bd326dc
[src] Add GPU version of MergeTaskOutput().. relates to batch decodin…
luitjens Mar 4, 2019
17b7f3f
[src] Add device options to enable tensor core math mode. (#3066)
luitjens Mar 4, 2019
0a1f827
[src] Log nnet3 computation to VLOG, not std::cout (#3072)
kkm000 Mar 5, 2019
f2a89c2
[src] Allow upsampling in compute-mfcc-feats, etc. (#3014)
danpovey Mar 5, 2019
98b45c8
[src] fix problem with rand_r being undefined on Android (#3037)
keli78 Mar 5, 2019
197214d
[egs] Update swbd1_map_words.pl, fix them_1's -> them's (#3052)
Mar 5, 2019
991a75c
[src] Add const overload OnlineNnet2FeaturePipeline::IvectorFeature (…
kkm000 Mar 6, 2019
4432371
[src] Fix syntax error in egs/bn_music_speech/v1/local/make_musan.py …
antonstakhouski Mar 6, 2019
8460fa3
[src] Memory optimization for online feature extraction of long recor…
pzelasko Mar 6, 2019
b801b98
[build] fixed a bug in linux_configure_redhat_fat when use_cuda=no (#…
kan-bayashi Mar 7, 2019
ce97c47
[scripts] Add missing '. ./path.sh' to get_utt2num_frames.sh (#3076)
hhadian Mar 7, 2019
4d61452
[src,scripts,egs] Add count-based biphone tree tying for flat-start c…
hhadian Mar 7, 2019
01cef69
[scripts,egs] Remove sed from various scripts (avoid compatibility pr…
desh2608 Mar 8, 2019
2f95609
[src] Rework error logging for safety and cleanliness (#3064)
kkm000 Mar 8, 2019
bcfe3f8
[src] Change warp-synchronous to cub::BlockReduce (safer but slower) …
desh2608 Mar 10, 2019
1209c07
[src] Fix && and || uses where & and | intended, and other weird erro…
kkm000 Mar 11, 2019
5a5696f
[build] Some fixes to Makefiles (#3088)
kkm000 Mar 11, 2019
abd4869
[src] Fixed -Wreordered warnings in feat (#3090)
pzelasko Mar 12, 2019
9c8ba0f
[egs] Replace bc with perl -e (#3093)
entn-at Mar 12, 2019
8cbd582
[scripts] Fix python3 compatibility issue in data-perturbing script (…
nikhilm16 Mar 12, 2019
7435661
[doc] fix some typos in doc. (#3097)
csukuangfj Mar 12, 2019
5bdea69
[build] Make sure expf() speed probe times sensibly (#3089)
kkm000 Mar 12, 2019
b7a4fec
[scripts] Make sure merge_targets.py works in python3 (#3094)
XIAOYixuan Mar 12, 2019
94475d6
[src] ifdef to fix compilation failure on CUDA 8 and earlier (#3103)
desh2608 Mar 13, 2019
fc8c17b
[doc] fix typos and broken links in doc. (#3102)
csukuangfj Mar 13, 2019
3f8b6b2
[scripts] Fix frame_shift bug in egs/swbd/s5c/local/score_sclite_conf…
freewym Mar 13, 2019
633e61c
[src] Fix wrong assertion failure in nnet3-am-compute (#3106)
MartinKocour Mar 14, 2019
8cafd32
[src] Cosmetic changes to natural-gradient code (#3108)
danpovey Mar 14, 2019
b1b230c
[src,scripts] Python2 compatibility fixes and code cleanup for nnet1 …
KarelVesely84 Mar 14, 2019
9c875ef
[doc] Small documentation fixes; update on Kaldi history (#3031)
KarelVesely84 Mar 14, 2019
7a1908f
[src] Various mostly-cosmetic changes (copying from another branch) (…
danpovey Mar 15, 2019
fcd70d3
[scripts] Simplify text encoding in RNNLM scripts (now only support …
saikiranvalluri Mar 16, 2019
b4c7ab6
[egs] Add "formosa_speech" recipe (Taiwanese Mandarin ASR) (#2474)
yfliao Mar 16, 2019
461b50c
[egs] python3 compatibility in csj example script (#3123)
rickychanhoyin Mar 16, 2019
61637e6
[egs] python3 compatibility in example scripts (#3126)
danpovey Mar 17, 2019
1f068cd
[scripts] Bug-fix for removing deleted words (#3116)
psmit Mar 17, 2019
8d60ee3
[scripts] Add fix regarding num-jobs for segment_long_utterances*.sh(…
vimalmanohar Mar 17, 2019
7fb716a
[src] Enable allow_{upsample,downsample} with online features (#3139)
jtrmal Mar 18, 2019
80c1437
[src] Fix bad assert in fstmakecontextsyms (#3142)
Mar 19, 2019
0d6ead5
[src] Fix to "Fixes to grammar-fst & LM-disambig symbols" (#3000) (#3…
daanzu Mar 19, 2019
338b586
[build] Make sure PaUtils exported from portaudio (#3144)
jtrmal Mar 19, 2019
73720e6
[src] cudamatrix: fixing a synchronization bug in 'normalize-per-row'…
KarelVesely84 Mar 20, 2019
f9276a5
[src] Fix typo in comment (#3147)
csukuangfj Mar 20, 2019
252690f
[src] Add binary that functions as a TCP server (#2938)
danijel3 Mar 20, 2019
6134c29
[scripts] Fix bug in comment (#3152)
Shujian2015 Mar 21, 2019
aead118
[scripts] Fix bug in steps/segmentation/ali_to_targets.sh (#3155)
saikiranvalluri Mar 21, 2019
213ae52
[scripts] Avoid holding out more data than the requested num-utts (du…
kkm000 Mar 21, 2019
1ac8c92
[src,scripts] Add support for two-pass agglomerative clustering. (#3058)
dogancan Mar 24, 2019
6bd9dad
[src] Disable unget warning in PeekToken (and other small fix) (#3163)
kkm000 Mar 24, 2019
37f4f44
[build] Add new nvidia tools to windows build (#3159)
btiplitz Mar 24, 2019
77ac79f
[doc] Fix documentation errors and add more docs for tcp-server decod…
danijel3 Mar 24, 2019
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
The table of contents is too big for display.
Diff view
Diff view
  •  
  •  
  •  
13 changes: 9 additions & 4 deletions .gitignore
Original file line number Diff line number Diff line change
Expand Up @@ -73,15 +73,17 @@ GSYMS
/src/kaldi.mk.bak

# /egs/
/egs/*/s*/mfcc
/egs/*/s*/plp
/egs/*/s*/exp
/egs/*/s*/data
/egs/*/*/mfcc
/egs/*/*/plp
/egs/*/*/exp
/egs/*/*/data

# /tools/
/tools/pocolm/
/tools/ATLAS/
/tools/atlas3.8.3.tar.gz
/tools/irstlm/
/tools/mitlm/
/tools/openfst
/tools/openfst-1.3.2.tar.gz
/tools/openfst-1.3.2/
Expand Down Expand Up @@ -143,3 +145,6 @@ GSYMS
/tools/mmseg-1.3.0.tar.gz
/tools/mmseg-1.3.0/
/kaldiwin_vs*
/tools/cub-1.8.0.zip
/tools/cub-1.8.0/
/tools/cub
2 changes: 1 addition & 1 deletion .travis.yml
Original file line number Diff line number Diff line change
Expand Up @@ -49,7 +49,7 @@ script:
# for the explanation why extra switches needed for clang with ccache.
- CXX="ccache clang++-3.8 -Qunused-arguments -fcolor-diagnostics -Wno-tautological-compare"
CFLAGS=""
LDFLAGS="-llapack"
LDFLAGS="-llapack -Wl,-fuse-ld=gold"
INCDIRS="$XROOT/usr/include"
LIBDIRS="$XROOT/usr/lib"
tools/extras/travis_script.sh
Expand Down
26 changes: 18 additions & 8 deletions egs/aishell/s5/RESULTS
Original file line number Diff line number Diff line change
@@ -1,8 +1,18 @@
%WER 33.82 [ 35432 / 104765, 743 ins, 3991 del, 30698 sub ] exp/mono/decode_test/cer_12_0.0
%WER 19.39 [ 20310 / 104765, 903 ins, 1452 del, 17955 sub ] exp/tri1/decode_test/cer_13_0.5
%WER 19.23 [ 20147 / 104765, 910 ins, 1287 del, 17950 sub ] exp/tri2/decode_test/cer_14_0.5
%WER 17.14 [ 17961 / 104765, 812 ins, 1024 del, 16125 sub ] exp/tri3a/decode_test/cer_14_0.0
%WER 13.64 [ 14294 / 104765, 669 ins, 736 del, 12889 sub ] exp/tri4a/decode_test/cer_14_0.5
%WER 12.23 [ 12809 / 104765, 656 ins, 580 del, 11573 sub ] exp/tri5a/decode_test/cer_13_1.0
%WER 8.45 [ 8849 / 104765, 312 ins, 538 del, 7999 sub ] exp/nnet3/tdnn_sp/decode_test/cer_13_1.0
%WER 7.46 [ 7813 / 104765, 287 ins, 472 del, 7054 sub ] exp/chain/tdnn_1a_sp/decode_test/cer_10_1.0
%WER 36.41 [ 38146 / 104765, 837 ins, 3114 del, 34195 sub ] exp/mono/decode_test/cer_10_0.0
%WER 18.76 [ 19654 / 104765, 949 ins, 1152 del, 17553 sub ] exp/tri1/decode_test/cer_13_0.5
%WER 18.64 [ 19531 / 104765, 941 ins, 1159 del, 17431 sub ] exp/tri2/decode_test/cer_14_0.5
%WER 17.04 [ 17849 / 104765, 810 ins, 1021 del, 16018 sub ] exp/tri3a/decode_test/cer_14_0.5
%WER 13.82 [ 14482 / 104765, 764 ins, 670 del, 13048 sub ] exp/tri4a/decode_test/cer_13_0.5
%WER 12.12 [ 12694 / 104765, 751 ins, 523 del, 11420 sub ] exp/tri5a/decode_test/cer_13_0.5
%WER 8.65 [ 9064 / 104765, 367 ins, 455 del, 8242 sub ] exp/nnet3/tdnn_sp/decode_test/cer_14_0.5
%WER 7.48 [ 7839 / 104765, 285 ins, 454 del, 7100 sub ] exp/chain/tdnn_1a_sp/decode_test/cer_10_1.0

# nnet3 tdnn with online pitch, local/nnet3/tuning/tun_tdnn_2a.sh
%WER 8.64 [ 9050 / 104765, 349 ins, 521 del, 8180 sub ] exp/nnet3/tdnn_sp/decode_test/cer_15_0.5
%WER 8.72 [ 9135 / 104765, 367 ins, 422 del, 8346 sub ] exp/nnet3/tdnn_sp_online/decode_test/cer_12_1.0
%WER 9.36 [ 9807 / 104765, 386 ins, 441 del, 8980 sub ] exp/nnet3/tdnn_sp_online/decode_test_per_utt/cer_13_1.0

# chain with online pitch, local/chain/tuning/run_tdnn_2a.sh
%WER 7.45 [ 7807 / 104765, 340 ins, 497 del, 6970 sub ] exp/chain/tdnn_2a_sp/decode_test/cer_11_0.5
%WER 7.43 [ 7780 / 104765, 341 ins, 469 del, 6970 sub ] exp/chain/tdnn_2a_sp_online/decode_test/cer_11_0.5
%WER 7.92 [ 8296 / 104765, 384 ins, 472 del, 7440 sub ] exp/chain/tdnn_2a_sp_online/decode_test_per_utt/cer_11_0.5
4 changes: 4 additions & 0 deletions egs/aishell/s5/conf/online_pitch.conf
Original file line number Diff line number Diff line change
@@ -0,0 +1,4 @@
--sample-frequency=16000
--simulate-first-pass-online=true
--normalization-right-context=25
--frames-per-chunk=10
18 changes: 3 additions & 15 deletions egs/aishell/s5/local/aishell_prepare_dict.sh
Original file line number Diff line number Diff line change
Expand Up @@ -15,21 +15,9 @@ mkdir -p $dict_dir
cp $res_dir/lexicon.txt $dict_dir

cat $dict_dir/lexicon.txt | awk '{ for(n=2;n<=NF;n++){ phones[$n] = 1; }} END{for (p in phones) print p;}'| \
sort -u |\
perl -e '
my %ph_cl;
while (<STDIN>) {
$phone = $_;
chomp($phone);
chomp($_);
$phone = $_;
next if ($phone eq "sil");
if (exists $ph_cl{$phone}) { push(@{$ph_cl{$phone}}, $_) }
else { $ph_cl{$phone} = [$_]; }
}
foreach $key ( keys %ph_cl ) {
print "@{ $ph_cl{$key} }\n"
}
perl -e 'while(<>){ chomp($_); $phone = $_; next if ($phone eq "sil");
m:^([^\d]+)(\d*)$: || die "Bad phone $_"; $q{$1} .= "$phone "; }
foreach $l (values %q) {print "$l\n";}
' | sort -k1 > $dict_dir/nonsilence_phones.txt || exit 1;

echo sil > $dict_dir/silence_phones.txt
Expand Down
2 changes: 1 addition & 1 deletion egs/aishell/s5/local/chain/tuning/run_tdnn_1a.sh
Original file line number Diff line number Diff line change
Expand Up @@ -90,7 +90,7 @@ if [ $stage -le 10 ]; then
echo "$0: creating neural net configs using the xconfig parser";

num_targets=$(tree-info $treedir/tree |grep num-pdfs|awk '{print $2}')
learning_rate_factor=$(echo "print 0.5/$xent_regularize" | python)
learning_rate_factor=$(echo "print (0.5/$xent_regularize)" | python)

mkdir -p $dir/configs
cat <<EOF > $dir/configs/network.xconfig
Expand Down
211 changes: 211 additions & 0 deletions egs/aishell/s5/local/chain/tuning/run_tdnn_2a.sh
Original file line number Diff line number Diff line change
@@ -0,0 +1,211 @@
#!/bin/bash

# This script is based on run_tdnn_1a.sh.
# This setup used online pitch to train the neural network.
# It requires a online_pitch.conf in the conf dir.

set -e

# configs for 'chain'
affix=
stage=0
train_stage=-10
get_egs_stage=-10
dir=exp/chain/tdnn_2a # Note: _sp will get added to this
decode_iter=

# training options
num_epochs=4
initial_effective_lrate=0.001
final_effective_lrate=0.0001
max_param_change=2.0
final_layer_normalize_target=0.5
num_jobs_initial=2
num_jobs_final=12
minibatch_size=128
frames_per_eg=150,110,90
remove_egs=true
common_egs_dir=
xent_regularize=0.1

# End configuration section.
echo "$0 $@" # Print the command line for logging

. ./cmd.sh
. ./path.sh
. ./utils/parse_options.sh

if ! cuda-compiled; then
cat <<EOF && exit 1
This script is intended to be used with GPUs but you have not compiled Kaldi with CUDA
If you want to use GPUs (and have them), go to src/, and configure and make on a machine
where "nvcc" is installed.
EOF
fi

# The iVector-extraction and feature-dumping parts are the same as the standard
# nnet3 setup, and you can skip them by setting "--stage 8" if you have already
# run those things.

dir=${dir}${affix:+_$affix}_sp
train_set=train_sp
ali_dir=exp/tri5a_sp_ali
treedir=exp/chain/tri6_7d_tree_sp
lang=data/lang_chain


# if we are using the speed-perturbed data we need to generate
# alignments for it.
local/nnet3/run_ivector_common.sh --stage $stage --online true || exit 1;

if [ $stage -le 7 ]; then
# Get the alignments as lattices (gives the LF-MMI training more freedom).
# use the same num-jobs as the alignments
nj=$(cat $ali_dir/num_jobs) || exit 1;
steps/align_fmllr_lats.sh --nj $nj --cmd "$train_cmd" data/$train_set \
data/lang exp/tri5a exp/tri5a_sp_lats
rm exp/tri5a_sp_lats/fsts.*.gz # save space
fi

if [ $stage -le 8 ]; then
# Create a version of the lang/ directory that has one state per phone in the
# topo file. [note, it really has two states.. the first one is only repeated
# once, the second one has zero or more repeats.]
rm -rf $lang
cp -r data/lang $lang
silphonelist=$(cat $lang/phones/silence.csl) || exit 1;
nonsilphonelist=$(cat $lang/phones/nonsilence.csl) || exit 1;
# Use our special topology... note that later on may have to tune this
# topology.
steps/nnet3/chain/gen_topo.py $nonsilphonelist $silphonelist >$lang/topo
fi

if [ $stage -le 9 ]; then
# Build a tree using our new topology. This is the critically different
# step compared with other recipes.
steps/nnet3/chain/build_tree.sh --frame-subsampling-factor 3 \
--context-opts "--context-width=2 --central-position=1" \
--cmd "$train_cmd" 5000 data/$train_set $lang $ali_dir $treedir
fi

if [ $stage -le 10 ]; then
echo "$0: creating neural net configs using the xconfig parser";

num_targets=$(tree-info $treedir/tree |grep num-pdfs|awk '{print $2}')
learning_rate_factor=$(echo "print (0.5/$xent_regularize)" | python)

mkdir -p $dir/configs
cat <<EOF > $dir/configs/network.xconfig
input dim=100 name=ivector
input dim=43 name=input

# please note that it is important to have input layer with the name=input
# as the layer immediately preceding the fixed-affine-layer to enable
# the use of short notation for the descriptor
fixed-affine-layer name=lda input=Append(-1,0,1,ReplaceIndex(ivector, t, 0)) affine-transform-file=$dir/configs/lda.mat

# the first splicing is moved before the lda layer, so no splicing here
relu-batchnorm-layer name=tdnn1 dim=625
relu-batchnorm-layer name=tdnn2 input=Append(-1,0,1) dim=625
relu-batchnorm-layer name=tdnn3 input=Append(-1,0,1) dim=625
relu-batchnorm-layer name=tdnn4 input=Append(-3,0,3) dim=625
relu-batchnorm-layer name=tdnn5 input=Append(-3,0,3) dim=625
relu-batchnorm-layer name=tdnn6 input=Append(-3,0,3) dim=625

## adding the layers for chain branch
relu-batchnorm-layer name=prefinal-chain input=tdnn6 dim=625 target-rms=0.5
output-layer name=output include-log-softmax=false dim=$num_targets max-change=1.5

# adding the layers for xent branch
# This block prints the configs for a separate output that will be
# trained with a cross-entropy objective in the 'chain' models... this
# has the effect of regularizing the hidden parts of the model. we use
# 0.5 / args.xent_regularize as the learning rate factor- the factor of
# 0.5 / args.xent_regularize is suitable as it means the xent
# final-layer learns at a rate independent of the regularization
# constant; and the 0.5 was tuned so as to make the relative progress
# similar in the xent and regular final layers.
relu-batchnorm-layer name=prefinal-xent input=tdnn6 dim=625 target-rms=0.5
output-layer name=output-xent dim=$num_targets learning-rate-factor=$learning_rate_factor max-change=1.5

EOF
steps/nnet3/xconfig_to_configs.py --xconfig-file $dir/configs/network.xconfig --config-dir $dir/configs/
fi

if [ $stage -le 11 ]; then
if [[ $(hostname -f) == *.clsp.jhu.edu ]] && [ ! -d $dir/egs/storage ]; then
utils/create_split_dir.pl \
/export/b0{5,6,7,8}/$USER/kaldi-data/egs/aishell-$(date +'%m_%d_%H_%M')/s5c/$dir/egs/storage $dir/egs/storage
fi

steps/nnet3/chain/train.py --stage $train_stage \
--cmd "$decode_cmd" \
--feat.online-ivector-dir exp/nnet3/ivectors_${train_set} \
--feat.cmvn-opts "--norm-means=false --norm-vars=false" \
--chain.xent-regularize $xent_regularize \
--chain.leaky-hmm-coefficient 0.1 \
--chain.l2-regularize 0.00005 \
--chain.apply-deriv-weights false \
--chain.lm-opts="--num-extra-lm-states=2000" \
--egs.dir "$common_egs_dir" \
--egs.stage $get_egs_stage \
--egs.opts "--frames-overlap-per-eg 0" \
--egs.chunk-width $frames_per_eg \
--trainer.num-chunk-per-minibatch $minibatch_size \
--trainer.frames-per-iter 1500000 \
--trainer.num-epochs $num_epochs \
--trainer.optimization.num-jobs-initial $num_jobs_initial \
--trainer.optimization.num-jobs-final $num_jobs_final \
--trainer.optimization.initial-effective-lrate $initial_effective_lrate \
--trainer.optimization.final-effective-lrate $final_effective_lrate \
--trainer.max-param-change $max_param_change \
--cleanup.remove-egs $remove_egs \
--feat-dir data/${train_set}_hires_online \
--tree-dir $treedir \
--lat-dir exp/tri5a_sp_lats \
--dir $dir || exit 1;
fi

if [ $stage -le 12 ]; then
# Note: it might appear that this $lang directory is mismatched, and it is as
# far as the 'topo' is concerned, but this script doesn't read the 'topo' from
# the lang directory.
utils/mkgraph.sh --self-loop-scale 1.0 data/lang_test $dir $dir/graph
fi

graph_dir=$dir/graph
if [ $stage -le 13 ]; then
for test_set in dev test; do
steps/nnet3/decode.sh --acwt 1.0 --post-decode-acwt 10.0 \
--nj 10 --cmd "$decode_cmd" \
--online-ivector-dir exp/nnet3/ivectors_$test_set \
$graph_dir data/${test_set}_hires_online $dir/decode_${test_set} || exit 1;
done
fi

if [ $stage -le 14 ]; then
steps/online/nnet3/prepare_online_decoding.sh --mfcc-config conf/mfcc_hires.conf \
--add-pitch true \
$lang exp/nnet3/extractor "$dir" ${dir}_online || exit 1;
fi

dir=${dir}_online
if [ $stage -le 15 ]; then
for test_set in dev test; do
steps/online/nnet3/decode.sh --acwt 1.0 --post-decode-acwt 10.0 \
--nj 10 --cmd "$decode_cmd" \
--config conf/decode.config \
$graph_dir data/${test_set}_hires_online $dir/decode_${test_set} || exit 1;
done
fi

if [ $stage -le 16 ]; then
for test_set in dev test; do
steps/online/nnet3/decode.sh --acwt 1.0 --post-decode-acwt 10.0 \
--nj 10 --cmd "$decode_cmd" --per-utt true \
--config conf/decode.config \
$graph_dir data/${test_set}_hires_online $dir/decode_${test_set}_per_utt || exit 1;
done
fi

exit;
23 changes: 14 additions & 9 deletions egs/aishell/s5/local/nnet3/run_ivector_common.sh
Original file line number Diff line number Diff line change
Expand Up @@ -14,7 +14,7 @@ stage=0
train_set=train
test_sets="dev test"
gmm=tri5a

online=false
nnet3_affix=

. ./cmd.sh
Expand All @@ -31,6 +31,11 @@ for f in data/${train_set}/feats.scp ${gmm_dir}/final.mdl; do
fi
done

online_affix=
if [ $online = true ]; then
online_affix=_online
fi

if [ $stage -le 1 ]; then
# Although the nnet will be trained by high resolution data, we still have to
# perturb the normal data to get the alignment _sp stands for speed-perturbed
Expand All @@ -54,26 +59,26 @@ if [ $stage -le 3 ]; then
# Create high-resolution MFCC features (with 40 cepstra instead of 13).
# this shows how you can split across multiple file-systems.
echo "$0: creating high-resolution MFCC features"
mfccdir=mfcc_perturbed_hires
mfccdir=mfcc_perturbed_hires$online_affix
if [[ $(hostname -f) == *.clsp.jhu.edu ]] && [ ! -d $mfccdir/storage ]; then
utils/create_split_dir.pl /export/b0{5,6,7,8}/$USER/kaldi-data/mfcc/aishell-$(date +'%m_%d_%H_%M')/s5/$mfccdir/storage $mfccdir/storage
fi

for datadir in ${train_set}_sp ${test_sets}; do
utils/copy_data_dir.sh data/$datadir data/${datadir}_hires
utils/copy_data_dir.sh data/$datadir data/${datadir}_hires$online_affix
done

# do volume-perturbation on the training data prior to extracting hires
# features; this helps make trained nnets more invariant to test data volume.
utils/data/perturb_data_dir_volume.sh data/${train_set}_sp_hires || exit 1;
utils/data/perturb_data_dir_volume.sh data/${train_set}_sp_hires$online_affix || exit 1;

for datadir in ${train_set}_sp ${test_sets}; do
steps/make_mfcc_pitch.sh --nj 10 --mfcc-config conf/mfcc_hires.conf \
--cmd "$train_cmd" data/${datadir}_hires exp/make_hires/$datadir $mfccdir || exit 1;
steps/compute_cmvn_stats.sh data/${datadir}_hires exp/make_hires/$datadir $mfccdir || exit 1;
utils/fix_data_dir.sh data/${datadir}_hires || exit 1;
steps/make_mfcc_pitch$online_affix.sh --nj 10 --mfcc-config conf/mfcc_hires.conf \
--cmd "$train_cmd" data/${datadir}_hires$online_affix exp/make_hires/$datadir $mfccdir || exit 1;
steps/compute_cmvn_stats.sh data/${datadir}_hires$online_affix exp/make_hires/$datadir $mfccdir || exit 1;
utils/fix_data_dir.sh data/${datadir}_hires$online_affix || exit 1;
# create MFCC data dir without pitch to extract iVector
utils/data/limit_feature_dim.sh 0:39 data/${datadir}_hires data/${datadir}_hires_nopitch || exit 1;
utils/data/limit_feature_dim.sh 0:39 data/${datadir}_hires$online_affix data/${datadir}_hires_nopitch || exit 1;
steps/compute_cmvn_stats.sh data/${datadir}_hires_nopitch exp/make_hires/$datadir $mfccdir || exit 1;
done
fi
Expand Down
Loading