Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
420 commits
Select commit Hold shift + click to select a range
39958f7
Merge pull request #485 from naxingyu/add-chain-rm
danpovey Feb 5, 2016
f18638f
Finishing some things relating to kStrideEqualNumCols and flags for c…
danpovey Feb 5, 2016
7af7ba5
Merge branch 'chain' of github.com:/kaldi-asr/kaldi into chain
danpovey Feb 5, 2016
ae15da0
Merge remote-tracking branch 'origin/chain-general-component' into chain
danpovey Feb 5, 2016
1e4a97d
bug-fix for CUDA-related code of kStrideEqualNumCols
danpovey Feb 5, 2016
f63270a
Added a tool to compute the confidence interval of WER
Feb 5, 2016
d037050
bug-fixes regarding optimization code handling of kStrideEqualNumCols.
danpovey Feb 5, 2016
4d4cf4c
new scripts, and script changes, testing statistics-averaging
danpovey Feb 5, 2016
9bf7062
Adding steps/cleanup/find_bad_utts_nnet.sh, nnet2 version of steps/cl…
realill Feb 5, 2016
d26522d
Merge pull request #488 from realill/master
danpovey Feb 5, 2016
44db487
Merge pull request #460 from tomkocse/3dmaxpooling
danpovey Feb 6, 2016
bccf2bd
cosmatic change for swbd nnet2 recipe
naxingyu Feb 6, 2016
6844446
modified NonlinearComponent::Info().
pegahgh Feb 6, 2016
8ce0ae1
Merge pull request #489 from naxingyu/cosmatic-swbd-nnet2
danpovey Feb 6, 2016
b86aeb3
Merge pull request #482 from pegahgh/chain-normalization-with-log
danpovey Feb 6, 2016
8bf234d
bug-fix to backprop of NormalizeComponent; new script; improvement to…
danpovey Feb 6, 2016
48c07f3
preliminary work relating to statistics-pooling and statistics-extrac…
danpovey Jan 21, 2016
e3b834b
adding more code for statistics-extraction component.
danpovey Jan 23, 2016
b2101fb
Further progress toward finishing statistics-extraction-component and…
danpovey Jan 29, 2016
cf73f14
Testing code and various bug fixes relating to statistics-extraction…
danpovey Jan 30, 2016
3aabb61
chain branch: adding script-level support, and example script, for st…
danpovey Jan 31, 2016
82b5b95
bug-fixes and example-script changes relating to moving-average stati…
danpovey Feb 1, 2016
9d8133b
regarding statistics-averaging component: script bug fix regarding or…
danpovey Feb 3, 2016
7c3d324
Some partial work towards efficient convolution in time. May rewrite…
danpovey Feb 3, 2016
7a161bf
removing some draft code regarding convolution in time that I won't i…
danpovey Feb 4, 2016
c0bb04e
Finishing some things relating to kStrideEqualNumCols and flags for c…
danpovey Feb 5, 2016
59df267
bug-fix for CUDA-related code of kStrideEqualNumCols
danpovey Feb 5, 2016
faa18ae
bug-fixes regarding optimization code handling of kStrideEqualNumCols.
danpovey Feb 5, 2016
cc7be76
new scripts, and script changes, testing statistics-averaging
danpovey Feb 5, 2016
58252ae
modified cmd.sh again
Feb 6, 2016
f5307d1
Merge pull request #475 from keli78/test
danpovey Feb 6, 2016
ea0cbdc
Merge remote-tracking branch 'upstream/master' into chain
danpovey Feb 6, 2016
2a51b02
refactoring, fixing and cleanup of component NormalizeComponent
danpovey Feb 6, 2016
6ad7a10
Adding ConstantFunctionComponent
danpovey Feb 7, 2016
5df3686
some script changes; testing improved randomization of frame-shift.
danpovey Feb 7, 2016
d916174
minor bug-fix to prevent compiler warning
danpovey Feb 7, 2016
0b86fd6
removing 'notes' file from chain branch
danpovey Feb 7, 2016
5df81f2
Merge branch 'chain-general-component-temp' into chain-temp5
danpovey Feb 7, 2016
5d7d87d
some script changes; testing improved randomization of frame-shift.
danpovey Feb 7, 2016
b2dbc17
Merge commit '5df3686' into chain-merged
danpovey Feb 7, 2016
1427e7a
merging changes in chain branch with some local changes I've been mak…
danpovey Feb 7, 2016
fe70a63
fix to bug in test code introduced by merge
danpovey Feb 7, 2016
692dd89
bug-fix to steps/nnet3/decode.sh
danpovey Feb 9, 2016
9291472
Merge remote-tracking branch 'upstream/master' into chain
danpovey Feb 9, 2016
bbcc66e
added I/O interfaces for NnetComputation and ComputationRequest.
xiaohui-zhang Feb 8, 2016
d27c96b
fixed a bug
xiaohui-zhang Feb 9, 2016
a9b6513
fix to validate_lang.pl regarding disambiguation symbols, and associa…
danpovey Feb 10, 2016
0b09531
Cleaned the style and new functionality
Feb 10, 2016
46adeed
minor fixes
xiaohui-zhang Feb 10, 2016
81e8389
Merge pull request #491 from xiaohui-zhang/chain
danpovey Feb 11, 2016
751af96
fixed bugs in steps/nnet3/lstm/train.sh from PR #491.
xiaohui-zhang Feb 11, 2016
4a3d499
a minor fix.
xiaohui-zhang Feb 11, 2016
60df952
Merge pull request #493 from xiaohui-zhang/chain
danpovey Feb 11, 2016
117c075
nnet1: added the removal of frames with 0 confidence from mini-batches
KarelVesely84 Feb 11, 2016
760a400
Adding self-repair code for ReLUs (to fix over- or under-saturated Re…
danpovey Feb 10, 2016
9bcd17a
some code and script to support higher-frame-rate feature input for c…
danpovey Feb 10, 2016
07952a6
some extra tuning script for chain models in switchboard setup
danpovey Feb 11, 2016
304d4a1
Fix to self-repair code of nonlinear components
danpovey Feb 11, 2016
c6c0f34
Efficiency improvement for ReLU backpropagation; change in default tu…
danpovey Feb 11, 2016
d5cd1d6
script changes to support caching of computations when training chain…
danpovey Feb 11, 2016
14954a0
adding some new tuning scripts; some script-level fixes/updates.
danpovey Feb 12, 2016
dd10008
some tuning-script updates for chain models on switchboard
danpovey Feb 12, 2016
57cfb24
Change to StoreStats functions of tanh, sigmoid and relu components t…
danpovey Feb 12, 2016
b24f10f
chain branch: Adding some more recent results for swbd tuning
danpovey Feb 12, 2016
3651ca8
some cosmetic improvements to slurm.pl and to the fisher_callhome_spa…
danpovey Feb 12, 2016
032aa24
Minor fix regarding adaptation configs
hhadian Feb 14, 2016
8aa016a
If compiling with MSVS 2015, dont redefine snprintf
DeepHorizons Feb 14, 2016
5439624
Merge pull request #499 from DeepHorizons/snprintf_redefinition_fix
jtrmal Feb 14, 2016
6df8653
Merge pull request #498 from hhadian/patch-1
danpovey Feb 14, 2016
1fa0e18
adding a couple more swbd+chain tuning experiments
danpovey Feb 15, 2016
d093ec0
Kaldified source code and added a system comparison script
Feb 15, 2016
d5b8b23
Modified bash command to avoid wildcard expansion when using phones w…
Feb 15, 2016
15faa6c
Merge pull request #501 from speechmatics/prepare_lang_bash
danpovey Feb 15, 2016
2a862b4
swbd+chain: Add some new example scripts and an associated change in …
danpovey Feb 15, 2016
d25785d
small bug fix for fisher_swbd data prep
naxingyu Feb 16, 2016
d1e1e3d
Merge pull request #502 from naxingyu/bugfix-fisherswbd-swbd-prep
danpovey Feb 16, 2016
e06745d
adding some newer results for swbd+chain tuning; reverting skip-splic…
danpovey Feb 16, 2016
341e0f0
Changes to allow the large spanish word list to be downloaded if not …
noisychannel Feb 17, 2016
b7aa6b1
Small changes. Fixes #494
noisychannel Feb 17, 2016
6fd0056
Merge pull request #508 from noisychannel/fisher_spanish
danpovey Feb 17, 2016
577659a
fix swbd1 data prep duplicates
naxingyu Feb 17, 2016
4bce24c
Merge pull request #509 from naxingyu/swbd1-prep-dup-fix
danpovey Feb 17, 2016
92994e2
Handle multiple pronunciations in lexicon. Fixes #506
noisychannel Feb 17, 2016
a059643
updating 'cmd.sh' for BUT cluster in various recipes,
KarelVesely84 Feb 17, 2016
04cd902
fixing tidigits data preparation,
KarelVesely84 Feb 17, 2016
77395a9
Merge pull request #510 from noisychannel/fisher_spanish
danpovey Feb 17, 2016
2646cfb
chain+swbd experiments: tuning-experiment results
danpovey Feb 17, 2016
187fa16
chain branch: changing the self-repair code so that it should work we…
danpovey Feb 17, 2016
ea98c12
Merge pull request #511 from vesis84/tidigits_debug
danpovey Feb 18, 2016
7bc34fe
cosmetic change: fix 'score' to 'cost'
danpovey Feb 18, 2016
b3bbc03
modifying cmd.sh in example recipes to encourage the use of new-style…
danpovey Feb 18, 2016
dbb028f
clarifying configuration process for CUDA, and give prototype Makefil…
danpovey Feb 19, 2016
90af624
chain branch: bug-fix in self-repair code for sigmoid units
danpovey Feb 19, 2016
bcbfca0
Merge remote-tracking branch 'upstream/master' into chain
danpovey Feb 20, 2016
e980e75
small cosmetic change to RM example script for chain models
danpovey Feb 20, 2016
bccc82e
Fix processing of phonemes when adding position dependency symbols:
OliverWalter Feb 21, 2016
5beecf2
Merge pull request #513 from OliverWalter/master
danpovey Feb 21, 2016
093feb6
Merge remote-tracking branch 'upstream/master' into chain
danpovey Feb 22, 2016
4bd255f
Added plotting feature to the steps/nnet3/report scripts. Refactored …
vijayaditya Feb 22, 2016
d27c4b3
Merge pull request #514 from vijayaditya/report_refactor
danpovey Feb 22, 2016
29ed1f2
chain+swbd experiments: new tuning scripts, some changes to jesus con…
danpovey Feb 22, 2016
71b3009
removing unhelpful feature in make_jesus_configs.py
danpovey Feb 22, 2016
075e471
steps/nnet3/report: Added support for comparing multiple experiments
vijayaditya Feb 22, 2016
adb0c26
steps/nnet3/report : Adding latex report generation
vijayaditya Feb 22, 2016
745e775
bug fix to FeatureFbank::Dim()
danpovey Feb 22, 2016
e08fc17
Fix warnings on undefined values. $ACT is optional and there is no ne…
nshmyrev Feb 22, 2016
3db63ff
Merge pull request #518 from cmusphinx/master
danpovey Feb 22, 2016
f024c75
Merge pull request #516 from vijayaditya/report_refactor
danpovey Feb 22, 2016
db16052
Added train_rnn.py, nnet3_train_lib.py and modified steps/nnet3/lstm/…
vijayaditya Feb 7, 2016
05703ea
Merge pull request #490 from vijayaditya/new_train
danpovey Feb 22, 2016
3e73f67
some new tuning experiments on chain+swbd setup; add --xent-separate-…
danpovey Feb 22, 2016
5a9397e
minor bug fix to accommodate changes in PR #514
vijayaditya Feb 22, 2016
1de12bf
Merge pull request #519 from vijayaditya/report_refactor
danpovey Feb 22, 2016
12b3396
steps/nnet3/report : modified regex to accept negative mean values an…
vijayaditya Feb 23, 2016
32f4d94
Merge pull request #521 from vijayaditya/report_refactor
danpovey Feb 23, 2016
911dfed
Update make_configs.py
vimalmanohar Feb 23, 2016
30e3b42
Merge pull request #522 from vimalmanohar/patch-1
danpovey Feb 23, 2016
0b720b8
steps/nnet3/report : minor cosmetic fix
vijayaditya Feb 23, 2016
1eb8076
Merge pull request #523 from vijayaditya/report_refactor
danpovey Feb 23, 2016
005e64b
Merge remote-tracking branch 'upstream/master' into chain
danpovey Feb 23, 2016
5b88f67
Some clarifying changes to comments
danpovey Feb 24, 2016
c524b8e
chain branch: merging some script changes/improvements from utils/ di…
danpovey Feb 24, 2016
cc0dcef
swbd+chain: miscellaneous new tuning scripts and bug-fixes
danpovey Feb 24, 2016
22e221a
chain+swbd expts: update to script; and minor fix to get_frame_shift.sh
danpovey Feb 24, 2016
1afcda8
minor correction in the way arguments are provided to python training…
vijayaditya Feb 25, 2016
ee4f4d4
Merge pull request #524 from vijayaditya/report_refactor
danpovey Feb 25, 2016
90769ef
fixes to debug_lexicon.sh script; fixes to TEDLIUM scripts (<unk> was…
danpovey Feb 25, 2016
05b94ce
fixes to debug_lexicon.sh script; fixes to TEDLIUM scripts (<unk> was…
danpovey Feb 25, 2016
dd0fe67
fix problem in tedlium recipe introduced by previous fix
danpovey Feb 25, 2016
c61d99a
Merge remote-tracking branch 'upstream/master' into chain
danpovey Feb 25, 2016
09673ef
Documenting the behavior of compute-atwv correctly, as suggested by V…
jtrmal Feb 25, 2016
890cc05
Improved documentation for compute-atwv.cc
jtrmal Feb 25, 2016
bd87207
Improved documentation on treatment of the csv alignment
jtrmal Feb 25, 2016
cf17987
Merge pull request #527 from jtrmal/compute-atwv-fix2
danpovey Feb 25, 2016
7cc032f
fix to minor bug in create_split_dir.pl (would affect no recipes)
danpovey Feb 26, 2016
fe49d90
Added steps/nnet3/train_dnn.py and made changes to swbd/tdnn recipe t…
vijayaditya Feb 25, 2016
1d3d854
steps/nnet3/train_*.py modified the name of the option command
vijayaditya Feb 26, 2016
415d5c7
Merge pull request #528 from vijayaditya/new_train
danpovey Feb 26, 2016
dce72c1
chain+swbd experiments: changes to tuning scripts
danpovey Feb 26, 2016
e190c61
Merge remote-tracking branch 'upstream/master' into chain
danpovey Feb 26, 2016
ae8e69b
nnet1: bugfix in NN training script (seeding of nnet-inititalize)
KarelVesely84 Feb 27, 2016
2f66de7
nnet3: Added self-repair option for LSTMs, modified swbd lstm recipe …
vijayaditya Feb 27, 2016
b9b728a
Merge pull request #529 from vijayaditya/new_train2
danpovey Feb 27, 2016
f70a416
Added python scripts for chain model training
vijayaditya Feb 25, 2016
950028c
minor modification to python chain script
vijayaditya Feb 27, 2016
f8e9ba2
Merge pull request #525 from vijayaditya/chain_new_train2
danpovey Feb 27, 2016
f463444
added compute-filter.cc binary, which computes LP filters per speaker.
pegahgh Feb 28, 2016
190c015
nnet3/tdnn : reduced reliance on numpy and scipy
vijayaditya Feb 28, 2016
89bef1e
Merge pull request #531 from vijayaditya/remove_numpy_req
danpovey Feb 28, 2016
8f9db59
bug-fix to use self-repair option in LSTMs
vijayaditya Feb 28, 2016
29481e8
Changes from master to enable automerge
vijayaditya Feb 28, 2016
70801e0
Merge pull request #532 from vijayaditya/remove_numpy_req
danpovey Feb 28, 2016
6f2662a
Merge pull request #533 from vijayaditya/chain_new_train2
danpovey Feb 28, 2016
0466014
chain branch: merging changes from upstream/master
danpovey Feb 28, 2016
7e06792
Merge branch 'chain' of https://github.com/kaldi-asr/kaldi into chain
danpovey Feb 28, 2016
a3694a4
Add chain library to nnet3bin dependencies
akreal Feb 28, 2016
1ba06bf
Merge pull request #534 from akreal/master
danpovey Feb 28, 2016
8d09b03
swbd/local/chain/: lstm recipe parameter modification. Using momentum…
vijayaditya Feb 29, 2016
b2ca6f6
Fix <unk> bug in TEDLIUM Cantab Research language model.
galv Feb 28, 2016
84ce638
add utils/perturb_data_signal.sh
tomkocse Feb 29, 2016
cbfa751
added --spk2utt option to compute-filter to consider computing speake…
pegahgh Feb 29, 2016
437849b
Merge pull request #537 from galv/tedlium-lm-fix
danpovey Feb 29, 2016
92a63f6
Merge pull request #535 from vijayaditya/chain_new_train2
danpovey Feb 29, 2016
a540543
modified compute-filter.cc and fixed some issues.
pegahgh Mar 1, 2016
737102f
Merge remote-tracking branch 'upstream/master' into chain
danpovey Mar 1, 2016
61dc4bb
Merge pull request #4 from tomkocse/chain-signal-perturb
pegahgh Mar 2, 2016
dcae01c
added apply-filter and fixed compute-filter.
pegahgh Mar 2, 2016
78ab2bf
added utils/perturb_data_signal.sh written by Tom Ko.
pegahgh Mar 2, 2016
eb615eb
Merge branch 'chain-signal-perturb' of https://github.com/pegahgh/kal…
pegahgh Mar 2, 2016
cf93278
minor bug fix in egs/*/*/local/*/run_lstm.sh scripts
vijayaditya Mar 2, 2016
9a3a4f1
Merge pull request #541 from vijayaditya/remove_numpy_req
danpovey Mar 2, 2016
81cfaff
master: some fixes to chain scripts to fix problems encountered by Ilya
danpovey Mar 2, 2016
9fc8834
Merge remote-tracking branch 'upstream/master' into chain
danpovey Mar 2, 2016
6092808
chain branch: updates to chain+swbd tuning scripts; cosmetic change t…
danpovey Mar 2, 2016
54a90f6
some minor cosmetic improvements to TEDLIUM recipe; improvements to c…
danpovey Mar 2, 2016
d77acd8
bug fix in nnet3 lstms
vijayaditya Mar 3, 2016
775fe3f
Merge pull request #542 from vijayaditya/remove_numpy_req
danpovey Mar 3, 2016
7d72974
small fixes/improvements to debug_lexicon.sh
danpovey Mar 3, 2016
82c5cc5
adding program lattice-arc-post
danpovey Mar 3, 2016
ebb239b
chain branch: fixes to lattice-arc-post.cc
danpovey Mar 3, 2016
2cfc989
Raw nnet3 training
vimalmanohar Mar 2, 2016
1148b4e
Additional information about Intel MKL and NVIDIA CUDA
ceefour Mar 3, 2016
771663f
Use curl instead of wget
ceefour Mar 3, 2016
03fbab2
Merge pull request #545 from ceefour/patch-3
jtrmal Mar 3, 2016
915d676
Fix directory name
ceefour Mar 3, 2016
6d58ce3
Merge pull request #544 from ceefour/patch-2
jtrmal Mar 3, 2016
c141c80
Merge pull request #546 from ceefour/patch-4
jtrmal Mar 3, 2016
94dcf92
add prefix argument in utils/perturb_data_signal.sh.
tomkocse Mar 3, 2016
fb16a92
Merge pull request #5 from tomkocse/chain-signal-perturb
pegahgh Mar 3, 2016
19c0615
raw_nnet: Some fixes
vimalmanohar Mar 3, 2016
2caf161
modify the way of print out
tomkocse Mar 3, 2016
77318db
use random filter if can't find speaker's own filter
tomkocse Mar 3, 2016
dd24876
bug fix in lattice-arc-post: apply acoustic and language model scale.
danpovey Mar 3, 2016
e250573
raw_nnet: Moved config generation to run script
vimalmanohar Mar 3, 2016
8d14382
raw_nnet: Minor fix
vimalmanohar Mar 3, 2016
cdcd92b
raw_nnet: Removed some unnecessary checks
vimalmanohar Mar 3, 2016
4a6daa1
raw_nnet: Added an option for configs_dir
vimalmanohar Mar 3, 2016
dc54009
minor bug-fix in steps/nnet3/chain/train_tdnn.sh (won't affect results)
danpovey Mar 3, 2016
e5a66f5
Merge remote-tracking branch 'upstream/master' into chain
danpovey Mar 3, 2016
babbb47
Merge pull request #6 from tomkocse/chain-signal-perturb
pegahgh Mar 3, 2016
6c59f5d
raw_nnet: Removed some unnecessary stuff from steps/nnet3/tdnn/train_…
vimalmanohar Mar 4, 2016
2ebfcf5
raw_nnet: Minor cosmetic change
vimalmanohar Mar 4, 2016
fbedf1d
raw_nnet: Added autoencoder recipe in aspire
vimalmanohar Mar 4, 2016
e0b3705
Merge pull request #539 from vimalmanohar/raw_nnet
danpovey Mar 4, 2016
4ecee3f
added simple TDNN comparison run_tdnn_6v.sh for jesus models run_tdnn…
vijayaditya Mar 4, 2016
6708551
Merge pull request #548 from vijayaditya/tdnn_chain
danpovey Mar 4, 2016
29e0aaf
minor fix to chain/train.py
vijayaditya Mar 4, 2016
bd884df
Merge pull request #549 from vijayaditya/tdnn_chain
danpovey Mar 4, 2016
596b382
use recording id to search the filter
tomkocse Mar 4, 2016
d2591c0
Merge pull request #7 from tomkocse/chain-signal-perturb
pegahgh Mar 4, 2016
e3410bc
Renamed, tweaked and added options to bootci WER
Mar 4, 2016
2590476
Merge remote-tracking branch 'upstream/master' into chain
danpovey Mar 4, 2016
3191fcd
chain branch: various new tuning script; some minor code improvements
danpovey Mar 4, 2016
3432509
new tuning scripts for swbd+chain; cosmetic change to chain/train.py
danpovey Mar 4, 2016
b706a96
Merge remote-tracking branch 'upstream/chain' into master (periodic s…
danpovey Mar 4, 2016
c3c4ccd
Merge pull request #487 from nsmartinez/compute-confidence-interval-wer
jtrmal Mar 4, 2016
9a0d071
some swbd+chain tuning script updates; minor cosmetic fixes to chain …
danpovey Mar 5, 2016
7219ee0
some fixes to nnet3 and chain python scripts; new tuning script.
danpovey Mar 5, 2016
d9266e7
Merge remote-tracking branch 'upstream/master' into chain
danpovey Mar 5, 2016
c36bd73
removed bugs introduced due automerge with master
vijayaditya Mar 5, 2016
7e8259c
Merge pull request #557 from vijayaditya/chain_remove
danpovey Mar 5, 2016
8156912
chain model training: fix to steps/nnet3/tdnn/make_configs.py to appl…
danpovey Mar 5, 2016
5f47da7
swbd : Added lstm chain recipe based on tdnn_6h
vijayaditya Mar 6, 2016
8c6d814
Merge pull request #561 from vijayaditya/chain_remove
danpovey Mar 6, 2016
1a7613c
chain+swbd: new tuning scripts
danpovey Mar 6, 2016
2a91756
chain scripts: fixing script-level default of right-tolerancde from 1…
danpovey Mar 6, 2016
7017322
changes to interfaces for how learning-rates are set, for greater cla…
danpovey Mar 6, 2016
e0e2065
sequence: Minor changes required for sequence training
vimalmanohar Mar 5, 2016
2c9bff8
Merge pull request #560 from vimalmanohar/sequence_changes
danpovey Mar 7, 2016
c928063
chain+swbd: some more recent tuning scripts
danpovey Mar 7, 2016
dd7cb32
swbd : Added blstm + chain recipe, results pending
vijayaditya Mar 7, 2016
76af6b6
Merge pull request #564 from vijayaditya/chain_remove
danpovey Mar 7, 2016
d75764e
sequence: nnet3 discriminative code added
vimalmanohar Jan 31, 2016
394ae72
add utils/perturb_data_signal_v2.sh
tomkocse Mar 8, 2016
8a15945
Merge pull request #8 from tomkocse/chain-signal-perturb
pegahgh Mar 8, 2016
d6ba876
exposed --frame-shift option at the script level in get_ctm.sh and ge…
danpovey Mar 8, 2016
7d66712
sequence: Restructured code to reduce some configs and rename some cl…
vimalmanohar Mar 9, 2016
4e3267a
Merge pull request #563 from vimalmanohar/sequence_changes
danpovey Mar 9, 2016
c1dc87c
fixed some issue with compute-filter and modified scripts.
pegahgh Mar 9, 2016
6932765
swbd+chain+lstm:updated results after Dan's fix to right-tolerance de…
vijayaditya Mar 9, 2016
ff2f2d0
Merge pull request #570 from vijayaditya/chain_remove
danpovey Mar 9, 2016
8d5c8b4
change to nbest-to-ctm.cc to automatically use higher precision if pr…
danpovey Mar 9, 2016
d5f4f8a
Add options for score.sh to rescoring script.
galv Mar 9, 2016
d930497
Merge pull request #573 from galv/bug-fix/add-rescore-score-options
danpovey Mar 9, 2016
c418561
swbd+chain+blstm: Added results
vijayaditya Mar 10, 2016
0ff500c
Merge pull request #575 from vijayaditya/chain_remove
danpovey Mar 10, 2016
4e48a71
fixed some bugs and added recent scripts.
pegahgh Mar 10, 2016
8c7b789
added changes to signal.*
pegahgh Mar 10, 2016
5de9b1d
Merge branch 'chain' of https://github.com/kaldi-asr/kaldi into chain…
pegahgh Mar 10, 2016
cf35c6c
added some changes
pegahgh Apr 5, 2016
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
The table of contents is too big for display.
Diff view
Diff view
  •  
  •  
  •  
2 changes: 2 additions & 0 deletions .gitattributes
Original file line number Diff line number Diff line change
Expand Up @@ -15,4 +15,6 @@ windows/INSTALL* eol=native
windows/NewGuidCmd.exe.config text eol=crlf
windows/NewGuidCmd.exe binary

# Prevent git changing CR-LF to LF when archiving (patch requires CR-LF on Windows).
**/*.patch -text

3 changes: 2 additions & 1 deletion .gitignore
Original file line number Diff line number Diff line change
Expand Up @@ -6,11 +6,12 @@
!/src/*/Makefile
!/src/*/README

# Compiled Object files
# Compiled Object files and python ciles
*.slo
*.lo
*.o
*.obj
*.pyc

# Compiled Dynamic libraries
*.so
Expand Down
49 changes: 25 additions & 24 deletions egs/ami/s5/cmd.sh
Original file line number Diff line number Diff line change
@@ -1,37 +1,38 @@
# "queue.pl" uses qsub. The options to it are
# options to qsub. If you have GridEngine installed,
# change this to a queue you have access to.
# Otherwise, use "run.pl", which will run jobs locally
# (make sure your --num-jobs options are no more than
# the number of cpus on your machine.
# you can change cmd.sh depending on what type of queue you are using.
# If you have no queueing system and want to run on a local machine, you
# can change all instances 'queue.pl' to run.pl (but be careful and run
# commands one by one: most recipes will exhaust the memory on your
# machine). queue.pl works with GridEngine (qsub). slurm.pl works
# with slurm. Different queues are configured differently, with different
# queue names and different ways of specifying things like memory;
# to account for these differences you can create and edit the file
# conf/queue.conf to match your queue's configuration. Search for
# conf/queue.conf in http://kaldi-asr.org/doc/queue.html for more information,
# or search for the string 'default_config' in utils/queue.pl or utils/slurm.pl.

export train_cmd="queue.pl --mem 1G"
export decode_cmd="queue.pl --mem 2G"
# the use of cuda_cmd is deprecated but it is sometimes still used in nnet1
# scripts.
export cuda_cmd="queue.pl --gpu 1 --mem 20G"

# the rest of this file is present for historical reasons.
# In general it's best to rely on conf/queue.conf for cluster-specific
# configuration.

# On Eddie use:
#export train_cmd="queue.pl -P inf_hcrc_cstr_nst -l h_rt=08:00:00"
#export decode_cmd="queue.pl -P inf_hcrc_cstr_nst -l h_rt=05:00:00 -pe memory-2G 4"
#export highmem_cmd="queue.pl -P inf_hcrc_cstr_nst -l h_rt=05:00:00 -pe memory-2G 4"
#export scoring_cmd="queue.pl -P inf_hcrc_cstr_nst -l h_rt=00:20:00"

# JSALT2015 workshop, cluster AWS-EC2, (setup from Vijay)
export train_cmd="queue.pl -l arch=*64* --mem 1G"
export decode_cmd="queue.pl -l arch=*64* --mem 2G"
export highmem_cmd="queue.pl -l arch=*64* --mem 4G"
export scoring_cmd="queue.pl -l arch=*64*"
export cuda_cmd="queue.pl --gpu 1 -l mem_free=20G,ram_free=20G"
export cntk_decode_cmd="queue.pl -l arch=*64* --mem 1G -pe smp 2"

# To run locally, use:
#export train_cmd=run.pl
#export decode_cmd=run.pl
#export highmem_cmd=run.pl
#export cuda_cmd=run.pl

if [ "$(hostname -d)" == "fit.vutbr.cz" ]; then
# BUT cluster:
queue="all.q@@blade,all.q@@speech"
gpu_queue="long.q@supergpu*,long.q@dellgpu*,long.q@pcspeech-gpu,long.q@pcgpu*"
gpu_queue="long.q@@gpu"
storage="matylda5"
export train_cmd="queue.pl -q $queue -l ram_free=1500M,mem_free=1500M,${storage}=1"
export decode_cmd="queue.pl -q $queue -l ram_free=2500M,mem_free=2500M,${storage}=0.5"
export train_cmd="queue.pl -q $queue -l ram_free=1.5G,mem_free=1.5G,${storage}=1"
export decode_cmd="queue.pl -q $queue -l ram_free=2.5G,mem_free=2.5G,${storage}=0.5"
export cuda_cmd="queue.pl -q $gpu_queue -l gpu=1"
fi
fi

4 changes: 2 additions & 2 deletions egs/ami/s5/local/nnet3/run_lstm.sh
Original file line number Diff line number Diff line change
Expand Up @@ -18,7 +18,7 @@
stage=0
train_stage=-10
mic=ihm
use_ihm_ali=false
use_ihm_ali=false
use_sat_alignments=false # if true, use tri4a alignments are used
# by default GMM-HMM systems are not built to this stage
# in SDM and MDM systems. So run the tri4a stage if you
Expand Down Expand Up @@ -66,7 +66,7 @@ decode_iter=

echo "$0 $@" # Print the command line for logging

. cmd.sh
. ./cmd.sh
. ./path.sh
. ./utils/parse_options.sh

Expand Down
26 changes: 13 additions & 13 deletions egs/ami/s5/run_ihm.sh
Original file line number Diff line number Diff line change
Expand Up @@ -10,13 +10,13 @@ mic=ihm
stage=0
. utils/parse_options.sh

# Set bash to 'debug' mode, it prints the commands (option '-x') and exits on :
# Set bash to 'debug' mode, it prints the commands (option '-x') and exits on :
# -e 'error', -u 'undefined variable', -o pipefail 'error in pipeline',
set -euxo pipefail

# Path where AMI gets downloaded (or where locally available):
AMI_DIR=$PWD/wav_db # Default,
case $(hostname -d) in
AMI_DIR=$PWD/wav_db # Default,
case $(hostname -d) in
fit.vutbr.cz) AMI_DIR=/mnt/scratch05/iveselyk/KALDI_AMI_WAV ;; # BUT,
clsp.jhu.edu) AMI_DIR=/export/corpora4/ami/amicorpus ;; # JHU,
cstr.ed.ac.uk) AMI_DIR= ;; # Edinburgh,
Expand Down Expand Up @@ -86,7 +86,7 @@ if [ $stage -le 5 ]; then
data/$mic/train data/lang exp/$mic/tri2a exp/$mic/tri2_ali
# Decode,
graph_dir=exp/$mic/tri2a/graph_${LM}
$highmem_cmd $graph_dir/mkgraph.log \
$cmd --mem 4G $graph_dir/mkgraph.log \
utils/mkgraph.sh data/lang_${LM} exp/$mic/tri2a $graph_dir
steps/decode.sh --nj $nj --cmd "$decode_cmd" --config conf/decode.conf \
$graph_dir data/$mic/dev exp/$mic/tri2a/decode_dev_${LM}
Expand All @@ -104,26 +104,26 @@ if [ $stage -le 6 ]; then
data/$mic/train data/lang exp/$mic/tri3a exp/$mic/tri3a_ali
# Decode,
graph_dir=exp/$mic/tri3a/graph_${LM}
$highmem_cmd $graph_dir/mkgraph.log \
$cmd --mem 4G $graph_dir/mkgraph.log \
utils/mkgraph.sh data/lang_${LM} exp/$mic/tri3a $graph_dir
steps/decode.sh --nj $nj --cmd "$decode_cmd" --config conf/decode.conf \
$graph_dir data/$mic/dev exp/$mic/tri3a/decode_dev_${LM}
$graph_dir data/$mic/dev exp/$mic/tri3a/decode_dev_${LM}
steps/decode.sh --nj $nj --cmd "$decode_cmd" --config conf/decode.conf \
$graph_dir data/$mic/eval exp/$mic/tri3a/decode_eval_${LM}
fi
fi

if [ $stage -le 7 ]; then
# Train tri4a, which is LDA+MLLT+SAT,
steps/train_sat.sh --cmd "$train_cmd" \
5000 80000 data/$mic/train data/lang exp/$mic/tri3a_ali exp/$mic/tri4a
# Decode,
# Decode,
graph_dir=exp/$mic/tri4a/graph_${LM}
$highmem_cmd $graph_dir/mkgraph.log \
utils/mkgraph.sh data/lang_${LM} exp/$mic/tri4a $graph_dir
steps/decode_fmllr.sh --nj $nj --cmd "$decode_cmd" --config conf/decode.conf \
$graph_dir data/$mic/dev exp/$mic/tri4a/decode_dev_${LM}
$graph_dir data/$mic/dev exp/$mic/tri4a/decode_dev_${LM}
steps/decode_fmllr.sh --nj $nj --cmd "$decode_cmd" --config conf/decode.conf \
$graph_dir data/$mic/eval exp/$mic/tri4a/decode_eval_${LM}
$graph_dir data/$mic/eval exp/$mic/tri4a/decode_eval_${LM}
fi

nj_mmi=80
Expand Down Expand Up @@ -160,11 +160,11 @@ if [ $stage -le 11 ]; then
decode_dir=exp/$mic/tri4a_mmi_b0.1/decode_dev_${i}.mdl_${LM}
steps/decode.sh --nj $nj --cmd "$decode_cmd" --config conf/decode.conf \
--transform-dir exp/$mic/tri4a/decode_dev_${LM} --iter $i \
$graph_dir data/$mic/dev $decode_dir
$graph_dir data/$mic/dev $decode_dir
decode_dir=exp/$mic/tri4a_mmi_b0.1/decode_eval_${i}.mdl_${LM}
steps/decode.sh --nj $nj --cmd "$decode_cmd" --config conf/decode.conf \
--transform-dir exp/$mic/tri4a/decode_eval_${LM} --iter $i \
$graph_dir data/$mic/eval $decode_dir
$graph_dir data/$mic/eval $decode_dir
done
fi

Expand All @@ -181,7 +181,7 @@ if [ $stage -le 13 ]; then
--hidden-dim 950 \
--splice-indexes "layer0/-2:-1:0:1:2 layer1/-1:2 layer2/-3:3 layer3/-7:2 layer4/-3:3" \
--use-sat-alignments true

local/online/run_nnet2_ms_sp_disc.sh \
--mic $mic \
--gmm-dir exp/$mic/tri4a \
Expand Down
88 changes: 88 additions & 0 deletions egs/aspire/s5/local/nnet3/run_autoencoder.sh
Original file line number Diff line number Diff line change
@@ -0,0 +1,88 @@
#!/bin/bash

# this is an example to show a "tdnn" system in raw nnet configuration
# i.e. without a transition model

. cmd.sh


# At this script level we don't support not running on GPU, as it would be painfully slow.
# If you want to run without GPU you'd have to call train_tdnn.sh with --gpu false,
# --num-threads 16 and --minibatch-size 128.

stage=0
affix=
train_stage=-10
common_egs_dir=
num_data_reps=10

remove_egs=true

. cmd.sh
. ./path.sh
. ./utils/parse_options.sh


if ! cuda-compiled; then
cat <<EOF && exit 1
This script is intended to be used with GPUs but you have not compiled Kaldi with CUDA
If you want to use GPUs (and have them), go to src/, and configure and make on a machine
where "nvcc" is installed.
EOF
fi

dir=exp/nnet3/tdnn_raw
dir=$dir${affix:+_$affix}

clean_data_dir=data/train
data_dir=data/train_rvb
targets_scp=$dir/targets.scp

mkdir -p $dir

# Create copies of clean feats with prefix "rev$x_" to match utterance names of
# the noisy feats
for x in `seq 1 $num_data_reps`; do
awk -v x=$x '{print "rev"x"_"$0}' $clean_data_dir/feats.scp | sort -k1,1 > $targets_scp
done

if [ $stage -le 9 ]; then
echo "$0: creating neural net configs";

num_targets=`feat-to-dim scp:$targets_scp - 2>/dev/null` || exit 1

# create the config files for nnet initialization
python steps/nnet3/tdnn/make_configs.py \
--splice-indexes "-2,-1,0,1,2 -1,2 -3,3 -7,2 0" \
--feat-dir ${data_dir} \
--relu-dim=1024 \
--add-lda=false \
--objective-type=quadratic \
--add-final-sigmoid=false \
--include-log-softmax=false \
--use-presoftmax-prior-scale=false \
--num-targets=$num_targets \
$dir/configs || exit 1;
fi

if [ $stage -le 10 ]; then
if [[ $(hostname -f) == *.clsp.jhu.edu ]] && [ ! -d $dir/egs/storage ]; then
utils/create_split_dir.pl \
/export/b0{3,4,5,6}/$USER/kaldi-data/egs/aspire-$(date +'%m_%d_%H_%M')/s5/$dir/egs/storage $dir/egs/storage
fi

steps/nnet3/tdnn/train_raw_nnet.sh --stage $train_stage \
--cmd "$decode_cmd" \
--cmvn-opts "--norm-means=false --norm-vars=false" \
--num-epochs 2 \
--num-jobs-initial 3 \
--num-jobs-final 16 \
--initial-effective-lrate 0.0017 \
--final-effective-lrate 0.00017 \
--egs-dir "$common_egs_dir" \
--remove-egs $remove_egs \
--use-gpu true \
--dense-targets true \
${data_dir} $targets_scp $dir || exit 1
fi

43 changes: 16 additions & 27 deletions egs/aurora4/s5/cmd.sh
Original file line number Diff line number Diff line change
@@ -1,29 +1,18 @@
# "queue.pl" uses qsub. The options to it are
# options to qsub. If you have GridEngine installed,
# change this to a queue you have access to.
# Otherwise, use "run.pl", which will run jobs locally
# (make sure your --num-jobs options are no more than
# the number of cpus on your machine.
# you can change cmd.sh depending on what type of queue you are using.
# If you have no queueing system and want to run on a local machine, you
# can change all instances 'queue.pl' to run.pl (but be careful and run
# commands one by one: most recipes will exhaust the memory on your
# machine). queue.pl works with GridEngine (qsub). slurm.pl works
# with slurm. Different queues are configured differently, with different
# queue names and different ways of specifying things like memory;
# to account for these differences you can create and edit the file
# conf/queue.conf to match your queue's configuration. Search for
# conf/queue.conf in http://kaldi-asr.org/doc/queue.html for more information,
# or search for the string 'default_config' in utils/queue.pl or utils/slurm.pl.

#a) JHU cluster options
export train_cmd="queue.pl -l arch=*64"
export decode_cmd="queue.pl -l arch=*64 --mem 2G"
export mkgraph_cmd="queue.pl -l arch=*64 --mem 4G"
export big_memory_cmd="queue.pl -l arch=*64 --mem 8G"
export train_cmd="queue.pl --mem 4G"
export decode_cmd="queue.pl --mem 4G"
export mkgraph_cmd="queue.pl --mem 8G"
# the use of cuda_cmd is deprecated but it's still used in some example scripts
# here.
export cuda_cmd="queue.pl --gpu 1"


#b) BUT cluster options
#export train_cmd="queue.pl -q all.q@@blade -l ram_free=1200M,mem_free=1200M"
#export decode_cmd="queue.pl -q all.q@@blade -l ram_free=1700M,mem_free=1700M"
#export decodebig_cmd="queue.pl -q all.q@@blade -l ram_free=4G,mem_free=4G"

#export cuda_cmd="queue.pl -q long.q@@pco203 -l gpu=1"
#export cuda_cmd="queue.pl -q long.q@pcspeech-gpu"
#export mkgraph_cmd="queue.pl -q all.q@@servers -l ram_free=4G,mem_free=4G"

#c) run it locally...
#export train_cmd=run.pl
#export decode_cmd=run.pl
#export cuda_cmd=run.pl
#export mkgraph_cmd=run.pl
44 changes: 15 additions & 29 deletions egs/babel/s5/cmd.sh
Original file line number Diff line number Diff line change
@@ -1,29 +1,15 @@
# "queue.pl" uses qsub. The options to it are
# options to qsub. If you have GridEngine installed,
# change this to a queue you have access to.
# Otherwise, use "run.pl", which will run jobs locally
# (make sure your --num-jobs options are no more than
# the number of cpus on your machine.

#a) JHU cluster options
export train_cmd="queue.pl -l arch=*64"
export decode_cmd="queue.pl -l arch=*64,mem_free=2G,ram_free=2G"
export mkgraph_cmd="queue.pl -l arch=*64,ram_free=4G,mem_free=4G"

#export cuda_cmd="..."


#b) BUT cluster options
#export train_cmd="queue.pl -q all.q@@blade -l ram_free=1200M,mem_free=1200M"
#export decode_cmd="queue.pl -q all.q@@blade -l ram_free=1700M,mem_free=1700M"
#export decodebig_cmd="queue.pl -q all.q@@blade -l ram_free=4G,mem_free=4G"

#export cuda_cmd="queue.pl -q long.q@@pco203 -l gpu=1"
#export cuda_cmd="queue.pl -q long.q@pcspeech-gpu"
#export mkgraph_cmd="queue.pl -q all.q@@servers -l ram_free=4G,mem_free=4G"

#c) run it locally...
#export train_cmd=run.pl
#export decode_cmd=run.pl
#export cuda_cmd=run.pl
#export mkgraph_cmd=run.pl
# you can change cmd.sh depending on what type of queue you are using.
# If you have no queueing system and want to run on a local machine, you
# can change all instances 'queue.pl' to run.pl (but be careful and run
# commands one by one: most recipes will exhaust the memory on your
# machine). queue.pl works with GridEngine (qsub). slurm.pl works
# with slurm. Different queues are configured differently, with different
# queue names and different ways of specifying things like memory;
# to account for these differences you can create and edit the file
# conf/queue.conf to match your queue's configuration. Search for
# conf/queue.conf in http://kaldi-asr.org/doc/queue.html for more information,
# or search for the string 'default_config' in utils/queue.pl or utils/slurm.pl.

export train_cmd="queue.pl --mem 2G"
export decode_cmd="queue.pl --mem 4G"
export mkgraph_cmd="queue.pl --mem 8G"
Loading