Neural Architecture Search For LF-MMI Trained Time Delay Neural Networks
Publication in refereed journal
Officially Accepted for Publication


Times Cited
Web of Science3WOS source URL (as at 19/08/2024) Click here for the latest count
Altmetrics Information
.

Other information
AbstractState-of-the-art automatic speech recognition (ASR) system development is data and computation intensive. The optimal design of deep neural networks (DNNs) for these systems often require expert knowledge and empirical evaluation. In this paper, a range of neural architecture search (NAS) techniques are used to automatically learn two types of hyper-parameters of factored time delay neural networks (TDNN-Fs): i) the left and right splicing context offsets; and ii) the dimensionality of the bottleneck linear projection at each hidden layer. These techniques include the differentiable neural architecture search (DARTS) method integrating architecture learning with lattice-free MMI training; Gumbel-Softmax and pipelined DARTS methods reducing the confusion over candidate architectures and improving the generalization of architecture selection; and Penalized DARTS incorporating resource constraints to balance the trade-off between performance and system complexity. Parameter sharing among TDNN-F architectures allows an efficient search over up to $7^{28}$ different systems. Statistically significant word error rate (WER) reductions of up to 1.2\% absolute and relative model size reduction of 31\% were obtained over a state-of-the-art 300-hour Switchboard corpus trained baseline LF-MMI TDNN-F system featuring speed perturbation, i-Vector and learning hidden unit contribution (LHUC) based speaker adaptation as well as RNNLM rescoring. Performance contrasts on the same task against recent end-to-end systems reported in the literature suggest the best NAS auto-configured system achieves state-of-the-art WERs of 9.9\% and 11.1\% on the NIST Hub5 00 and Rt03s test sets respectively with up to 96\% model size reduction. Further analysis using Bayesian learning shows that the proposed NAS approaches can effectively minimize the structural redundancy in the TDNN-F systems and reduce their model parameter uncertainty. Consistent performance improvements were also obtained on a UASpeech dysarthric speech recognition task.
Acceptance Date13/02/2022
All Author(s) ListShoukang Hu, Xurong Xie, Mingyu Cui, Jiajun Deng, Shansong Liu, Jianwei Yu, Mengzhe Geng, Xunying Liu, Helen Meng
Journal nameIEEE/ACM Transactions on Audio, Speech and Language Processing
Year2022
ISSN2329-9290
LanguagesEnglish-United States

Last updated on 2024-20-08 at 00:28