We study model pruning methods applied to Transformer-based neural network language models for automatic speech recognition. We explore three aspects of the pruning frame work, namely criterion, method and scheduler, analyzing their contribution in terms of accuracy and inference speed. To the best of our knowledge, such in-depth analyses on large-scale recognition systems has not been reported in the literature. In addition, we propose a variant of low-rank approximation suitable for incrementally compressing models, and delivering multiple models with varied target sizes. Among other results, we show that a) data-driven pruning outperforms magnitude-driven in several scenarios; b) incremental pruning achieves higher accuracy compared to one-shot pruning, especially when targeting smaller sizes; and c) low-rank approximation presents the best trade-off between size reduction and inference speed-up for moderate compression.
Neural Language Model Pruning for Automatic Speech Recognition
Leonardo Emili,Thiago Fraga-Silva,Ernest Pusateri,M. Nußbaum-Thom,Youssef Oualil
Published 2023 in arXiv.org
ABSTRACT
PUBLICATION RECORD
- Publication year
2023
- Venue
arXiv.org
- Publication date
2023-10-05
- Fields of study
Computer Science
- Identifiers
- External record
- Source metadata
Semantic Scholar
CITATION MAP
EXTRACTION MAP
CLAIMS
- No claims are published for this paper.
CONCEPTS
- No concepts are published for this paper.
REFERENCES
Showing 1-50 of 50 references · Page 1 of 1
CITED BY
Showing 1-3 of 3 citing papers · Page 1 of 1