To undertake machine lip-reading, we try to recognise speech from a visual signal. Current work often uses viseme classification supported by language models with varying degrees of success. A few recent works suggest phoneme classification, in the right circumstances, can outperform viseme classification. In this work we present a novel two-pass method of training phoneme classifiers which uses previously trained visemes in the first pass. With our new training algorithm, we show classification performance which significantly improves on previous lip-reading results.
Decoding visemes: Improving machine lip-reading
Published 2016 in IEEE International Conference on Acoustics, Speech, and Signal Processing
ABSTRACT
PUBLICATION RECORD
- Publication year
2016
- Venue
IEEE International Conference on Acoustics, Speech, and Signal Processing
- Publication date
2016-03-20
- Fields of study
Computer Science
- Identifiers
- External record
- Source metadata
Semantic Scholar
CITATION MAP
EXTRACTION MAP
CLAIMS
- No claims are published for this paper.
CONCEPTS
- No concepts are published for this paper.
REFERENCES
CITED BY
Showing 1-45 of 45 citing papers · Page 1 of 1