Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Speech recognition with deep recurrent neural networks
8.784
Zitationen
3
Autoren
2013
Jahr
Abstract
Recurrent neural networks (RNNs) are a powerful model for sequential data. End-to-end training methods such as Connectionist Temporal Classification make it possible to train RNNs for sequence labelling problems where the input-output alignment is unknown. The combination of these methods with the Long Short-term Memory RNN architecture has proved particularly fruitful, delivering state-of-the-art results in cursive handwriting recognition. However RNN performance in speech recognition has so far been disappointing, with better results returned by deep feedforward networks. This paper investigates deep recurrent neural networks, which combine the multiple levels of representation that have proved so effective in deep networks with the flexible use of long range context that empowers RNNs. When trained end-to-end with suitable regularisation, we find that deep Long Short-term Memory RNNs achieve a test set error of 17.7% on the TIMIT phoneme recognition benchmark, which to our knowledge is the best recorded score.
Ähnliche Arbeiten
AI-Assisted Pipeline for Dynamic Generation of Trustworthy Health Supplement Content at Scale
2018 · 45.495 Zit.
Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation
2014 · 24.073 Zit.
A tutorial on hidden Markov models and selected applications in speech recognition
1989 · 22.674 Zit.
Efficient Estimation of Word Representations in Vector Space
2013 · 18.100 Zit.
Conditional Random Fields: Probabilistic Models for Segmenting and Labeling Sequence Data
2001 · 12.995 Zit.