Supervised Sequence Labelling with Recurrent Neural Networks

This document was uploaded by one of our users. The uploader already confirmed that they had the permission to publish it. If you are author/publisher or own the copyright of this documents, please report to us by using this DMCA report form.

Simply click on the Download Book button.

Yes, Book downloads on Ebookily are 100% Free.

Sometimes the book is free on Amazon As well, so go ahead and hit "Search on Amazon"

Supervised sequence labelling is a vital area of machine learning, encompassing tasks such as speech, handwriting and gesture recognition, protein secondary structure prediction and part-of-speech tagging. Recurrent neural networks are powerful sequence learning tools—robust to input noise and distortion, able to exploit long-range contextual information—that would seem ideally suited to such problems. However their role in large-scale sequence labelling systems has so far been auxiliary.

The goal of this book is a complete framework for classifying and transcribing sequential data with recurrent neural networks only. Three main innovations are introduced in order to realise this goal. Firstly, the connectionist temporal classification output layer allows the framework to be trained with unsegmented target sequences, such as phoneme-level speech transcriptions; this is in contrast to previous connectionist approaches, which were dependent on error-prone prior segmentation. Secondly, multidimensional recurrent neural networks extend the framework in a natural way to data with more than one spatio-temporal dimension, such as images and videos. Thirdly, the use of hierarchical subsampling makes it feasible to apply the framework to very large or high resolution sequences, such as raw audio or video.

Experimental validation is provided by state-of-the-art results in speech and handwriting recognition.

Author(s): Alex Graves (auth.)
Series: Studies in Computational Intelligence 385
Edition: 1
Publisher: Springer-Verlag Berlin Heidelberg
Year: 2012

Language: English
Pages: 146
Tags: Computational Intelligence; Artificial Intelligence (incl. Robotics)

Front Matter....Pages 1-11
Introduction....Pages 1-3
Supervised Sequence Labelling....Pages 5-13
Neural Networks....Pages 15-35
Long Short-Term Memory....Pages 37-45
A Comparison of Network Architectures....Pages 47-56
Hidden Markov Model Hybrids....Pages 57-60
Connectionist Temporal Classification....Pages 61-93
Multidimensional Networks....Pages 95-108
Hierarchical Subsampling Networks....Pages 109-131
Back Matter....Pages 0--1