pdf文档 Connectionist Temporal Classification Labelling Unsegmented Sequence Data with Recurrent Neural Network

专业资料 > IT/计算机 > 人工智能 > 文档预览
8 页 1 下载 1776 浏览 0 评论 0 收藏 3.0分
温馨提示:如果当前文档出现乱码或未能正常浏览,请先下载原文档进行浏览。
Connectionist Temporal Classification Labelling Unsegmented Sequence Data with Recurrent Neural Network 第 1 页 Connectionist Temporal Classification Labelling Unsegmented Sequence Data with Recurrent Neural Network 第 2 页 Connectionist Temporal Classification Labelling Unsegmented Sequence Data with Recurrent Neural Network 第 3 页 Connectionist Temporal Classification Labelling Unsegmented Sequence Data with Recurrent Neural Network 第 4 页 Connectionist Temporal Classification Labelling Unsegmented Sequence Data with Recurrent Neural Network 第 5 页
下载文档到电脑,方便使用
还有 3 页可预览,继续阅读

Connectionist Temporal Classification Labelling Unsegmented Sequence Data with Recurrent Neural Network内容摘要:

Connectionist Temporal Classification: Labelling Unsegmented Sequence Data with Recurrent Neural Networks Alex Graves1 alex@idsia.ch Santiago Fern´ andez1 santiago@idsia.ch Faustino Gomez1 tino@idsia.ch J¨ urgen Schmidhuber1,2 juergen@idsia.ch 1 Istituto Dalle Molle di Studi sull’Intelligenza Artificiale (IDSIA), Galleria 2, 6928 Manno-Lugano, Switzerland 2 Technische Universit¨ at M¨ unchen (TUM), Boltzmannstr. 3, 85748 Garching, Munich, Germany Abstract Many real-world sequence learning tasks require the prediction of sequences of labels from noisy, unsegmented input data. In speech recognition, for example, an acoustic signal is transcribed into words or sub-word units. Recurrent neural networks (RNNs) are powerful sequence learners that would seem well suited to such tasks. However, because they require pre-segmented training data, and post-processing to transform their outputs into label sequences, their applicability has so far been limited. This paper presents a novel method for training RNNs to label unsegmented sequences directly, thereby solving both problems. An experiment on the TIMIT speech corpus demonstrates its advantages over both a baseline HMM and a hybrid HMM-RNN. 1. Introduction Labelling unsegmented sequence data is a ubiquitous problem in real-world sequence learning. It is particularly common in perceptual tasks (e.g. handwriting recognition, speech recognition, gesture recognition) where noisy, real-valued input streams are annotated with strings of discrete labels, such as letters or words. Currently, graphical models such as hidden Markov Models (HMMs; Rabiner, 1989), conditional random fields (CRFs; Lafferty et al., 2001) and their variants, are the predominant framework for sequence laAppearing in Proceedings of the 23 rd International Conference on Machine Learning, Pittsburgh, PA, 2006. Copyright 2006 by the author(s)/owner(s). belling. While these approaches have proved successful for many problems, they have several drawbacks: (1) they usually require a significant amount of task specific knowledge, e.g. to design the state models for HMMs, or choose the input features for CRFs; (2) they require explicit (and often questionable) dependency assumptions to make inference tractable, e.g. the assumption that observations are independent for HMMs; (3) for standard HMMs, training is generative, even though sequence labelling is discriminative. Recurrent neural networks (RNNs), on the other hand, require no prior knowledge of the data, beyond the choice of input and output representation. They can be trained discriminatively, and their internal state provides a powerful, general mechanism for modelling time series. In addition, they tend to be robust to temporal and spatial noise. So far, however, it has not been possible to apply RNNs directly to sequence labelling. The problem is that the standard neural network objective functions are defined separately for each point in the training sequence; in other words, RNNs can only be trained to make a series of independent label classifications. This means that the training data must be pre-segmented, and that the network outputs must be post-processed to give the final label sequence. At present, the most effective use of RNNs for sequence labelling is to combine them with HMMs in the so-called hybrid approach (Bourlard & Morgan, 1994; Bengio., 1999). Hybrid systems use HMMs to model the long-range sequential structure of the data, and neural nets to provide localised classifications. The HMM component is able to automatically segment the sequence during training, and to transform the network classifications into label sequences. However, as well as inheriting the aforementioned drawbacks of Connectionist Temporal Classification HMMs, hybrid systems do not exploit the full potential of RNNs for sequence modelling. This paper presents a novel method for labelling sequence data with RNNs that removes the need for presegmented training data and post-processed outputs, and models all aspects of the sequence within a single network architecture. The basic idea is to interpret the network outputs as a probability distribution over all possible label sequences, conditioned on a given input sequence. Given this distribution, an objective function can be derived that directly maximises the probabilities of the correct labellings. Since the objective function is differentiable, the network can then be trained with standard backpropagation through t

本文档由 sddwt2021-06-13 01:04:20上传分享
给文档打分
您好可以输入 255 个字符
本站的域名是什么?( 答案:sciwk.com )
评论列表
  • 暂时还没有评论,期待您的金玉良言