Here, we introduce a fully local index named "sensitivity" for each neuron to
control chaoticity or gradient globally in a neural network (NN). We also
propose a learning method to adjust it named "se
This paper proposes a modification to RNN-Transducer (RNN-T) models for
automatic speech recognition (ASR). In standard RNN-T, the emission of a blank
symbol consumes exactly one input frame; in our p
Learning hierarchical structures in sequential data -- from simple
algorithmic patterns to natural language -- in a reliable, generalizable way
remains a challenging problem for neural language models
The two most popular loss functions for streaming end-to-end automatic speech
recognition (ASR) are the RNN-Transducer (RNN-T) and the connectionist temporal
classification (CTC) objectives. Both perf
This paper aims to discuss and analyze the potentialities of Recurrent Neural
Networks (RNN) in control design applications. The main families of RNN are
considered, namely Neural Nonlinear AutoRegres
Previous works on the Recurrent Neural Network-Transducer (RNN-T) models have
shown that, under some conditions, it is possible to simplify its prediction
network with little or no loss in recognition
Spatiotemporal predictive learning is to predict future frames changes
through historical prior knowledge. Previous work improves prediction
performance by making the network wider and deeper, but thi
Recently, there has been a strong push to transition from hybrid models to end-to-end (E2E) models for automatic speech recognition. Currently, there are three promising E2E methods: recurrent neural
The recurrent neural network transducer (RNN-T) objective plays a major role
in building today's best automatic speech recognition (ASR) systems for
production. Similarly to the connectionist temporal
In this paper, we explore different ways to extend a recurrent neural network
(RNN) to a \textit{deep} RNN. We start by arguing that the concept of depth in
an RNN is not as clear as it is in feedforw
As high-throughput biological sequencing becomes faster and cheaper, the need
to extract useful information from sequencing becomes ever more paramount,
often limited by low-throughput experimental ch
Long short-term memory (LSTM) and recurrent neural network (RNN) has achieved
great successes on time-series prediction. In this paper, a methodology of
using LSTM-based deep-RNN for two-phase flow re
Lower leg prostheses could improve the life quality of amputees by increasing
comfort and reducing energy to locomote, but currently control methods are
limited in modulating behaviors based upon the
Because of its streaming nature, recurrent neural network transducer (RNN-T) is a very promising end-to-end (E2E) model that may replace the popular hybrid model for automatic speech recognition. In t
Sequence-to-sequence models have delivered impressive results in wordformation tasks such as morphological inflection, often learning to model subtle morphophonological details with limited training data.
Despite theperformance, the opacity of neural models makes it difficult to determine whether complex generalizations are learned, or whether a kind of separate rote memororization of each morphophonological process takes place.
Because of their effectiveness in broad practical applications, LSTM networks
have received a wealth of coverage in scientific journals, technical blogs, and
implementation guides. However, in most ar
Irregular time series data are prevalent in the real world and are
challenging to model with a simple recurrent neural network (RNN). Hence, a
model that combines the use of ordinary differential equa
Sparse neural networks have been widely applied to reduce the necessary
resource requirements to train and deploy over-parameterized deep neural
networks. For inference acceleration, methods that indu
Performance RNN is a machine-learning system designed primarily for the
generation of solo piano performances using an event-based (rather than audio)
representation. More specifically, Performance RN