In this article, we propose a hybrid model for spotting words and regular expressions (REGEX) in handwritten
documents. The model is made of the state-of-the-art BLSTM (Bidirectional Long Short Time Memory) neural
network for recognizing and segmenting characters, coupled with a HMM to build line models able to spot the
desired sequences. Experiments on the Rimes database show very promising results.
KEYWORDS: Neural networks, Databases, Signal processing, Feature extraction, Detection and tracking algorithms, Systems modeling, Image processing, Data processing, Data modeling, Stochastic processes
The BLSTM-CTC is a novel recurrent neural network architecture that has outperformed previous state
of the art algorithms in tasks such as speech recognition or handwriting recognition. It has the ability to
process long term dependencies in temporal signals in order to label unsegmented data. This paper describes
different ways of combining features using a BLSTM-CTC architecture. Not only do we explore the low level
combination (feature space combination) but we also explore high level combination (decoding combination)
and mid-level (internal system representation combination). The results are compared on the RIMES word
database. Our results show that the low level combination works best, thanks to the powerful data modeling
of the LSTM neurons.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
INSTITUTIONAL Select your institution to access the SPIE Digital Library.
PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.