Back to Search Start Over

Transcription of Spanish Historical Handwritten Documents with Deep Neural Networks

Authors :
Emilio Granell
Edgard Chammas
Laurence Likforman-Sulem
Carlos-D. Martínez-Hinarejos
Chafic Mokbel
Bogdan-Ionuţ Cîrstea
Source :
Journal of Imaging, Vol 4, Iss 1, p 15 (2018)
Publication Year :
2018
Publisher :
MDPI AG, 2018.

Abstract

The digitization of historical handwritten document images is important for the preservation of cultural heritage. Moreover, the transcription of text images obtained from digitization is necessary to provide efficient information access to the content of these documents. Handwritten Text Recognition (HTR) has become an important research topic in the areas of image and computational language processing that allows us to obtain transcriptions from text images. State-of-the-art HTR systems are, however, far from perfect. One difficulty is that they have to cope with image noise and handwriting variability. Another difficulty is the presence of a large amount of Out-Of-Vocabulary (OOV) words in ancient historical texts. A solution to this problem is to use external lexical resources, but such resources might be scarce or unavailable given the nature and the age of such documents. This work proposes a solution to avoid this limitation. It consists of associating a powerful optical recognition system that will cope with image noise and variability, with a language model based on sub-lexical units that will model OOV words. Such a language modeling approach reduces the size of the lexicon while increasing the lexicon coverage. Experiments are first conducted on the publicly available Rodrigo dataset, which contains the digitization of an ancient Spanish manuscript, with a recognizer based on Hidden Markov Models (HMMs). They show that sub-lexical units outperform word units in terms of Word Error Rate (WER), Character Error Rate (CER) and OOV word accuracy rate. This approach is then applied to deep net classifiers, namely Bi-directional Long-Short Term Memory (BLSTMs) and Convolutional Recurrent Neural Nets (CRNNs). Results show that CRNNs outperform HMMs and BLSTMs, reaching the lowest WER and CER for this image dataset and significantly improving OOV recognition.

Details

Language :
English
ISSN :
2313433X
Volume :
4
Issue :
1
Database :
Directory of Open Access Journals
Journal :
Journal of Imaging
Publication Type :
Academic Journal
Accession number :
edsdoj.b3e58cffbfd8442eafa6dff6483d8747
Document Type :
article
Full Text :
https://doi.org/10.3390/jimaging4010015