Research Output
Deep Sparse Auto-Encoder Features Learning for Arabic Text Recognition
  One of the most recent challenging issues of pattern recognition and artificial intelligence is Arabic text recognition. This research topic is still a pervasive and unaddressed research field, because of several factors. Complications arise due to the cursive nature of the Arabic writing, character similarities, unlimited vocabulary, use of multi-size and mixed-fonts, etc. To handle these challenges, an automatic Arabic text recognition requires building a robust system by computing discriminative features and applying a rigorous classifier together to achieve an improved performance. In this work, we introduce a new deep learning based system that recognizes Arabic text contained in images. We propose a novel hybrid network, combining a Bag-of-Feature (BoF) framework for feature extraction based on a deep Sparse Auto-Encoder (SAE), and Hidden Markov Models (HMMs), for sequence recognition. Our proposed system, termed BoF-deep SAE-HMM, is tested on four datasets, namely the printed Arabic line images Printed KHATT (P-KHATT), the benchmark printed word images Arabic Printed Text Image (APTI), the benchmark handwritten Arabic word images IFN/ENIT, and the benchmark handwritten digits images Modified National Institute of Standards and Technology (MNIST).

  • Type:

    Article

  • Date:

    22 January 2021

  • Publication Status:

    Published

  • Publisher

    Institute of Electrical and Electronics Engineers (IEEE)

  • DOI:

    10.1109/access.2021.3053618

  • Cross Ref:

    10.1109/access.2021.3053618

  • Funders:

    Ministry of Higher Education and Scientific Research of Tunisia

Citation

Rahal, N., Tounsi, M., Hussain, A., & Alimi, A. M. (2021). Deep Sparse Auto-Encoder Features Learning for Arabic Text Recognition. IEEE Access, 9, 18569-18584. https://doi.org/10.1109/access.2021.3053618

Authors

Keywords

Arabic text recognition, feature learning, bag of features, sparse auto-encoder, hidden Markov models

Monthly Views:

Available Documents