Transformer-Based Bidirectional Attention Network for Segmentation-Free Word-Level Text Recognition with Overlapping Characters

A. Pandey,Arun Kumar Shukla

Published 2025 in 2025 International Conference on Electronics and Computing, Communication Networking Automation Technologies (ICEC2NT)

ABSTRACT

Word-level text recognition continues to present a challenging task in document and scene analysis, particularly in the case of overcrowded characters, where conventional methods based on segmentation fail. This paper proposes a non-segmentation approach for text recognition based on transformer architecture with bidirectional attention. The model exploits an encoding-decoding transformer on top of a convolutional visual feature extraction module that permits a global contextual understanding across character sequences. This method does not rely on temporal convolutional networks or recurrent units but otherwise incorporates the bidirectional attention mechanisms added to improve alignment in sequences, character prediction, and so forth, without clear segmentation. We evaluate our model on benchmark datasets of synthetic and real-world overlapping words, which exhibit dramatic improvements over previous CNN-CTC and TCN methods in terms of accuracy and robustness. The architecture sets the limits of state of the art on end-to-end text recognition through complex visual distortions and brings to light the promise of transformer-based solutions in the area of fine-Grain character interaction.

PUBLICATION RECORD

  • Publication year

    2025

  • Venue

    2025 International Conference on Electronics and Computing, Communication Networking Automation Technologies (ICEC2NT)

  • Publication date

    2025-09-03

  • Fields of study

    Not labeled

  • Identifiers
  • External record

    Open on Semantic Scholar

  • Source metadata

    Semantic Scholar

CITATION MAP

EXTRACTION MAP

CLAIMS

  • No claims are published for this paper.

CONCEPTS

  • No concepts are published for this paper.

REFERENCES

Showing 1-24 of 24 references · Page 1 of 1

CITED BY

  • No citing papers are available for this paper.

Showing 0-0 of 0 citing papers · Page 1 of 1