File Download

There are no files associated with this item.

Supplementary

Conference Paper: Char-Net: A Character-Aware Neural Network for Distorted Scene Text Recognition

TitleChar-Net: A Character-Aware Neural Network for Distorted Scene Text Recognition
Authors
KeywordsText Recognition
Attention Mechanism
RNN
Issue Date2018
PublisherAssociation for the Advancement of Artificial Intelligence (AAAI) Press.
Citation
Proceedings of the Thirty-Second Association for the Advancement of Artificial Intelligence (AAAI) Conference on Artificial Intelligence (AAAI-18), New Orleans, Louisiana, USA, 2-7 February 2018, p. 7154-7161 How to Cite?
AbstractIn this paper, we present a Character-Aware Neural Network (Char-Net) for recognizing distorted scene text. Our Char-Net is composed of a word-level encoder, a character-level encoder, and a LSTM-based decoder. Unlike previous work which employed a global spatial transformer network to rectify the entire distorted text image, we take an approach of detecting and rectifying individual characters. To this end, we introduce a novel hierarchical attention mechanism (HAM) which consists of a recurrent RoIWarp layer and a character-level attention layer. The recurrent RoIWarp layer sequentially extracts a feature region corresponding to a character from the feature map produced by the word-level encoder, and feeds it to the character-level encoder which removes the distortion of the character through a simple spatial transformer and further encodes the character region. The character-level attention layer then attends to the most relevant features of the feature map produced by the character-level encoder and composes a context vector, which is finally fed to the LSTM-based decoder for decoding. This approach of adopting a simple local transformation to model the distortion of individual characters not only results in an improved efficiency, but can also handle different types of distortion that are hard, if not impossible, to be modelled by a single global transformation. Experiments have been conducted on six public benchmark datasets. Our results show that Char-Net can achieve state-of-the-art performance on all the benchmarks, especially on the IC-IST which contains scene text with large distortion. Code will be made available.
DescriptionSession: AAAI18 - Vision
Persistent Identifierhttp://hdl.handle.net/10722/250580

 

DC FieldValueLanguage
dc.contributor.authorLiu, W-
dc.contributor.authorChen, C-
dc.contributor.authorWong, KKY-
dc.date.accessioned2018-01-18T04:29:17Z-
dc.date.available2018-01-18T04:29:17Z-
dc.date.issued2018-
dc.identifier.citationProceedings of the Thirty-Second Association for the Advancement of Artificial Intelligence (AAAI) Conference on Artificial Intelligence (AAAI-18), New Orleans, Louisiana, USA, 2-7 February 2018, p. 7154-7161-
dc.identifier.urihttp://hdl.handle.net/10722/250580-
dc.descriptionSession: AAAI18 - Vision-
dc.description.abstractIn this paper, we present a Character-Aware Neural Network (Char-Net) for recognizing distorted scene text. Our Char-Net is composed of a word-level encoder, a character-level encoder, and a LSTM-based decoder. Unlike previous work which employed a global spatial transformer network to rectify the entire distorted text image, we take an approach of detecting and rectifying individual characters. To this end, we introduce a novel hierarchical attention mechanism (HAM) which consists of a recurrent RoIWarp layer and a character-level attention layer. The recurrent RoIWarp layer sequentially extracts a feature region corresponding to a character from the feature map produced by the word-level encoder, and feeds it to the character-level encoder which removes the distortion of the character through a simple spatial transformer and further encodes the character region. The character-level attention layer then attends to the most relevant features of the feature map produced by the character-level encoder and composes a context vector, which is finally fed to the LSTM-based decoder for decoding. This approach of adopting a simple local transformation to model the distortion of individual characters not only results in an improved efficiency, but can also handle different types of distortion that are hard, if not impossible, to be modelled by a single global transformation. Experiments have been conducted on six public benchmark datasets. Our results show that Char-Net can achieve state-of-the-art performance on all the benchmarks, especially on the IC-IST which contains scene text with large distortion. Code will be made available.-
dc.languageeng-
dc.publisherAssociation for the Advancement of Artificial Intelligence (AAAI) Press. -
dc.relation.ispartofAAAI Conference on Artificial Intelligence, AAAI-18-
dc.subjectText Recognition-
dc.subjectAttention Mechanism-
dc.subjectRNN-
dc.titleChar-Net: A Character-Aware Neural Network for Distorted Scene Text Recognition-
dc.typeConference_Paper-
dc.identifier.emailWong, KKY: kykwong@cs.hku.hk-
dc.identifier.authorityWong, KKY=rp01393-
dc.identifier.hkuros284067-
dc.identifier.spage7154-
dc.identifier.epage7161-
dc.publisher.placeUnited States-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats