File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Conference Paper: Structured Attentions for Visual Question Answering

TitleStructured Attentions for Visual Question Answering
Authors
Issue Date2017
Citation
Proceedings of the IEEE International Conference on Computer Vision, 2017, v. 2017-October, p. 1300-1309 How to Cite?
AbstractVisual attention, which assigns weights to image regions according to their relevance to a question, is considered as an indispensable part by most Visual Question Answering models. Although the questions may involve complex rela- tions among multiple regions, few attention models can ef- fectively encode such cross-region relations. In this paper, we demonstrate the importance of encoding such relations by showing the limited effective receptive field of ResNet on two datasets, and propose to model the visual attention as a multivariate distribution over a grid-structured Con- ditional Random Field on image regions. We demonstrate how to convert the iterative inference algorithms, Mean Field and Loopy Belief Propagation, as recurrent layers of an end-to-end neural network. We empirically evalu- ated our model on 3 datasets, in which it surpasses the best baseline model of the newly released CLEVR dataset [13] by 9.5%, and the best published model on the VQA dataset [3] by 1.25%. Source code is available at https://github.com/zhuchen03/vqa-sva.
Persistent Identifierhttp://hdl.handle.net/10722/327173
ISSN
2023 SCImago Journal Rankings: 12.263
ISI Accession Number ID

 

DC FieldValueLanguage
dc.contributor.authorZhu, Chen-
dc.contributor.authorZhao, Yanpeng-
dc.contributor.authorHuang, Shuaiyi-
dc.contributor.authorTu, Kewei-
dc.contributor.authorMa, Yi-
dc.date.accessioned2023-03-31T05:29:29Z-
dc.date.available2023-03-31T05:29:29Z-
dc.date.issued2017-
dc.identifier.citationProceedings of the IEEE International Conference on Computer Vision, 2017, v. 2017-October, p. 1300-1309-
dc.identifier.issn1550-5499-
dc.identifier.urihttp://hdl.handle.net/10722/327173-
dc.description.abstractVisual attention, which assigns weights to image regions according to their relevance to a question, is considered as an indispensable part by most Visual Question Answering models. Although the questions may involve complex rela- tions among multiple regions, few attention models can ef- fectively encode such cross-region relations. In this paper, we demonstrate the importance of encoding such relations by showing the limited effective receptive field of ResNet on two datasets, and propose to model the visual attention as a multivariate distribution over a grid-structured Con- ditional Random Field on image regions. We demonstrate how to convert the iterative inference algorithms, Mean Field and Loopy Belief Propagation, as recurrent layers of an end-to-end neural network. We empirically evalu- ated our model on 3 datasets, in which it surpasses the best baseline model of the newly released CLEVR dataset [13] by 9.5%, and the best published model on the VQA dataset [3] by 1.25%. Source code is available at https://github.com/zhuchen03/vqa-sva.-
dc.languageeng-
dc.relation.ispartofProceedings of the IEEE International Conference on Computer Vision-
dc.titleStructured Attentions for Visual Question Answering-
dc.typeConference_Paper-
dc.description.naturelink_to_subscribed_fulltext-
dc.identifier.doi10.1109/ICCV.2017.145-
dc.identifier.scopuseid_2-s2.0-85041915476-
dc.identifier.volume2017-October-
dc.identifier.spage1300-
dc.identifier.epage1309-
dc.identifier.isiWOS:000425498401038-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats