Overcoming language priors in VQA via decomposed linguistic representations
| dc.contributor.author | Jing, C. | |
| dc.contributor.author | Wu, Y. | |
| dc.contributor.author | Zhang, X. | |
| dc.contributor.author | Jia, Y. | |
| dc.contributor.author | Wu, Q. | |
| dc.contributor.conference | AAAI Conference on Artificial Intelligence (AAAI) (7 Feb 2020 - 12 Feb 2020 : New York, USA) | |
| dc.date.issued | 2020 | |
| dc.description.abstract | Most existing Visual Question Answering (VQA) models overly rely on language priors between questions and answers. In this paper, we present a novel method of language attention-based VQA that learns decomposed linguistic representations of questions and utilizes the representations to infer answers for overcoming language priors. We introduce a modular language attention mechanism to parse a question into three phrase representations: type representation, object representation, and concept representation. We use the type representation to identify the question type and the possible answer set (yes/no or specific concepts such as colors or numbers), and the object representation to focus on the relevant region of an image. The concept representation is verified with the attended region to infer the final answer. The proposed method decouples the language-based concept discovery and vision-based concept verification in the process of answer inference to prevent language priors from dominating the answering process. Experiments on the VQA-CP dataset demonstrate the effectiveness of our method. | |
| dc.description.statementofresponsibility | Chenchen Jing, Yuwei Wu, Xiaoxun Zhang, Yunde Jia, Qi Wu | |
| dc.identifier.citation | Proceedings of the ... AAAI Conference on Artificial Intelligence. AAAI Conference on Artificial Intelligence, 2020, vol.34, iss.7, pp.11181-11188 | |
| dc.identifier.doi | 10.1609/aaai.v34i07.6776 | |
| dc.identifier.isbn | 9781577358350 | |
| dc.identifier.issn | 2159-5399 | |
| dc.identifier.issn | 2374-3468 | |
| dc.identifier.orcid | Wu, Q. [0000-0003-3631-256X] | |
| dc.identifier.uri | https://hdl.handle.net/2440/133677 | |
| dc.language.iso | en | |
| dc.publisher | AAAI | |
| dc.publisher.place | online | |
| dc.relation.grant | http://purl.org/au-research/grants/arc/DE190100539 | |
| dc.relation.ispartofseries | AAAI Conference on Artificial Intelligence | |
| dc.rights | © 2020, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved. | |
| dc.source.uri | https://aaai.org/Library/AAAI/aaai20contents-issue07.php | |
| dc.title | Overcoming language priors in VQA via decomposed linguistic representations | |
| dc.type | Conference paper | |
| pubs.publication-status | Published |