1 code implementation • EMNLP 2017 • Aroma Mahendru, Viraj Prabhu, Akrit Mohapatra, Dhruv Batra, Stefan Lee
In this paper, we make a simple observation that questions about images often contain premises - objects and relationships implied by the question - and that reasoning about premises can help Visual Question Answering (VQA) models respond more intelligently to irrelevant or previously unseen questions.
no code implementations • 31 Aug 2016 • Yash Goyal, Akrit Mohapatra, Devi Parikh, Dhruv Batra
In this paper, we address the problem of interpreting Visual Question Answering (VQA) models.
no code implementations • 12 Jun 2015 • Harsh Agrawal, Clint Solomon Mathialagan, Yash Goyal, Neelima Chavali, Prakriti Banik, Akrit Mohapatra, Ahmed Osman, Dhruv Batra
We are witnessing a proliferation of massive visual data.