VQA Challenge and Visual Dialog Workshop
Location: Room 155A, Calvin L. Rampton Salt Palace Convention Center
at CVPR 2018, June 18, Salt Lake City, Utah, USA
Home Program SubmissionAccepted Abstracts
Accepted Abstracts
Bilinear Attention Networks for Visual Question Answering
Jin-Hwa Kim, Jaehyun Jun, Byoung-Tak Zhang
Two can play this Game: Visual Dialog with Discriminative Question Generation and Answering
Unnat Jain, Svetlana Lazebnik, Alexander Schwing
Learning to Count Objects in Natural Images for Visual Question Answering
Yan Zhang, Jonathon Hare, Adam Prugel-Bennett
Joint Image Captioning and Question Answering
Jialin Wu, Zeyuan Hu, Raymond J. Mooney
Robustness Analysis of Visual QA Models by Basic Questions
Jia-Hong Huang, Cuong Duc Dao*, Modar Alfadly*, C. Huck Yang, Bernard Ghanem
Mind Your Language: Learning Visually Grounded Dialog in a Multi-Agent Setting
Akshat Agarwal*, Swaminathan Gurumurthy*, Vasu Sharma*, Katia Sycara
Dialog-based Interactive Image Retrieval
Hui Wu*, Xiaoxiao Guo*, Yu Cheng, Steven Rennie, Gerald Tesauro, Rogerio Schmidt Feris
Cross Domain Normalization for Natural Language Object Retrieval
Asi Sheffer, Michael Elhadad
DVQA: Understanding Data Visualizations via Question Answering
Kushal Kafle, Brian Price, Scott Cohen, Christopher Kanan
Block Superdiagonal Fusion
Remi Cadene, Hedi Ben-younes, Matthieu Cord, Nicolas Thome
Convolutional Image Captioning
Jyoti Aneja, Aditya Deshpande, Alexander Schwing
Visual Question Answering System as a Teaching Aid
Sourav Suresh, Varun Nagaraj Rao, Gowri Srinivasa
VizWiz Grand Challenge: Answering Visual Questions from Blind People
Danna Gurari, Qing Li, Abigale J. Stangl, Anhong Guo, Chi Lin, Kristen Grauman, Jiebo Luo, Jeffrey P. Bigham
Women also Snowboard: Overcoming Bias in Captioning Models
Kaylee Burns, Lisa Anne Hendricks, Kate Saenko, Trevor Darrell, Anna Rohrbach
Visual Question Answering as a Meta Learning Task
Damien Teney, Anton van den Hengel
Improving Visual Question Answering Using Dropout and Enhanced Question Encoder
Zhiwei Fang, Jing Liu, Qu Tang, Yanyuan Qiao, Fei Liu, Yong Li, Hanqing Lu
Focal Visual-Text Attention for Visual Question Answering
Junwei Liang, Lu Jiang, Liangliang Cao, Li-Jia Li, Alexander Hauptmann
Vision-and-Language Navigation: Interpreting visually-grounded navigation instructions in real environments
Peter Anderson, Qi Wu, Damien Teney, Jake Bruce, Mark Johnson, Niko Sunderhauf, Ian Reid, Stephen Gould, Anton van den Hengel
Bottom-Up and Top-Down Attention for Image Captioning and Visual Question Answering
Peter Anderson, Xiaodong He, Chris Buehler, Damien Teney, Mark Johnson, Stephen Gould, Lei Zhang
CS-VQA: Visual Question Answering with Compressively Sensed Images
Li-Chi Huang, Kuldeep Kulkarni, Anik Jha, Suhas Lohit, Suren Jayasuriya, Pavan Turaga
Bottom Up and Top Down Attention on Steroids: Architecture Tweaks, Learning Rate Schedules, and Ensembling
Yu Jiang*, Vivek Natarajan*, Xinlei Chen*, Marcus Rohrbach, Dhruv Batra, Devi Parikh
Audio Visual Scene-Aware Dialog (AVSD) Challenge at DSTC7
Huda Alamri, Vincent Cartillier, Raphael Gontijo Lopes, Abhishek Das, Jue Wang, Irfan Essa, Dhruv Batra, Devi Parikh, Anoop Cherian, Tim K. Marks, Chiori Hori
Answer Again: Multi-Answering Model for Visual Question Answering
Liang Peng, Yang Yang, Heng Tao Shen
Dual Recurrent Attention Units for Visual Question Answering
Ahmed Osman, Wojciech Samek
Straight to the Facts: Learning Knowledge Base Retrieval for Factual Visual Question Answering
Medhini G. Narasimhan, Alexander Schwing
Evaluating Visual-Semantic Explanations using a Collaborative Image Guessing Game
Arijit Ray, Yi Yao, Avi Ziskind, Rakesh Kumar, Giedrius Burachas