VQA Challenge and Visual Dialog Workshop
Location: Room 155A, Calvin L. Rampton Salt Palace Convention Center
at CVPR 2018, June 18, Salt Lake City, Utah, USA

Home Program SubmissionAccepted Abstracts


Accepted Abstracts

Bilinear Attention Networks for Visual Question Answering
Jin-Hwa Kim, Jaehyun Jun, Byoung-Tak Zhang

Two can play this Game: Visual Dialog with Discriminative Question Generation and Answering
Unnat Jain, Svetlana Lazebnik, Alexander Schwing

Learning to Count Objects in Natural Images for Visual Question Answering
Yan Zhang, Jonathon Hare, Adam Prugel-Bennett

Joint Image Captioning and Question Answering
Jialin Wu, Zeyuan Hu, Raymond J. Mooney

Robustness Analysis of Visual QA Models by Basic Questions
Jia-Hong Huang, Cuong Duc Dao*, Modar Alfadly*, C. Huck Yang, Bernard Ghanem

Mind Your Language: Learning Visually Grounded Dialog in a Multi-Agent Setting
Akshat Agarwal*, Swaminathan Gurumurthy*, Vasu Sharma*, Katia Sycara

Dialog-based Interactive Image Retrieval
Hui Wu*, Xiaoxiao Guo*, Yu Cheng, Steven Rennie, Gerald Tesauro, Rogerio Schmidt Feris

Cross Domain Normalization for Natural Language Object Retrieval
Asi Sheffer, Michael Elhadad

DVQA: Understanding Data Visualizations via Question Answering
Kushal Kafle, Brian Price, Scott Cohen, Christopher Kanan

Block Superdiagonal Fusion
Remi Cadene, Hedi Ben-younes, Matthieu Cord, Nicolas Thome

Convolutional Image Captioning
Jyoti Aneja, Aditya Deshpande, Alexander Schwing

Visual Question Answering System as a Teaching Aid
Sourav Suresh, Varun Nagaraj Rao, Gowri Srinivasa

VizWiz Grand Challenge: Answering Visual Questions from Blind People
Danna Gurari, Qing Li, Abigale J. Stangl, Anhong Guo, Chi Lin, Kristen Grauman, Jiebo Luo, Jeffrey P. Bigham

Women also Snowboard: Overcoming Bias in Captioning Models
Kaylee Burns, Lisa Anne Hendricks, Kate Saenko, Trevor Darrell, Anna Rohrbach

Visual Question Answering as a Meta Learning Task
Damien Teney, Anton van den Hengel

Improving Visual Question Answering Using Dropout and Enhanced Question Encoder
Zhiwei Fang, Jing Liu, Qu Tang, Yanyuan Qiao, Fei Liu, Yong Li, Hanqing Lu

Focal Visual-Text Attention for Visual Question Answering
Junwei Liang, Lu Jiang, Liangliang Cao, Li-Jia Li, Alexander Hauptmann

Vision-and-Language Navigation: Interpreting visually-grounded navigation instructions in real environments
Peter Anderson, Qi Wu, Damien Teney, Jake Bruce, Mark Johnson, Niko Sunderhauf, Ian Reid, Stephen Gould, Anton van den Hengel

Bottom-Up and Top-Down Attention for Image Captioning and Visual Question Answering
Peter Anderson, Xiaodong He, Chris Buehler, Damien Teney, Mark Johnson, Stephen Gould, Lei Zhang

CS-VQA: Visual Question Answering with Compressively Sensed Images
Li-Chi Huang, Kuldeep Kulkarni, Anik Jha, Suhas Lohit, Suren Jayasuriya, Pavan Turaga

Bottom Up and Top Down Attention on Steroids: Architecture Tweaks, Learning Rate Schedules, and Ensembling
Yu Jiang*, Vivek Natarajan*, Xinlei Chen*, Marcus Rohrbach, Dhruv Batra, Devi Parikh

Audio Visual Scene-Aware Dialog (AVSD) Challenge at DSTC7
Huda Alamri, Vincent Cartillier, Raphael Gontijo Lopes, Abhishek Das, Jue Wang, Irfan Essa, Dhruv Batra, Devi Parikh, Anoop Cherian, Tim K. Marks, Chiori Hori

Answer Again: Multi-Answering Model for Visual Question Answering
Liang Peng, Yang Yang, Heng Tao Shen

Dual Recurrent Attention Units for Visual Question Answering
Ahmed Osman, Wojciech Samek

Straight to the Facts: Learning Knowledge Base Retrieval for Factual Visual Question Answering
Medhini G. Narasimhan, Alexander Schwing

Evaluating Visual-Semantic Explanations using a Collaborative Image Guessing Game
Arijit Ray, Yi Yao, Avi Ziskind, Rakesh Kumar, Giedrius Burachas