Multimodal question answering over structured data with ambiguous entities

Huadong Li, Yafang Wang, Gerard De Melo, Changhe Tu, Baoquan Chen

Research output: Chapter in Book/Report/Conference proceedingConference contribution

9 Scopus citations

Abstract

In recent years, we have witnessed profound changes in the way people satisfy their information needs. For instance, with the ubiquitous 24/7 availability of mobile devices, the number of search engine queries on mobile devices has reportedly overtaken that of queries on regular personal computers. In this paper, we consider the task of multimodal question answering over structured data, in which a user supplies not just a natural language query but also an image. Our system addresses this by optimizing a non-convex objective function capturing multimodal constraints. Our experiments show that this enables it to answer even very challenging ambiguous entity queries with high accuracy.

Original languageEnglish (US)
Title of host publication26th International World Wide Web Conference 2017, WWW 2017 Companion
PublisherInternational World Wide Web Conferences Steering Committee
Pages79-88
Number of pages10
ISBN (Electronic)9781450349147
DOIs
StatePublished - 2017
Event26th International World Wide Web Conference, WWW 2017 Companion - Perth, Australia
Duration: Apr 3 2017Apr 7 2017

Publication series

Name26th International World Wide Web Conference 2017, WWW 2017 Companion

Other

Other26th International World Wide Web Conference, WWW 2017 Companion
Country/TerritoryAustralia
CityPerth
Period4/3/174/7/17

All Science Journal Classification (ASJC) codes

  • Software
  • Computer Networks and Communications

Keywords

  • Multimedia knowledge bases
  • Multimodal
  • Question answering

Fingerprint

Dive into the research topics of 'Multimodal question answering over structured data with ambiguous entities'. Together they form a unique fingerprint.

Cite this