A discriminative model for perceptually-grounded incremental reference resolution

Casey Kennington, Livia Dia, David Schlangen

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

8 Scopus citations

Abstract

A large part of human communication involves referring to entities in the world, and often these entities are objects that are visually present for the interlocutors. A computer system that aims to resolve such references needs to tackle a complex task: objects and their visual features must be determined, the referring expressions must be recognised, extra-linguistic information such as eye gaze or pointing gestures must be incorporated - and the intended connection between words and world must be reconstructed. In this paper, we introduce a discriminative model of reference resolution that processes incrementally (i.e., word for word), is perceptually-grounded, and improves when interpolated with information from gaze and pointing gestures. We evaluated our model and found that it performed robustly in a realistic reference resolution task, when compared to a generative model.

Original languageEnglish
Title of host publicationIWCS 2015 - Proceedings of the 11th International Conference on Computational Semantics
Pages195-205
Number of pages11
ISBN (Electronic)9781941643334
StatePublished - 2015
Event11th International Conference on Computational Semantics, IWCS 2015 - London, United Kingdom
Duration: 15 Apr 201517 Apr 2015

Publication series

NameIWCS 2015 - Proceedings of the 11th International Conference on Computational Semantics

Conference

Conference11th International Conference on Computational Semantics, IWCS 2015
Country/TerritoryUnited Kingdom
CityLondon
Period15/04/1517/04/15

Fingerprint

Dive into the research topics of 'A discriminative model for perceptually-grounded incremental reference resolution'. Together they form a unique fingerprint.

Cite this