• Keine Ergebnisse gefunden

MeetUp! A Task For Modelling Visual Dialogue

N/A
N/A
Protected

Academic year: 2022

Aktie "MeetUp! A Task For Modelling Visual Dialogue"

Copied!
2
0
0

Wird geladen.... (Jetzt Volltext ansehen)

Volltext

(1)

MeetUp! A Task For Modelling Visual Dialogue

David Schlangen, Nikolai Ilinykh, Sina Zarrieß

Dialogue Systems Group // CITEC // Linguistics & Literary Studies Bielefeld University, Germany

first.last@uni-bielefeld.de

1 Introduction

After achieving impressive success representing image content textually (as done by captioning models (Fang et al., 2015; Devlin et al., 2015; Chen and Lawrence Zitnick, 2015; Vinyals et al., 2015; Bernardi et al., 2016); and referring expression resolution and generation (Kazemzadeh et al., 2014; Mao et al., 2015;

Yu et al., 2016; Schlangen et al., 2016)), the Vision and Language community has recently established

“Visual Dialogue” as the more challenging follow up task (Das et al., 2017; De Vries et al., 2017). In that task, a Questioner, prompted by some textual information (a caption) can ask an Answerer questions about an image that only the latter sees. We argue here that this setup leads to an impoverished form of dialogue and hence to data that is not substantially more informative than captioning data, if the goal is to model visual dialogue. We describe our ongoing work on the MeetUp setting, where two players navigate separately through a visually represented environment, with the goal of being at the same location. This goal gives them a reason to describe visual content, leading to motivated descriptions, and the dynamic setting induces an interesting split between private and shared information.

2 Visual Dialogue

(a) What the ‘questioner’ sees. (b) What the ‘answerer’ sees. (c) Example dialog from our VisDial dataset.

Figure 3: Collecting visually-grounded dialog data on Amazon Mechanical Turk via a live chat interface where one person is assigned the role of ‘questioner’ and the second person is the ‘answerer’. We show the first two questions being collected via the interface as Turkers interact with each other in Fig.3aand Fig.3b. Remaining questions are shown in Fig.3c.

Context (COCO) [25] dataset, which contains multiple ob- jects in everyday scenes. The visual complexity of these images allows for engaging and diverse conversations to be held about them.

Live Chat Interface. Good data for this task should in- clude dialogs that have (1) temporal continuity, (2) ground- ing in the image, and (3) mimic natural ‘conversational’

exchanges. To elicit such responses, we paired 2 work- ers on AMT to chat with each other in real-time (Fig.3).

Each worker was assigned a specific role. One worker (the

‘questioner’) sees only a single line of text describing an image (caption from COCO); the image remains hidden to the questioner. Their task is to ask questions about this hid- den image so as to ‘imagine the scene better’. The sec- ond worker (the ‘answerer’) sees the image and the cap- tion. Their task is to answer the questions asked by their chat partner. Unlike VQA [4], answers are not restricted to be short or concise, instead workers will be encouraged to reply as naturally and ‘conversationally’ as possible. An example dialog is shown in Fig.3c.

This process is an unconstrained ‘live’ chat, with the only exception that the questioner must wait to receive an answer before posting the next question. The workers are allowed to end the conversation after 20 messages are exchanged (10 pairs of questions and answers). Further details about our final interface can be found in the supplement.

We also piloted a different setup where the questioner saw a highly blurred version of the image, instead of the caption.

The conversations seeded with blurred images resulted in questions that were essentially ‘blob recognition’ –‘What is the pink patch at the bottom right?’. For our full-scale data-collection, we decided to seed with just the captions since it resulted in more ‘natural’ questions and more closely modeled the real-world applications discussed in Section1where no visual signal is available to the human.

Building a 2-person chat on AMT.Despite the popular-

ity of AMT as a data collection platform in computer vi- sion, our setup had to design for and overcome some unique challenges – the key issue being that AMT is simply not designed for multi-user Human Intelligence Tasks (HITs).

Hosting a live two-person chat on AMT meant that none of the Amazon tools could be used and we developed our own backend messaging and data-storage infrastructure based on Redis messaging queues and Node.js. To support data qual- ity, we ensured that a worker could not chat with themselves (using say, two different browser tabs) by maintaining a pool of worker IDs paired. To minimize wait time for one worker while the second was being searched for, we ensured that there was always a significant pool of available HITs. If one of the workers abandoned a HIT (or was disconnected) midway, automatic conditions in the code kicked in asking the remaining worker to either continue asking questions or providing facts (captions) about the image (depending on their role) till 10 messages were sent by them. Workers who completed the task in this way were fully compensated, but our backend discarded this data and automatically launched a new HIT on this image so a real two-person conversation could be recorded. Our entire data-collection infrastructure (front-end UI, chat interface, backend storage and messag- ing system, error handling protocols) will be publicly avail- able to help future efforts.

4. VisDial Dataset Analysis

We now analyze the v0.5 subset of our VisDial dataset col- lected so far – it contains 1 dialog (10 question-answer pairs) on 68k images from COCO (58ktrainand 10k val), or a total of 680,000 QA pairs.

4.1. Analyzing VisDial Questions

Visual Priming Bias.One key difference between VisDial and previous image question-answering datasets (VQA [4], Visual 7W [62], Baidu mQA [12]) is the lack of a ‘vi- sual priming bias’ in VisDial. Specifically, in all previ-

Figure 1: The Visual Dialogue Collection Task and an Example Dialogue (from (Das et al., 2017)) Figure 1 shows the environment in which the visual dialogue dataset (Das et al., 2017) was collected.

As the example dialogue on the right indicates, this rather artificial setting (“you have to ask questions about the image”) seem to encourage a pairwise structuring of question and answer. That the string of pairs forms a dialogue is only recognisable in the fact that each pair concerns a different aspect of the image, and that later questions may refer to entities previously mentioned. Since there is no way for the questioner to provide feedback on the answers, it is unlikely that a model could learn from data of this type that dialogue is more than a sequence of loosely related question/answer pairs, and that even such sequences typically would have structure in human dialogue. (For reasons of space, we cannot argue this point more deeply here.)

3 The MeetUp Task

In contrast, we designed the MeetUp task to elicit more structured dialogue. The task is based on a

dynamic environment with several “rooms” (in the instantiation presented here, represented as images)

where two dialogue participants (players) are placed in different rooms and have to find each other. As

the players cannot see each other, but can communicate (via text messages), the only way they can solve

the task is to establish verbally whether they both currently see the same room/image.

(2)

Figure 2: The scene discussed in the ex- cerpt below

Our set-up extends recent efforts along the following dimensions: 1) the task’s main goal can be defined inde- pendently of reference, in high-level communicative terms (namely “try to meet up in an unknown environment”), 2) the task is symmetric and does not need a rigid in- teraction protocol (there is no instruction giver/follower), 3) there is a clear division between private information (that only one player has access to) and public information (facts that have been publicly asserted), and reaching the goal involves moving information from the former state to the latter (i.e., it involves conversational grounding (Clark, 1996)), 4) reference can be made to things not currently

seen, if they have been introduced into the discourse earlier (see line 59, “I found the kitchen”). We have conducted a pilot data collection which indicates that this setting indeed leads to interesting dialogues.

We aim to collect a sufficient number of dialogues (in the thousands) in the upcoming weeks, in order to be able to train agents on this task. Project

URL

:

https://github.com/dsg-bielefeld/meetup

.

Time Private to A Public Private to B

31 (01:45) A: I am now in a kitchen withwood floorsanda posterthat saysCONTRATTO

. . . .

59 (02:50) B: Wait– I found the kitchen!

. . . .

60 (02:55) −→N kitchen

61 (02:55) You can go [/n]orth [/e]ast

[/s]outh [/w]est

62 (03:13) A: I am back in kitchen. It hasa white marble dining tablein center

63 (03:29) B: Yes. There arefour chairsonthe island.

64 (03:35) A: Exactly

65 (03:37) B: Andthe big Contratto poster.

66 (03:48) B:Three lightsabovethe island?

67 (03:53) A: yep

71 (04:05) B: /done

72 (04:07) A: /done

73 (04:10)

Well done! You are all indeed in the same room!

Table 1: (Discontinuous) excerpt from a MeetUp dialogue References

Raffaella Bernardi, Ruket Cakici, Desmond Elliott, Aykut Erdem, Erkut Erdem, Nazli Ikizler-Cinbis, Frank Keller, Adrian Muscat, and Barbara Plank. 2016. Automatic description generation from images: A survey of models, datasets, and evaluation measures.J. Artif. Int. Res., 55(1):409–442, January.

Xinlei Chen and C Lawrence Zitnick. 2015. Mind’s eye: A recurrent visual representation for image caption generation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 2422–2431.

Herbert H Clark. 1996. Using language. 1996.Cambridge University Press: Cambridge, pages 274–296.

Abhishek Das, Satwik Kottur, Khushi Gupta, Avi Singh, Deshraj Yadav, Jos´e MF Moura, Devi Parikh, and Dhruv Batra. 2017.

Visual dialog. InProceedings of the IEEE Conference on Computer Vision and Pattern Recognition, volume 2.

Harm De Vries, Florian Strub, Sarath Chandar, Olivier Pietquin, Hugo Larochelle, and Aaron Courville. 2017. Guesswhat?!

visual object discovery through multi-modal dialogue. InProc. of CVPR.

Jacob Devlin, Hao Cheng, Hao Fang, Saurabh Gupta, Li Deng, Xiaodong He, Geoffrey Zweig, and Margaret Mitchell. 2015.

Language models for image captioning: The quirks and what works. InProceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 2: Short Papers), pages 100–105, Beijing, China, July. Association for Computational Linguistics.

Hao Fang, Saurabh Gupta, Forrest Iandola, Rupesh Srivastava, Li Deng, Piotr Dollar, Jianfeng Gao, Xiaodong He, Margaret Mitchell, John Platt, Lawrence Zitnick, and Geoffrey Zweig. 2015. From captions to visual concepts and back. InProceed- ings of CVPR, Boston, MA, USA, June. IEEE.

Sahar Kazemzadeh, Vicente Ordonez, Mark Matten, and Tamara L Berg. 2014. ReferItGame: Referring to Objects in Pho- tographs of Natural Scenes. In Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP 2014), pages 787–798, Doha, Qatar.

Junhua Mao, Jonathan Huang, Alexander Toshev, Oana Camburu, Alan L. Yuille, and Kevin Murphy. 2015. Generation and comprehension of unambiguous object descriptions.CoRR, abs/1511.02283.

David Schlangen, Sina Zarriess, and Casey Kennington. 2016. Resolving references to objects in photographs using the words-as-classifiers model. InProceedings of the 54rd Annual Meeting of the Association for Computational Linguistics (ACL 2016).

Oriol Vinyals, Alexander Toshev, Samy Bengio, and Dumitru Erhan. 2015. Show and tell: A neural image caption generator.

InComputer Vision and Pattern Recognition.

Licheng Yu, Patrick Poirson, Shan Yang, Alexander C. Berg, and Tamara L. Berg, 2016. Modeling Context in Referring Expressions, pages 69–85. Springer International Publishing, Cham.

Referenzen

ÄHNLICHE DOKUMENTE

20 In the meantime, however, separate AUHIP-mediated talks between the government and SRF components (Sudan People’s Liberation Move- ment-North (SPLM-N) on the Two Areas,

the credibility and the acceptance of the OSCE as a mediator to the conflict parties; the political support of participating States; the mandate of an OSCE institution

Cooperation between LCPCs and the population is based on accountability (reporting), which takes place annually, together with local self-government and neighbourhood police

21 While EULEX (EU Rule of Law Mission in Kosovo) and Kosovo Customs maintained a shaky presence at the crossing points which allowed for the movement of goods across

Kosovo Liberation Army – both now prime ministers of their respective countries – to the table for direct talks and towards an accord in barely seven months is no small diplomatic

Turkey and the Gulf countries, however, have progressively developed strategic dialogue on pressing Middle East issues in an effort to achieve policy coordination, including

The model has been proposed for and used with K-12 to university teachers with varying teaching experiences (pre-service, in- service) and is also not dedicated to any

(c) t-SNE of the entire reference cohort, described in [2] confirms the affiliation of the two cases to the respective DNA methylation classes: glioblastoma, IDH wildtype, H3.3