• Keine Ergebnisse gefunden

Improved  Vote  Aggregation  Techniques  for  the  Geo-­Wiki   Cropland  Capture  Crowdsourcing  Game  

N/A
N/A
Protected

Academic year: 2022

Aktie "Improved  Vote  Aggregation  Techniques  for  the  Geo-­Wiki   Cropland  Capture  Crowdsourcing  Game  "

Copied!
1
0
0

Wird geladen.... (Jetzt Volltext ansehen)

Volltext

(1)

Improved  Vote  Aggregation  Techniques  for  the  Geo-­Wiki   Cropland  Capture  Crowdsourcing  Game  

Data preprocessing

Baklanov Artem

1,2

, Fritz Steffen

1

, Khachay Mikhail

2

, Nurmukhametov Oleg

2

Salk Carl

1

,See Linda

1

, and Shchepashchenko Dmitry

1

.

1 — International Institute for Applied Systems Analysis (IIASA);; Schlossplatz 1, Laxenburg, Austria, A-­2361;;

2 — N.N. Krasovskii Institute of Mathematics and Mechanics (Russian Academy of Sciences).

Over 5 million opinions from

non-experts

Expert-quality decisions about

190 000 images

HOW?

Challenge Results

1) Detection of similar images using pHash (perceptual hash) [Zauner, 2010].

è5% of images are not unique

2) Detection of low quality images using Blur detection algorithm [H Tang, 2012].

è2% of images are discarded

95% 98% 99%

0% 10% 30%

Volunteers’  ROCs Benchmark

Land cover map

We compare machine learning algorithms and state-­of-­the-­art vote aggregation algorithms:

EM [Dawid, 1979];;

KOS, KOS+ [Karger, 2011];;

Hard Penalty [Jagabathula, 2014].

ü Improved quality of image dataset;;

ü Improved majority voting estimates;;

ü Benchmarked state-­of-­the-­art algorithms;;

ü Demonstrated that these algorithms perform on a par with majority voting.

Explanation: all volunteers are reliable, the task assignment is highly irregular.

ü Accuracy is 96% for images with more than 9 votes.

We  increased  the  accuracy  of  “Cropland   Capture”  data  from   76%  to  91%

The  Cropland  Capture Game

How  to  aggregate  votes  from  non-­experts?

Approach

Individual performance of volunteers is studied with respect to the number of votes [Rayker, 2012].

Spammers

Malicious   Annotators Good  

Annotators Biased  

Annotators

Biased Annotators

Spammers

Malicious   Annotators Good  

Annotators Biased  

Annotators

Biased Annotators

Spammers

Malicious   Annotators Good  

Annotators Biased  

Annotators

Biased Annotators

Spammers

Malicious   Annotators Good  

Annotators Biased  

Annotators

Biased Annotators

ü There  are  no  spammers  among volunteers  with  more  than  12  

votes;;  

ü Good  volunteers  prevail;;  

ü Volunteers  with  >100  votes  show   higher  accuracy  than  any  tested   algorithm.  

*We  use  publicly  available  code  (https://github.com/ashwin90/Penalty-­based-­clustering)  

3:

Referenzen

ÄHNLICHE DOKUMENTE

In addition to validation, the data presented here also represent a valuable training tool that can be used to develop new land cover or cropland extent maps as well as to

In this paper, on the basis of several known machine learning techniques, we propose a technical approach to improve the overall performance of the majority voting decision rule used

This paper presents a case study that aims to compare the performance of several state-of-the-art vote aggregation techniques specifically developed for the analysis of

‘cropland’ and ‘non-cropland.’ Difficult images came from several categories: images with many ratings in both the non-cropland and cropland categories, images with many

Data collected from Geo-Wiki crowdsourcing campaigns and serious gaming applications have led to improved global maps of cropland, forest cover, and land cover distributions..

1) Detection of similar images using pHash (perceptual hash) [Zauner, 2010].. è5% of images are

A second product was created using a similar methodology but at a global level and with a number of modifications such as the integration of many more input data sets

As many types of crowdsourced data are georeferenced, the term volunteer- ed geographic information (VGI) is also used to describe these types of contributions [16],