crowdtruth poster @iswc2014

Post on 06-Dec-2014

243 Views

Category:

Technology

0 Downloads

Preview:

Click to see full reader

DESCRIPTION

Presented at the ISWC2014, RDBS Track

TRANSCRIPT

Crowd Truth Machine-Human Computation Framework for Harnessing Disagreement in Semantic Interpretation

Goal: gather ground truth data to train, test, evaluate cognitive computing systems

Humanannotationtasks:

use case #1: text annotation

concept: drug

concept: disease

relation:

treats

use case #2: image annotation use case #3: video annotation

Batavia

Army

Military

Indonesia

Triangle ofdisagreement

Oana Inel, Khalid Khamkham, Tatiana Cristea, Arne Rutjes, Jelle van der Ploeg, Lora Aroyo,Robert-Jan Sips, Anca Dumitrache and Lukasz Romaszko

crowdsourcing task

sentence vector

unit vectors for the same sentence

unit vector

sentence-annotation score

● sentence-annotation score:measures how clearly the annotation is expressed in the sentence

● sentence clarity:measures the maximum sentence-annotation score for the sentence

VS. The CrowdTruth approach

ask a large crowd

● allows for different interpretations● minimal instructions● large crowds of annotators● harnessing disagreement● continuously updated with new data

Traditional Ground Truth approach

ask few experts

● assumes one correct interpretation● guidelines limit interpretations● examples evaluated by single expert● eliminating disagreement● ground truth reused over time

Approach:

● worker-sentence score: measures quality of worker for one sentence

● worker-worker disagreement:● measures pairwise agreement between workers● average worker agreement:

measures overall worker quality

worker-sentence score

Disagreementmetrics:

(use case #1:text annotation)

top related