discourse annotation for improving spoken dialogue systems

22
Discourse Annotation for Improving Spoken Dialogue Systems Joel Tetreault, Mary Swift, Preethum Prithviraj, Myroslava Dzikovska, James Allen University of Rochester Department of Computer Science ACL Workshop on Discourse Annotation July 25, 2004

Upload: valencia-cruz

Post on 31-Dec-2015

34 views

Category:

Documents


0 download

DESCRIPTION

Discourse Annotation for Improving Spoken Dialogue Systems. Joel Tetreault, Mary Swift, Preethum Prithviraj, Myroslava Dzikovska, James Allen University of Rochester Department of Computer Science ACL Workshop on Discourse Annotation July 25, 2004. Reference in Spoken Dialogue. - PowerPoint PPT Presentation

TRANSCRIPT

Page 1: Discourse Annotation for Improving Spoken Dialogue Systems

Discourse Annotation for Improving Spoken Dialogue Systems

Joel Tetreault, Mary Swift, Preethum Prithviraj, Myroslava Dzikovska, James Allen University of RochesterDepartment of Computer ScienceACL Workshop on Discourse AnnotationJuly 25, 2004

Page 2: Discourse Annotation for Improving Spoken Dialogue Systems

Reference in Spoken Dialogue Resolving anaphoric expressions correctly is critical

in task-oriented domains Makes conversation easier for humans

Reference resolution module provides feedback to other components in system Ie. Incremental Parsing, Interpretation Module

Investigate how to improve RRM: Does deep semantic information provide an improvement

over syntactic approaches? Discourse Structure could be effective in reducing search

space of antecedents and improving accuracy (Grosz and Sidner, 1986)

Page 3: Discourse Annotation for Improving Spoken Dialogue Systems

Goal

Construct a linguistically rich parsed corpus to test algorithms and theories on reference in spoken dialogue, to provide overall system improvement Implicit roles

Paucity of empirical work on reference in spoken dialogue (Bryon and Stent 1998, Eckert & Strube, 2000; etc.)

Page 4: Discourse Annotation for Improving Spoken Dialogue Systems

Outline

Corpus Construction Parsing Monroe Domain Reference Annotation Dialogue Structure Annotation

Results Personal pronoun evaluation Dialogue Structure

Summary

Page 5: Discourse Annotation for Improving Spoken Dialogue Systems

Parsing Monroe Domain Domain: Monroe Corpus of 20 transcriptions (Stent, 2001) of

human subjects collaborating on Emergency Rescue 911 tasks Each dialogue was at least 10 minutes long, and most were over

300 utterances long Work presented here focuses on 5 of the dialogues 17 (1756

utterances) Goals: develop a corpus of sentences parsed with rich syntactic,

semantic, discourse information to: Improve TRIPS parser (Swift et al., 2004) Train statistical parser for comparison with existing parser Develop incremental parser (Stoness et al., 2004) Develop automated techniques for marking repairs

Page 6: Discourse Annotation for Improving Spoken Dialogue Systems

Parser information for Reference Rich parser output is helpful for discourse

annotation and reference resolution: Referring expressions identified (pronoun, NP, impros) Verb roles and temporal information (tense, aspect)

identified Noun phrases have semantic information associated

with them Speech act information (question, acknowledgment) Discourse markers (so, but) Semi-automatic annotation increases reliability

Page 7: Discourse Annotation for Improving Spoken Dialogue Systems

Monroe Corpus Example

UTT SPK:SA TEXTUtt53 S: TELL and so we're going to take an ambulance

from saint mary's hospital Utt54 U: TELL oh you never told me about the ambulancesUtt55 U: WH-QU how many do you have Utt56 S: TELL there's one at saint mary's hospital and

two at rochester general hospitalUtt57 U: IDENTIFY two Utt58 U: CONFIRM okay Utt59 S: TELL and we're going to take an ambulance from

saint mary's to east main street Utt60 S: CCA and that is as far as i have planned Utt61 U: CONFIRM okay Utt62A U: CONFIRM okay

Page 8: Discourse Annotation for Improving Spoken Dialogue Systems

TRIPS Parser

Broad-coverage, deep parser Uses bottom-up algorithm with CFG and

domain independent ontology combined with a domain model

Flat unscoped LF with events and labeled semantic roles based on FrameNet

Semantic information for noun phrases based on EuroWordNet

Page 9: Discourse Annotation for Improving Spoken Dialogue Systems

Semantics Example: “an ambulance” (TERM :VAR V213818

:LF (A V213818 (:* LF::LAND-VEHICLE W::AMBULANCE) :INPUT (AN AMBULANCE))

:SEM ($ F::PHYS-OBJ (SPATIAL-ABSTRACTION SPATIAL-POINT)

(GROUP -) (MOBILITY LAND-MOVABLE) (FORM ENCLOSURE) (ORIGIN ARTIFACT) (OBJECT-FUNCTION VEHICLE) (INTENTIONAL -) (INFORMATION -) (CONTAINER (OR + -))

(TRAJECTORY -)))

Page 10: Discourse Annotation for Improving Spoken Dialogue Systems

Semantic Representations for “Them” “and then send them to Strong Hospital” (TERM :VAR V3337536

:LF (PRO V3337536

(SET-OF (:* LF::REFERENTIAL-SEM THEM))

:SEM ($ F::PHYS-OBJ (F::MOBILITY F::MOVABLE)))

Page 11: Discourse Annotation for Improving Spoken Dialogue Systems

Corpus Construction Mark sentence status (ungrammatical, incomplete,

conjoined) and mark speech repairs Parse with domain-specific semantic restrictions for

better coverage Handcheck sentences, marking GOOD or BAD

Criteria for GOOD: both syntactic and semantic must be correct

Update parser to cover BAD cases Reparse and repeat handchecking

DataCollection

CorpusAnnotation

RunParser

ManualUpdate

ParserUpdate

Reparse &Merge

Page 12: Discourse Annotation for Improving Spoken Dialogue Systems

Current CoverageCorpus % Good Good Bad NA Total

S2 90.8% 325 34 37 405

S4 76.1% 246 78 61 388

S12 89.9% 151 17 21 189

S16 84.2% 298 56 29 383

S17 85.2% 311 54 26 392

Overall 84.1% 1331 239 174 1757

Page 13: Discourse Annotation for Improving Spoken Dialogue Systems

Reference Annotation

Annotated dialogues for reference w/undergraduate researchers (created a Java Tool: PronounTool)

Markables determined by LF terms Identification numbers determined by :VAR field of LF

term Used stand-off file to encode what each pronoun refers

to (refers-to) and the relation between pronoun and antecedent (relation)

Post-processing phase assigns an unique identification number to coreference chains

Also annotated coreference between definite noun phrases

Page 14: Discourse Annotation for Improving Spoken Dialogue Systems

Reference Annotation

Used slightly modified MATE scheme: pronouns divided into the following types: IDENTITY (Coreference) (278) FUNCTIONAL (20) PROPOSITON/D.DEXEIS (41) ACTION/EVENT (22) INDEXICAL (417) EXPLETIVE (97) DIFFICULT (5)

Page 15: Discourse Annotation for Improving Spoken Dialogue Systems

Dialogue Structure

How to integrate discourse structure into a reference module? Is it worth it?

Shallow techniques may work better: may not be necessary to get a fine embedding to improve reference resolution

Implemented QUD-based technique and Dialogue Act model (Eckert and Strube, 2000)

Annotated in a stand-off file

Page 16: Discourse Annotation for Improving Spoken Dialogue Systems

literal QUD

Questions Under Discussion (Craige Roberts, Jonathan Ginzburg) – questions or modals can be viewed as creating a discourse segment

Result – questions provide a shallow discourse structuring, but that maybe enough to improve performance

Entities in QUD main segment can be viewed as the topic Segment closed when question is answered (use ack

sequences, change in entities used) only entities from answer and entities in question are

accessible Can be used in TRIPS to reduce search space of entities –

set context size

Page 17: Discourse Annotation for Improving Spoken Dialogue Systems

QUD Annotation Scheme

Annotate: Start utterance End utterance Type (aside, repeated question, unanswered, nil)

Page 18: Discourse Annotation for Improving Spoken Dialogue Systems

QUD

Issue 1: easy to detect Q’s (use Speech-Act information), but how do you know Q is answered?

Cue words, multiple acknowledgements, changes in entities discussed provide strong clues that question is finishing

Issue 2: what is more salient to a QUD pronoun – the QUD topic or a more recent entity?

Page 19: Discourse Annotation for Improving Spoken Dialogue Systems

Dialogue Act Segmentation

Utterances that are not acknowledged by the listener may not be in common ground and thus not accessible to pronominal reference

Evaluation showed improvement for pronouns referring to abstract entities, and strong annotator reliability

Each utterance marked as I: contains content (initiation), A: acknowledgment, C: combination of the above

Page 20: Discourse Annotation for Improving Spoken Dialogue Systems

Results

Incorporating semantics into reference resolution algorithm (LRC) improves performance from 61.5% to 66.9% (CATALOG ’04)

Preliminary QUD results show an additional boost to 67.3% (DAARC ’04)

E&S Automated: 63.4% E&S Manual: 60.0%

Page 21: Discourse Annotation for Improving Spoken Dialogue Systems

Issues

Inter-annotator agreement for QUD annotation Segment ends are hardest to synch

Ungrammatical and fragmented utterances: Parse automatically or manually?

Small corpus size: need more data for statistical evaluations

Parser freeze? important for annotators to stay abreast of latest changes

Page 22: Discourse Annotation for Improving Spoken Dialogue Systems

Summary

Semi-automated parsing process to produce reliable discourse annotation

Discourse annotation done manually, but automated data helps guide manual annotation

Result: spoken dialogue corpus with rich linguistic data