icassp paper survey

24
1 ICASSP Paper Survey Presenter: Chen Yi-Ting

Upload: henry-pittman

Post on 01-Jan-2016

36 views

Category:

Documents


0 download

DESCRIPTION

ICASSP Paper Survey. Presenter: Chen Yi-Ting. Improved Spoken Document Retrieval With Dynamic Key Term Lexicon and Probabilistic Latent Semantic Analysis (PLSA) Improved Spoken Document Summarization Using Probabilistic Latent Semantic Analysis (PLSA) - PowerPoint PPT Presentation

TRANSCRIPT

Page 1: ICASSP Paper Survey

1

ICASSP Paper Survey

Presenter: Chen Yi-Ting

Page 2: ICASSP Paper Survey

2

• Improved Spoken Document Retrieval With Dynamic Key Term Lexicon and Probabilistic Latent Semantic Analysis (PLSA)

• Improved Spoken Document Summarization Using Probabilistic Latent Semantic Analysis (PLSA)

• Topic and Stylistic Adaptation for Speech Summarisation• Automatic Sentence Segmentation of Speech for Automa

tic Summarization

Page 3: ICASSP Paper Survey

3

Improved Spoken Document Retrieval With Dynamic Key Term Lexicon and Probabilistic Latent

Semantic Analysis (PLSA)

Page 4: ICASSP Paper Survey

4

• In this paper, using a “dynamic key term lexicon” automatically extracted from the ever-changing document archives as an extra feature set in the retrieval task.

Page 5: ICASSP Paper Survey

5

• An important part of the proposed approach is the automatic key term extraction from the archives

• The second important part of the proposed approach is the key term recognition from the user spoken query– Here special approaches to recognize correctly the key term in

the user query were developed, including emphasizing the possible key term candidates during search through the phone lattice, and key term matching using a phone similarity matrix including two different distance measures.

– Two different versions of models can be used: the general lexicon including all terms except those stop terms deleted and the other based on the much smaller but semantically rich key term lexicon

Page 6: ICASSP Paper Survey

6

• Named Entity Recogniion– The first is to recognize the NEs from a text document (or the tra

nscription of a spoken document) using global information– The second special approach used here is for spoken document

s to recover the OOV NEs using external knowledge

• Key Term extraction by term entropy based on PLSA

)(

)|(

)(

)()|()|(

j

kj

j

kkjjk tP

TtP

tP

TPTtPtTP

)|(log)|()(1

jk

K

kjkj tTPtTPtH

Page 7: ICASSP Paper Survey

7

• Key term recognition from user query– The user spoken query is transcribed not only into a word graph as usua

l recognition process, but into a phone lattice as well

– Then matching the phone lattice with the phone sequences of the key terms in the dynamic lexicon using dynamic programming (threshold)

– The price paid here is of course the overall word error rate may be increased

• The experimental conditions– Word error rates 27%、 character error rates 14.29%、 syllable error ra

te 8.91%

– 32 topics were used in PLSA modeling

– 1000new stories (test set) 50 queries

– The length of the queries is roughly 8-11 words

– A lexicon of 61521 word was used here

– A total of 1708 NE were obtained (from 9836 news stories)

– Picked up the top 2000 terms ranked by term entropy

Page 8: ICASSP Paper Survey

8

• Experimental Results

Page 9: ICASSP Paper Survey

9

Improved Spoken Document Summarization Using Probabilistic Latent Semantic Analysis (PLSA)

Page 10: ICASSP Paper Survey

10

• where some statistical measure s(tj) (such as TF/IDF or the like)linguistic measure l(tj) (e.g., named entities)c(tj) is calculated from the confidence scoreg(tj) is N-gram score for the term tj b(S) is calculated from the grammatical structure of the sentence Sλ1, λ2, λ3, λ4 and λ5 are weighting parameters

• Two useful measures, referred to as topic significance and term entropy in this paper are proposed based on the PLSA modeling to determine the terms and thus sentences important for the document which can then be used to construct the summary

• The statistical measure s(tj) which has been proved extremely useful is called “significance score”:

)()()()()(1

)( 51

4321 Sbtgtctltsn

SIn

jjjjj

t

Aijj F

Fdtnts log),()(

Page 11: ICASSP Paper Survey

11

• Topic significance– The topic significance score of a term tj with respect to a topic Tk

– The statistical measure:

• Term Entropy

– is a scaling factor

klTil

ik

Ddijkt

l

i

j dTP

dTPdtnTS

,

)|(

)|(),()(

)|()()(1

ikk

K

ktj dTPTStsj

)(

)|(

)(

)()|()|(

j

kj

j

kkjjk tP

TtP

tP

TPTtPtTP

K

kjkjkijjEN tTPtTPdtnts

1

)|(log)|(),()(

)|( jk tTP )|( jk tTP

Page 12: ICASSP Paper Survey

12

• Experiments configuration– The test corpus included 200 news stories broadcast– Word accuracy 66.46%、 character accuracy 74.95%、 syllable

accuracy 81.70%– Sentence recall/precision is the evaluation metric for automatic s

ummarization of documents

Page 13: ICASSP Paper Survey

13

• Experiments configuration

Page 14: ICASSP Paper Survey

14

Topic and Stylistic Adaptation for Speech Summarisation

Page 15: ICASSP Paper Survey

15

• In this paper they investigate LiM topic and stylistic adaptation using combinations of LiMs each trained on different adaptation data

• Focusing on adapting the linguistic component, which is not related at all to the language model used during the recognition process, to make it more suited for the summarisation task

• Experiments were performed both on spontaneous speech, using 9 talks from the Translanguage English Database (TED) corpus, and speech read from text, using 5 talks from CNN broadcast news from 1995

• The measure of summary quality used in this paper is summarisation accuracy (SumACCY)

[%]100*/)( LenDelInsSubLenAccuracy

Page 16: ICASSP Paper Survey

16

• Automatic speech summarisation system

Page 17: ICASSP Paper Survey

17

• Summarisation Method• Important sentences are first extracted according to the followi

ng score for each sentence, obtained from the automatic speech recognition out (ASR)

• Starting with a baseline LiM (LiMB) we perform LiM adaptation by linearly interpolating the baseline model with other component models trained on different data

Where • Different types of component LiM are built, coming from differe

nt sources of data, and using either unigram, bigram or trigram information

n

iiLiIiC wLwIwC

NWS

1

)()()(1

)(

)....|(.....

)....|()....|(

11

111111

iniinn

iniiinii

wwwP

wwwPwwwP

k k 1

)....|(log)( 11 iniii wwwPwL

Page 18: ICASSP Paper Survey

18

• Experimental Setup– Due to lack of data we had to use the talks both for development

and evaluation with a rotating form of cross-validation: all talks but one are used for development, the remaining talk being used for testing

– Summaries from the development talks are generated automatically by the system using different sets of parameters

– For the TED data, two type of component linguistic models:• The first type are built on the small corpus of hand-made sum

maries, made for the same summarisation ratio.• The second type are built from the papers in the conference

proceeding for talk we want to summarise– For the CNN data, one type of component linguistic models

• the small corpus of hand-made summaries

Page 19: ICASSP Paper Survey

19

• Experimental Setup– Reference results: random summarisation, the humman summari

es and the baseline

CNNTED

Page 20: ICASSP Paper Survey

20

Page 21: ICASSP Paper Survey

21

Automatic Sentence Segmentation of Speech for Automatic Summarization

Page 22: ICASSP Paper Survey

22

• This paper presents an automatic sentence segmentation method for an automatic speech summarization system

• The segmentation method is based on combining word- and class-based statistical language models to predict sentence and non-sentence boundaries

• Studying both the effect of the segmentation on the sentence segmentation system itself and the effect of the segmentation on the summarization accuracy

• To judge the quality of the sentence segmentation we used the F-measure metrics

Page 23: ICASSP Paper Survey

23

• Automatic sentence segmentation

• This probability was combined with the matching recursive path probability from

• Three LMs were used in sentence segmentation, two word-based LMs and a class-based LM

• The LMs were combined by linear interpolation as follows:

)|()|()....(

)|()|()....()....(

1211

111 1

SwpwwSpwwP

SwpSSpwwPwwP

iiiiSNO

iwiSiS i

)|()....(

)|()....()....(

1211

111 1

iiiiSNO

wiiSiSNO

wwwpwwP

SwpwwPwwPi

1iw

))(),...,(|)(())(|()|(

),....,|()|(

)|()|(

11

11

iniiiiim

iniiim

mimmi

wCwCwCPwCwPhwP

orwwwPhwP

hwPhwP

Page 24: ICASSP Paper Survey

24

• Experimental results