simnet: stepwise image-topic merging network for ... · topic extractor: multiple instance learning...
Post on 21-Jul-2020
15 Views
Preview:
TRANSCRIPT
|X
simNet: Stepwise Image-Topic Merging Network for
Generating Detailed and Comprehensive Image Captions
1
Fenglin Liu¹, Xuancheng Ren²*, Yuanxin Liu¹, Houfeng Wang² and Xu Sun²
¹ Beijing University of Posts and Telecommunications, Beijing, China
² Peking University, Beijing, China
* Equal Contributions
|X
CONTENTS
1 Introduction
2 Approach
3 Experiment
4 Analysis
|X
Introduction
|X
simNet: Stepwise Image-Topic Merging Network for Generating Detailed and Comprehensive Image Captions
4
·Soft-Attention: Show, attend and tell: Neural image caption generation with visual attention. In PMLR 2015
·ATT-FCN : Image captioning with semantic attention. In CVPR 2016
Figure 1: Examples of using different attention mechanisms.
|X
Introduction: Soft-Attention
5
Image Image Features Caption
encode decodeSoft-Attention:
·Soft-Attention: Show, attend and tell: Neural image caption generation with visual attention. In PMLR 2015
omitting “dog” and “mouse”
|X
Introduction: ATT-FCN
6
Image Image Keywords Caption
encode decodeATT-FCN:
·ATT-FCN : Image captioning with semantic attention. In CVPR 2016
missing “open” and mislocating “dog”
|X
Introduction: SimNet
7
Image
Image Features
decode
Image Keywords
encode CaptionsimNet:
|X
Introduction: Main idea
8
The visual information captured by CNN
The topics extracted by a topic extractor
The merging gate then adaptively adjusts the weight between visual attention and topic attention
Figure 2: Illustration of the main idea.
|X
Contributions
9
We propose a novel approach that can effectively merge the information in the image and the topics.
The generated captions are both detailed and comprehensive.
The proposed approach outperforms previous works in terms of SPICE, which correlates the best with human judgments.
|X
Approach
|X 11
Overview
|X 12
Approach: Image Encoder
Image Encoder: ResNet152
He et al., 2016: Deep residual learning for image recognition. In CVPR 2016.
|X 13
Approach: Image Encoder
Feature map:
|X 14
Approach: Topic Extractor
Topic Extractor: Multiple Instance Learning
Zhang et al., 2006: Multiple instance boosting for object detection. In NIPS 2006.
Fang et al., 2015: From captions to visual concepts and back. In CVPR2015
|X 15
Approach: Input Attention
Input Attention:
Xu et al., 2015 : Show, attend and tell: Neural image caption generation with visual attention. In PMLR 2015
|X 16
Approach: Output Attention
Output Attention:
You et al., 2016 : Image captioning with semantic attention. In CVPR 2016
Lu et al ., 2017 : Knowing when to look: Adaptive attention via a visual sentinel for image captioning. In CVPR 2017
|X 17
Approach: Visual Information
the visual information:
Output Attention:
|X 18
Approach: Previous Topic Attention
Topic Attention (Previous work):
Lacking the attentive visual informationwhen selecting topic!
You et al., 2016 : Image captioning with semantic attention. In CVPR 2016
|X 19
Approach: Our Topic Attention
Topic Attention (Our):
|X 20
Approach: Contextual Information
Topic Attention (Our):
the contextual information:
|X 21
Approach: Merging Gate
How to make full use of the visual information and the contextual information?
|X 22
Approach: Merging Gate
Visual information(e.g., “behind”, “red” is better)
Contextual information(e.g., “people”, “table” is better)
|X 23
Approach: Merging Gate
( Where σ is the sigmoid function )
|X 24
Approach: Merging Gate
The scoring function S is designed to evaluate the importance of the topic attention.
|X 25
Approach: Merging Gate
|X 26
Approach: Merging Gate
Share Weights
Share Weights
|X 27
Generating Words
the contextual information:
|X
Experiments
|X
Experiments
29
Microsoft COCO(MSCOCO) and Flickr30k
DatasetEvaluation Metrics
Sparrow bird on branch, withbeak inspecting leaves on branch.
A bird sitting on the branch of atree near leaves.
A bird that is sitting in a tree. a bird sitting on a branch of a
tree. a bird that is on a small branch
of a tree.
SPICE
CIDEr
BLEU
METEOR
ROUGE
Correlates the best with human Judgments !
|X
Experiments: Results (MSCOCO)
30
ComparableModels
|X
Experiments: Results (MSCOCO)
31
Competitive
|X
Analysis
|X
Analysis: The Contributions of The Sub-modules
33
Comprehensiveness Detailedness
|X
Analysis: Output Attention
34
The output attention is much more effective than the input attention
|X
Analysis: Visual Attention
35
A combination of the input attention and the output attention makes the results even better
|X
Analysis: Topic Attention
36
The topic attention is better at identifying objects but worse at identifying attributes.
|X
Analysis: Visual Attention + Topic Attention
37
Combing the visual attention and the topic attention directly results in a huge boost in performance
|X
Analysis: Full Model
38
Applying the merging gate is essential to the overall performance.
|X
Analysis: Visualization
39
The upper part shows the attention weights of each of 5 extracted topics. Deeper color means larger in value. The middle part shows the value of the merging gate. Determines the importance of the topic attention. The lower part shows the visualization of visual attention. The blue shade indicates the output attention.
The red shade indicates the input attention.
|X
Analysis: Visualization
40
Visual information “chair” is more important than contextual information “bed”
|X
Analysis: Examples
41
erroneoustopic “kitchen”
lacking “mouse”
missing “wooden” error count
|X
Conclusion
42
Stepwise image-topic merging network can adaptively combine the visual and the semantic attention to achieve substantial improvements.
The generated captions are both detailed and comprehensive
Our approach outperforms previous works in terms of SPICE on COCO and Flickr datasets.
|X 43
Thank you!If you have any questions about our paper, you can send a email to lfl@bupt.edu.cn
|X
Experiments: Results (Flickr30k)
44
|X
Analysis: Topic Extraction
45
The reason of using objects as topics is that they are easier to identify so that the generation suffers less from erroneous predictions.
It proves that for semantic attention, it is also important to limit the absolute number of incorrect visual words instead of merely the precision or the recall.
|X
Analysis: Merging Gate
46
|X
Analysis: Error Analysis
47
There are mainly three types of errors, i.e. distance (32, 26%), movement (22,18%), and object(60, 49%), with 9 (7%) other errors.
|X
Analysis: Error Analysis
48
There are mainly three types of errors, i.e. distance (32, 26%), movement (22,18%), and object(60, 49%), with 9 (7%) other errors.
|X
Experiments: Results (MSCOCO)
49
|X
Analysis: Topic Attention
50
top related