transfer learning for bioimaging and bilingual ... learning for bioimaging and bilingual...

137
Transfer Learning for BioImaging and Bilingual Applications by Qian Sun A Dissertation Presented in Partial Fulfillment of the Requirements for the Degree Doctor of Philosophy Approved November 2015 by the Graduate Supervisory Committee: Jieping Ye, Co-Chair Guoliang Xue, Co-Chair Huan Liu Jing Li ARIZONA STATE UNIVERSITY December 2015

Upload: lemien

Post on 16-May-2018

217 views

Category:

Documents


1 download

TRANSCRIPT

Page 1: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Transfer Learning for BioImaging and Bilingual Applications

by

Qian Sun

A Dissertation Presented in Partial Fulfillmentof the Requirements for the Degree

Doctor of Philosophy

Approved November 2015 by theGraduate Supervisory Committee:

Jieping Ye, Co-ChairGuoliang Xue, Co-Chair

Huan LiuJing Li

ARIZONA STATE UNIVERSITY

December 2015

Page 2: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

ABSTRACT

Discriminative learning when training and test data belong to different distributions

is a challenging and complex task. Often times we have very few or no labeled data

from the test or target distribution, but we may have plenty of labeled data from one

or multiple related sources with different distributions. Due to its capability of mi-

grating knowledge from related domains, transfer learning has shown to be effective

for cross-domain learning problems. In this dissertation, I carry out research along

this direction with a particular focus on designing efficient and effective algorithms for

BioImaging and Bilingual applications. Specifically, I propose deep transfer learning

algorithms which combine transfer learning and deep learning to improve image anno-

tation performance. Firstly, I propose to generate the deep features for the Drosophila

embryo images via pretrained deep models and build linear classifiers on top of the

deep features. Secondly, I propose to fine-tune the pretrained model with a small

amount of labeled images. The time complexity and performance of deep transfer

learning methodologies are investigated. Promising results have demonstrated the

knowledge transfer ability of proposed deep transfer algorithms. Moreover, I pro-

pose a novel Robust Principal Component Analysis (RPCA) approach to process the

noisy images in advance. In addition, I also present a two-stage re-weighting frame-

work for general domain adaptation problems. The distribution of source domain is

mapped towards the target domain in the first stage, and an adaptive learning model

is proposed in the second stage to incorporate label information from the target do-

main if it is available. Then the proposed model is applied to tackle cross-lingual

spam detection problem at LinkedIn’s website. Our experimental results on real data

demonstrate the efficiency and effectiveness of the proposed algorithms.

i

Page 3: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

To my parents

and to my husband Shuo

for their love

and supporting me all the way.

ii

Page 4: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

ACKNOWLEDGMENTS

First and foremost, I would like to gratefully thank my Ph.D advisor Dr. Jieping

Ye, for his guidance, understanding and patience during my Ph.D studies at ASU.

His helps are always around, from research methodologies, technical writing, and pre-

sentation skill. This dissertation would have never been possible without his support.

I would also like to express my sincere gratitude to my committee members: Dr.

Guoliang Xue, Dr. Huan Liu and Dr. Jing Li, for their guidance, support and

encouragement. Dr. Xue’s optimization class helps me to explore efficient solutions

for my proposed machine learning models. Dr. Liu’s Data Mining class is the first

course related to my research I took at ASU and it plays a fundamental role in all

my thesis work. Dr. Li provides a lot of valuable discussions and suggestions during

the thesis work.

My completion of this thesis could not have been accomplished without the sup-

port from my colleagues and friends in Dr. Jieping Ye’s research group. They are one

of the sources for incessant innovation and I benefit from them on a day to day basis.

Many thanks to: Jun Liu, Liang Sun, Shuiwang Ji, Jianhui Chen, Lei Yuan, Rita

Chattopadhyay, Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, Qingyang Li, Tao Yang,

Chao Zhang, Zheng Wang, Binbin Lin, Pinghua Gong, Jie Wang, Sherin Muckatira,

and Kefei Liu.

My time at ASU was made enjoyable in another large part due to many other

friends. I want to thank: Xilun Chen, Huiji Gao, Shanshan Liang, Liqi Ren, Yilin

Wang, Xiaolan Wang. Last but not least, no word could express my appreciation to

my family. Unconditional and invaluable love from my parents makes me feel I am

never alone. And thank my husband Shuo for being supportive and accompanying

me during this process.

iii

Page 5: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

TABLE OF CONTENTS

Page

LIST OF TABLES . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . vii

LIST OF FIGURES . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ix

CHAPTER

1 INTRODUCTION . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1

1.1 Transfer Learning . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1

1.2 Background . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3

1.2.1 BioImaging Annotation. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3

1.2.2 Bilingual Spam Detection . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7

2 ROBUST PRINCIPLE COMPONENT ANALYSIS FOR IMAGE PRO-

CESSING . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12

2.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12

2.2 Problem Formulation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15

2.2.1 Capped Norms. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15

2.2.2 Proposed Non-Convex Formulation . . . . . . . . . . . . . . . . . . . . . . . . 16

2.3 A DC Programming-Based Algorithm . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17

2.3.1 Solving the Sub-Problem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19

2.4 A Fast Alternating Algorithm . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23

2.4.1 Computing the Optimal Y . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24

2.4.2 Computing the Optimal X . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26

2.5 Experimental Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28

2.5.1 Synthetic Data . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28

2.5.2 Foreground and Background Separation on Surveillance Video 30

2.5.3 Shadows Removing from Face Images . . . . . . . . . . . . . . . . . . . . . 31

2.6 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32

iv

Page 6: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

CHAPTER Page

3 KEYWORDS PREDICTION FOR DROSOPHILA EMBRYO IMAGES . 39

3.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40

3.1.1 Image Representation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40

3.1.2 Sqrt Pooling . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 45

3.1.3 Keywords Annotation Schemes . . . . . . . . . . . . . . . . . . . . . . . . . . . . 46

3.1.4 Undersampling and Majority Vote. . . . . . . . . . . . . . . . . . . . . . . . . 49

3.2 Results and Discussion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 50

3.3 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 54

4 DEEP TRANSFER LEARNING FOR DROSOPHILA EMBRYO AN-

NOTATION. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 65

4.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 65

4.2 Deep Learning . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 67

4.3 Deep Transfer Learning . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 73

4.3.1 Deep Features . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 75

4.3.2 Stage Classification with Deep Features . . . . . . . . . . . . . . . . . . . . 81

4.3.3 Fine-Tuning . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 84

4.3.4 Fine-Tuning VS Training-From-Scratch . . . . . . . . . . . . . . . . . . . . 87

4.4 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 89

5 A TWO-STAGE TRANSFER LEARNING ALGORITHM . . . . . . . . . . . . . 90

5.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 90

5.2 Proposed Pipeline . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 90

5.2.1 Feature Generation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 91

5.2.2 Feature Selection . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 92

5.3 Transfer Learning . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 95

v

Page 7: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

CHAPTER Page

5.3.1 Instance Re-weighting . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 96

5.3.2 Adaptive Learning . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 98

5.4 Experiments . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 100

5.4.1 Experimental Settings . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 101

5.4.2 Experimental Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 103

5.4.3 Use of Less Unlabeled Samples in H . . . . . . . . . . . . . . . . . . . . . . . 106

5.5 Related Work . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 107

5.6 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 108

6 CONCLUSION . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 110

6.1 Summary of Contributions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 110

6.2 Future Directions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 111

REFERENCES . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 113

vi

Page 8: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

LIST OF TABLES

Table Page

2.1 Synthetic Results for Varying n with Fixed ρ = 0.001 and cr = 0.05.

Our Algorithms Perform Much Better Than ALM and NSA in Terms

of Capturing Sparse Locations in Y . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 34

2.2 Synthetic Results for Varying ρ with Fixed n = 100 and cr = 0.05.

Our Algorithms Perform Much Better than ALM and NSA in Terms

of Capturing Sparse Locations in Y . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 35

2.3 Synthetic Results for Varying cr with Fixed n = 100 and ρ = 0.001.

Our Algorithms Perform Much Better than ALM and NSA in Terms

of Capturing Sparse Locations in Y . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 36

2.4 Recovery Results of Airport Surveillance. Our Algorithm Produces X

with a Lower Rank and Y with a Smaller ℓ0-norm, While Keeping the

Relative Error Comparable with ALM and NSA. . . . . . . . . . . . . . . . . . . . . . 37

2.5 YaleFaceB Recovery Results. Our Algorithm Produces X with a Lower

Rank and Y with a Smaller ℓ0-norm, While Keeping the Relative Error

Comparable with ALM and NSA. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 37

3.1 In This Table, We Show the Total Number of Groups and the Corre-

sponding Groups which Contain the Top 10 Terms. For Most of the

Terms, the Number of Positive Samples (Groups Which Contain the

Term) Is Less Than the Number of Negative Samples (Groups Which

Do Not Contain the Term). . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 55

3.2 NonSpatial VS Spatial Group-Level Annotation for Top 10 Terms in

All Stages. Spatial BOW Performs Slightly Better than NonSpatial

BOW.. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 57

vii

Page 9: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Table Page

3.3 Image-Level with Union vs Image-Level without Union. Image-Level

with Union Performs Much Better than Without Union (Sparse Coding). 59

3.4 Group-level VS Image-level Scheme for Top 10 Terms in All Stages.

Image-level Spatial BOW Performs Slightly Better than Group-level

Spatial BOW. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 60

3.5 BoW vs Sparse Coding for Top 10 Terms in All Stages. Spatial Sparse

Coding Performs Slightly Better than Spatial BOW (Image-Level). . . . . 62

3.6 Max Pooling . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 63

3.7 Average Pooling . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 63

3.8 Sqrt Pooling . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 64

3.9 Comparison with Previous Research. Here is a Sample Comparison

Between the Previous Results (Refer to Table 5 in Yuan et al. (2012))

and Our Results (Refer to Table 5) for Stage Range 11-12. . . . . . . . . . . . . 64

4.1 Details of BVLC Reference CaffeNet. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 77

4.2 Performance of Features Extracted from Pretrained Deep Models. . . . . . 83

4.3 Performance of Features Extracted from Fine-Tuned Deep Models. . . . . 86

4.4 Performance of Features Extracted from Deep Models with Net Surgery. 87

5.1 Unigram Features for Two Domains . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 94

5.2 Results for Different Dimensions of Features . . . . . . . . . . . . . . . . . . . . . . . . . 104

5.3 KMM Results For a Small Number of Samples in the Target Domain . . 106

viii

Page 10: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

LIST OF FIGURES

Figure Page

1.1 Sample Images from BDGP Database. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4

1.2 Sample Images from FlySISH Database. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4

1.3 Hierarchical Structure of Stage Ranges and Stages in BDGP. . . . . . . . . . . 5

1.4 BDGP Images with Keywords Annotation. . . . . . . . . . . . . . . . . . . . . . . . . . . . 5

1.5 One Example of Group Discussions at LinkedIn. Members in This

Group Interact with Each Other by Posting Questions about Interior

Design and Answering Other Members’ Question. . . . . . . . . . . . . . . . . . . . . 8

1.6 A Sample of Spam. S.S Posted a Status Update at LinkedIn. A.A

Replied with a Post Searching for Jobs, Which is Clearly a Spam in

This Case. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9

1.7 Multi-lingual Uniform Model. Messages are Translated into Other Lan-

guages at First. The Keywords for All the Languages are Then Ex-

tracted. Frequency-based Values are Assigned to Each Entry of the

Data Matrix. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10

2.1 Illustration of Algorithm 2. Y Is Initialized as Z. We First Sort the En-

tries of Y to Form a Non-Decreasing Sequence {Yi}. Then We Reduce

Y as Much as Possible Sequentially within the Threshold σ, Finally

We Put Y Back in the Original Order. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 27

2.2 Background Extraction Results of Different Algorithms. In Each Sub-

figure, the 15-th, 150-th and 200-th Flames after Adding Noise are

Shown in the First Row. The Low-Rank Recoveries by Certain Al-

gorithm are Shown in the Second Row, and the Sparse Components

Extracted by the Algorithm are Shown in the Last Row. . . . . . . . . . . . . . . 31

ix

Page 11: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Figure Page

2.3 Shadow Removing Results of Different Algorithms (Top: ALM, Mid-

dle: NSA, Bottom: Our Proposed Algorithm). In Each Subfigure, the

YaleB Images after Adding Noise are Shown in the Left Column. The

Low-Rank Recoveries of Each Algorithm are Shown in the Middle Col-

umn, and the Sparse Errors Corresponding to Illumination are Shown

in the Right Column. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 38

3.1 The Proposed Framework of Drosophila Gene Pattern Annotation.

Given an Image, the SIFT Detector is Utilized to Generate Descriptors

for Local Patches of This Image. After SIFT Detection, the BoW or

Sparse Coding is Applied to Transform the Descriptors into Represen-

ters. Then the Pooling Functions Map the Representers into Features.

We Then Use These Features to Perform Keywords Annotation via

Different Schemes. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41

3.2 The Training Process of the Image-Level Scheme Compared with Group-

Level Scheme. In the Training Process, One Group of Images is Con-

sidered as One Sample in the Group-Level Scheme, While Each Image

within a Group is Treated as a Sample in the Image-Level Scheme. . . . 47

3.3 The Testing Process of the Image-Level Scheme Compared with Group-

Level Scheme. In the Testing Stage, the Group-Level Model Will Pro-

vide Prediction for Each Group of Images, while the Image-Level Model

will Predict Keywords for Each Image within this Group. In Addition,

in the Image-Level Scheme, We Make a Union of All the Predicted

Keywords in This Group, Which Will Form the Final Prediction for

This Group of Images. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 48

x

Page 12: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Figure Page

3.4 Average VS Majority Vote for All Stage Ranges. The Y-Axis of the

Figure Indicates the AUC. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 56

3.5 Image-Level Scheme with Union VS Image-Level Scheme without Union.

The Y-Axis of the Figure Indicates the AUC. . . . . . . . . . . . . . . . . . . . . . . . . 58

3.6 Group-level Scheme VS Image-level Scheme with Union. The Y-Axis

of the Figure Indicates the AUC. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 61

4.1 Learning Multiple Levels of Representation from Image. Raw Image

Is from Yoshua Bengio. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 66

4.2 LeNet 5: A Layered Model Composed of Convolution and Subsampling

Operations Followed by a Holistic Representation and Ultimately a

Classifier for Handwritten Digits LeCun et al. (1998) . . . . . . . . . . . . . . . . . 68

4.3 AlexNet: A Layered Model Composed of Convolution, Subsampling,

and Further Operations Followed by a Holistic Representation and All-

In-All a Landmark classifier on ILSVRC12 Krizhevsky et al. (2012) . . . . 70

4.4 A Typical Three-Layer Neural Network. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 71

4.5 Caffe Community in Github . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 76

4.6 Filters for Hidden Layer . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 79

4.7 Input Image . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 80

4.8 Output of the First Convolution Layer . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 81

4.9 Output of the Second Convolution Layer . . . . . . . . . . . . . . . . . . . . . . . . . . . . 81

4.10 Output of the Fifth Pooling Layer . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 82

4.11 Output of Fully-Connected Layer . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 82

4.12 The Workflow of Fine-Tuning . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 85

4.13 Fine-Tuning Convergence . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 85

xi

Page 13: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Figure Page

4.14 Comparison between Performance of Training-From-Scratch and Fine-

Tuning for Stage Range Classification. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 88

5.1 Proposed Pipeline for Spanish Spam Detection. Machine Translator is

Employed to Translate the Labeled Content from English to Spanish.

We Adopt Transfer Learning to Build a Spam Detection Model. Model

Evaluation is Done on Human Written Spanish. . . . . . . . . . . . . . . . . . . . . . . 91

5.2 Illustration of Adaptive Learning. Model w0 Is Borrowed from the

Domain T , and the Model w for the Domain H Should Be Similar to

w0. ∆w is a Small Perturbation We Aim to Learn. . . . . . . . . . . . . . . . . . . 99

5.3 Experiment Results for Varied Dimensions of Features: 600, 800, 1000

and 2000. In All These Settings, Our Proposed Algorithm Achieves

the Best Performance Compared with Other Algorithms. . . . . . . . . . . . . . 105

xii

Page 14: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Chapter 1

INTRODUCTION

1.1 Transfer Learning

We consider the scenarios where we have a few or no labeled data from the domain

we are interested in (target domain) but a large amount of labeled data available from

one or multiple related domains (source domains) with different-but-related data dis-

tributions. The difference of data distributions may lie in both marginal distribution

and conditional distribution. Under such scenarios, learning a single or multiple hy-

potheses on source domains using traditional machine learning methodologies and

applying them to target domain may lead to poor prediction performance. This is

because traditional machine learning algorithms assume that both the source and

target domain data are drawn independently and identically from the same distribu-

tion. One effective approach under such situations is Transfer Learning (or Do-

main Adaptation), which enables transfer of knowledge between source and target

domains with dissimilar distributions Pan and Yang (2009). It has been applied suc-

cessfully in various applications including text classification (parts of speech tagging,

webpage tagging, etc.) Daume III (2009), video concept detection across different

TV channels Duan et al. (2009a), and sentiment analysis (identifying positive and

negative reviews across domains) Blitzer et al. (2007).

Based on the way of migrating knowledge, transfer learning algorithms can be

classified into the following three categories Aytar (2014):

• Instance-based methods. This category of transfer learning algorithms aims

to re-weigh the instances in the source domain, in order to match the distribu-

1

Page 15: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

tion of target domain. Therefore, the classifier built from re-weighted source

domain can be applied to target domain Huang et al. (2006); Jiang and Zhai

(2007); Duan et al. (2009b); Daume III (2009).

• Feature-based methods. This category of transfer learning algorithms maps

the feature space of source domain and target domain into a common subspace,

thus the classifier can be built and applied on the same subspace. This kind of

algorithms work even when the source domain and target domain have different

feature dimensions Fink (2005); Yao et al. (2011); Kulis et al. (2011); Hoffman

et al. (2013).

• Model-based methods. This category of transfer learning algorithms makes

an assumption that the target classifier should be similar to the source classifier.

Thus the model is built via penalizing the difference between source model and

target model Li (2007); Donahue et al. (2013a); Fei-Fei et al. (2006); Gao et al.

(2012).

I present a two-stage transfer learning framework in this dissertation, which bridges

the gap induced by different distributions between source and target domains and

combine both instance-based method as well as model-based method. Specifically,

the distribution of source domain is mapped towards the target domain in the first

stage, and an adaptive learning model is proposed in the second stage to incorporate

label information from the target domain if it is available. Then I extend the proposed

transfer learning pipeline to the spam detection problem at LinkedIn’s website.

Deep learning has achieved a lot of success in image classification recently. Deep

learning model is essentially a cascade of many layers containing nonlinear processing

units for feature extraction and transformation Deep Learning (2015). Each succes-

sive layer uses the output from the previous layer as input. The algorithms may be

2

Page 16: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

supervised or unsupervised and applications include pattern analysis (unsupervised)

and classification (supervised). I propose to apply deep learning algorithms on our

Drosophila embryo images in this dissertation. Due to the fact that a large amount of

data are required for training deep models, the parameter transfer is utilized for ini-

tializing the deep model. Specifically, I propose to transfer the pretrained deep model

trained from ImageNet dataset and fine-tune it for our application. The performance

and time complexity will be investigated between training a deep model from scratch

and fine-tuning pretrained models.

1.2 Background

1.2.1 BioImaging Annotation

The development of a multi-cellular organism begins from a single fertilized egg

and proceeds as cell division and differentiation, controlled by the spatio-temporal

expression of a multitude of genes over time Lecuyer et al. (2007). Study of the

gene regulatory network is a crucial step towards unveiling the mechanism governing

cell-fate differentiation and embryonic development Fowlkes et al. (2008)). For many

years, Drosophila melanogaster has been established as the canonical model organ-

ism for the study of the fundamental principles of animal development Sean Carroll

(2005); Levine and Davidson (2005); Matthews et al. (2005)). To facilitate a system-

atic understanding of transcriptional regulation during Drosophila embryogenesis, the

Berkeley Drosophila Genome Project (BDGP) Tomancak et al. (2002, 2007)) and Fly-

FISH Lecuyer et al. (2007)) both produced a comprehensive atlas of gene expression

patterns in the form of two-dimensional (2D) digital images using high-throughput

RNA in situ hybridization Grumbling et al. (2006)). Sample images are shown in

Figure 1.1 and Figure 1.2.

3

Page 17: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

stage 9 stage 10 stage 11

stage 12 stage 13 stage 14

Figure 1.1: Sample Images from BDGP Database.

Stage 4 Stage 5 Stage 6

Stage 7 Stage 8 Stage 9

Figure 1.2: Sample Images from FlySISH Database.

Each of the images in BDGP and FlyFISH is labeled with stage range, but only a

small portion of them are annotated with stage-specific labels. The underlying reason

behind the label pattern is that the difference between stage ranges is more obvious

than the difference between nearby stages. For example, stage 5 may be similar to

stage 4, but it should be different from stage 1. Therefore, it is much easier to label

an image with stage range (from stage range 1 to stage range 6) than to label an

image with fine-grained stage (from stage 1 to stage 17) for a domain expert. The

hierarchical structure of stage ranges and stages for BDGP is illustrated in Figure 1.3.

4

Page 18: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Overlapping patterns from the same developmental stage imply strong co-regulations

in the development. Automated and efficient tools for stage annotation on images are

a prerequisite for efficient generation of biological insights into gene functions, interac-

tions and networks, because these images are being produced at a higher throughput

than ever before Yuan et al. (2014)). In this dissertation, I present an automated

computational system using novel machine learning methods to annotate the devel-

opmental stages for Drosophila embryo images. Furthermore, I propose to utilize

the hierarchical structure between stage range and stage via transferring knowledge

learnt from range models to stage models.

The BDGP images capture the spatial pattern of individual genes at a particu-

lar developmental stage as aforementioned. These images are also annotated with

anatomical and developmental ontology terms using a controlled vocabulary to fa-

cilitate text-based search of gene expression patterns Tomancak et al. (2002), see

Figure 1.4.

... .........

2 3 4 5Stage Range

Stage...

1

1 3 4 6 7 8 9 10 11 12

6

13 17...

Figure 1.3: Hierarchical Structure of Stage Ranges and Stages in BDGP.

Currently, the annotation of keywords are performed manually by human cura-

tors. With the rapid accumulation of available images generated by high throughput

technologies, it is imperative to design effective computational methods to automate

the annotation Ji et al. (2009a); Yuan et al. (2012)). The multi-image multi-label

nature of this problem poses significant challenges to traditional image annotation

methodologies. This is partially due to the fact that the assignment of a particular

term to a group does not imply that all images in this group are associated with this

5

Page 19: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Figure 1.4: BDGP Images with Keywords Annotation.

term. Hence, special formalism needs to be designed to retain the group membership

information. In addition, the shape and appearance of the same body part may vary

from image to image due to the effects of stochastic processes during embryogenesis

and the distortions introduced by the current image acquisition techniques. Hence,

invariance to local distortions is a crucial requirement for an accurate annotation

system. Several prior studies on the automatic annotation of Drosophila gene ex-

pression images have been reported. Zhou and Peng (2007) constructed their system

based on the assumption that each image in the group is annotated by all the terms

assigned to that group; Ji et al. (2009b) considered a learning framework that in-

corporates both the image group information and the term-term interactions; Yuan

et al. (2012) designed a bag-of-words based approach in which the spatial information

of the images is utilized. To date, no systematic comparison between the image-level

and the group-level methods has been reported. In addition, it is not clear which

scheme performs the best in combining and pooling images in the same group. In

this dissertation, I propose an end-to-end pipeline to compare the image-level and

group-level schemes for keywords annotation. Moreover, different pooling methods

are investigated for image feature extraction Sun et al. (2013a).

6

Page 20: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

In both BDGP and FlyFISH databases, samples are represented in the form of

digital images and features are to be extracted from the images. Due to the short life-

cycle of Drosophila, the embryo images may possess a low-rank structure capturing

the overall contour of embryos and a sparse component capturing the subtle changes

at different developmental stages. How to accurately extract the low-rank and sparse

components is a major challenge to process the images. Robust Principal Component

Analysis (RPCA) is a general framework to extract such structures Sun et al. (2013b).

I present a novel non-convex formulation for the RPCA problem using the capped

trace norm and the capped ℓ1-norm. In addition, I present a fast alternating algorithm

to solve the non-convex optimization. Compared to existing convex formulations, my

proposed approach gives proper interpretation and the proposed algorithm achieves

better accuracy for the RPCA problem which can be verified by empirical results.

Based on the low-rank structure produced by RPCA, I propose to apply the trans-

fer learning algorithm to investigate the stage annotation and keywords annotation

problem of BDGP.

1.2.2 Bilingual Spam Detection

Transfer learning exhibits its capability of solving the lack-of-label problem in

various scenarios. Besides the BioImaging application I introduced before, I also

investigate the efficacy of transfer learning in Bilingual spam detection problem in

this thesis.

LinkedIn is the world’s largest professional network with more than 332 million

members in over 200 countries and territories. LinkedIn provides a wide variety

of products and services that allow members to connect with one another in order

to be more productive and successful in their professional lives. LinkedIn Groups in

particular is a product which allows professionals in the same industry or with similar

7

Page 21: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

interests to gather, share content, find answers to questions, and establish themselves

as industry experts.

LinkedIn members are able to freely join public groups, or may be invited to

private groups by a group manager. So far, there are more than 2 million groups

at LinkedIn, and 154, 000 pieces of content are being posted every day. Members

who use Groups are 5 times more likely to get profile views from people outside their

network. Figure 1.5 shows a LinkedIn group for professional interior designers, and

a snapshot of discussions amongst designers sharing expertise and knowledge about

their industry. The best way to participate in LinkedIn groups in general is to focus

on intelligent, meaningful posts that add value to a discussion and that will benefit

other members.

There are rules for participating in LinkedIn Groups discussions. First, different

groups have different norms and expectations around discussions, established by the

participants in the group and the group manager(s). For example, different groups

have different tolerance for promotional content in group discussions. Some groups

provide a Jobs area for posting job openings, talking about the current job market,

asking job related questions, or starting career discussions. Other groups allow nei-

ther promotions nor job postings. In other words, users posting to a group should

understand the community norms and expectations for that group, and use good

judgement when making a post. Content posted to a group that does not adhere to

the group’s norms and expectations is considered spam.

Spam content posted in group discussion at LinkedIn will jeopardize other user’s

return site visit. Figure 1.6 is one of the spam examples in group discussion 1 , where

A.A puts a job post in reply to S.S’s profile update. A.A’s response would probably

1For the purpose of privacy protection, we anonymize the name and profile pictures in all exam-ples.

8

Page 22: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Figure 1.5: One Example of Group Discussions at LinkedIn. Members in This Group

Interact with Each Other by Posting Questions about Interior Design and Answering

Other Members’ Question.

annoy S.S since her reply is not related to S.S’s update. Simply blocking A.A’s account

is not a smart solution for LinkedIn, as she may be a legitimate user and may also

put informative posts somewhere else on the site. A better solution is to examine

each piece of content and filter those deemed to be spammy. With the development

of globalization in social network, LinkedIn is available in more than 20 languages.

Therefore such spam content may be written in different languages, such as English,

Spanish, Chinese, etc. Efficient tools for spam detection in multiple languages are

highly desired at LinkedIn. The multilingual text categorization problem has gained

9

Page 23: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Figure 1.6: A Sample of Spam. S.S Posted a Status Update at LinkedIn. A.A Replied

with a Post Searching for Jobs, Which is Clearly a Spam in This Case.

its attention recently Amini and Goutte (2010); Ling et al. (2008); Pan et al. (2011);

Vinokourov et al. (2002); Wang and Domeniconi (2009).

One may solve this multi-lingual spam detection problem by enriching data repre-

sentation with machine translated features in different languages Banea et al. (2010);

Tang et al. (2012)), as illustrated in Figure 1.7. If we are dealing with multilingual

spam detection problem in English, Spanish and French, we can translate any con-

tent into the other two languages. In this way, we will have bag-of-words features

concatenated from three languages, and each piece of content needs to be translated

into multiple languages in order to apply the uniform model. However, the machine

translation to multiple languages leads to higher latency and consumes more budget.

Therefore, the uniform model is not the best choice for an online product.

The motivation of this work is to develop a computational method to automatically

maintain the order of group discussion, i.e., to detect and filter the unwanted posts.

For each group, spam detection is conducted based on its specific rules. For example,

in the groups which do not allow job posting, any posts related to job annoucement

10

Page 24: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Figure 1.7: Multi-lingual Uniform Model. Messages are Translated into Other Lan-

guages at First. The Keywords for All the Languages are Then Extracted. Frequency-

based Values are Assigned to Each Entry of the Data Matrix.

are considered as spam and appropriate actions need to be taken. Users who post

spam content under group discussion are not necessarily spamers, instead they may

be legitimate users but unfamiliar with the rules of the group. For this reason, we

choose to build a content classifier rather than to block the users’ accounts.

Users can post their updates in up to 20 different languages. Thus, to maintain the

order of group discussion, spam detection is required to work in different languages.

From the perspective of production, we build one content classifier for each language

and keep a classifier pool in our system. In this dissertation, I design a generic pipeline

for spam detection in different languages. Starting with Spanish, I develop a two-

stage transfer learning approach to solve the cross-language spam detection problem.

I test and validate the proposed algorithm on real data from LinkedIn.

The rest of this dissertation is organized as follows. In Chapter 2, I introduce a

Robust Principle Component Analysis algorithm to process the Drosophila embryo

11

Page 25: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

images before classification. Specifically, I preprocess our images via filtering the

sparse component. In Chapter 3, I proposed an end-to-end pipeline to tackle the

multi-instance multi-label keywords annotation problem. In Chapter 4, I show that

by integrating deep learning and transfer learning further improves the classification

performance for our stage annotation task in terms of both accuracy and efficiency.

In Chapter 5, I present a two-stage transfer learning algorithm to bridge the gap

between two domains and apply it to bilingual spam detection problem. Last but not

least, Chapter 6 concludes the entire thesis and points out some promising directions

for future work.

12

Page 26: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Chapter 2

ROBUST PRINCIPLE COMPONENT ANALYSIS FOR IMAGE PROCESSING

2.1 Introduction

In many applications, we encounter very high-dimensional data such as images,

texts, and genomic data. Analysis of such data is challenging due to the curse of

dimensionality. One promising approach is to exploit the special structures of the

data and it has recently achieved great success in many applications Benoıt et al.

(2011); Chen et al. (2012); Mairal et al. (2011); Singer and Cucuringu (2010); Ye and

Liu (2012). Two particularly interesting structures are the low-rank structure Xiang

et al. (2012b) and the sparse structure. For example, the images of the same scene may

be taken from different illusions, thus the shadows represent the sparse component

and the scene relates to the low-rank part Aybat et al. (2011). For a collection

of text documents, the low-rank component could capture common words from all

the documents while the sparse component may capture the few key words that

best distinguish each document from others Candes et al. (2009). Robust Principal

Component Analysis (RPCA) Candes et al. (2009), or Stable Principal Component

Pursuit (SPCP) Aybat et al. (2011) is an efficient tool for such analysis and has

drawn increasing attentions in many areas Cai et al. (2010); Candes et al. (2009);

Peng et al. (2010); Recht et al. (2010); Tao and Yuan (2011); Toh and Yun (2010);

Wright et al. (2009).

The most general form of the RPCA problem can be formulated as follows:

minimizeX,Y

rank(X) + λ∥Y ∥0,

subject to A = X + Y,

(2.1)

13

Page 27: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

where we assume that the data matrix A is the summation of a low-rank matrixX and

a sparse component Y . We minimize the rank of X as well as the number of non-zero

entries in Y . Due to the discrete nature of the rank function, such a problem has been

proven to be NP-hard. Therefore computing a global optimal solution of eq. (2.1) is

a challenge.

Recently, one proper relaxation with theoretical guarantees has been proposed

in Candes et al. (2009). In particular, the authors approximate the original problem

by minimizing the weighted sum of trace norm ofX and ℓ1-norm of Y with an equality

constraint:

minimizeX,Y

∥X∥∗ + λ∥Y ∥1,

subject to A = X + Y,

(2.2)

where the trace norm ∥X∥∗ is defined as the sum of all singular values of X, and

∥Y ∥1 =∑

ij |Yij| denotes the sum of absolute values of all entries in Y . Notice

that (2.2) is a convex optimization problem, therefore a global optimal can be com-

puted. Moreover, the authors of Candes et al. (2009) also provide an efficient algo-

rithm for (2.2). It is well known that the trace norm and the ℓ1-norm are capable of

inducing low-rank and sparse structures Tao and Yuan (2011), achieving our desired

goal.

Furthermore, it has been shown in Candes et al. (2009) that with the balance

parameter λ equal to 1√max(m,n)

, model (2.2) will provide the correct answer.

Problem (2.2) can recover a low-rank matrix only with sparse corruptions. In

practice, as discussed in Wright et al. (2009), it is necessary to consider the observed

data matrix under more realistic conditions. Particularly, the given data may not

only be corrupted by impulse noise which is sparse and large, but they may also

be corrupted by Gaussian noise which is dense and small. To deal with this more

realistic issue, a modified model with tolerance of both Gaussian noise and impulse

14

Page 28: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

noise has been built in Tao and Yuan (2011). It replaced the equality constraint with

an inequality constraint:

minimizeX,Y

∥X∥∗ + λ∥Y ∥1,

subject to ∥A−X − Y ∥F ≤ σ,

(2.3)

where σ denotes the level of the Gaussian noise. The authors propose an augmented

lagrangian approaches to efficiently solve problem (2.3). Based on the same model,

an Augmented Lagrange Method of Multipliers (ADMM) is proposed in Lin et al.

(2010) to solve the optimization problem, which iteratively solves X and Y using

soft-thresholding. In Aybat et al. (2011), the authors developed a non-smooth aug-

mented lagrange method to efficiently solve problem (2.3). There are also many other

approaches which aim to solve a similar problem. For example, in Chandrasekaran

et al. (2011), the authors develop a notion of rank-sparsity incoherence and uses it to

characterize both of the fundamental identifiability and the sufficient conditions for

exact recovery. In Xu et al. (2010), the authors presented an approach to recover the

correct column space of the uncorrupted matrix, rather than the matrix itself.

To our best knowledge, most of the recent research focuses on solving RPCA

via solving a convex problem with certain constraints Cai et al. (2010); Peng et al.

(2010); Recht et al. (2010); Toh and Yun (2010). By contrast, this chapter proceeds

this interesting problem from a different point of view: non-convex optimization using

mixture of capped trace norm and capped ℓ1-norm. Non-convex formulation is shown

to deliver better performance then convex formulation for some applications Xiang

et al. (2012a). I propose two algorithms to solve the problem. First, we apply the

Difference of Convex functions (DC programming) framework to iteratively solve the

non-convex formulation, and apply ADMM to solve the sub-problem. In our second

approach, instead of using ADMM to solve the sub-problem, we present a greedy

algorithm to solve the sub-problem, based on which we propose a fast alternating

15

Page 29: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

optimization algorithm. Both low-rank part X and sparse component Y can be

recovered by our algorithms with higher accuracy than most of the exsited work.

The RPCA formulation has applications in many areas Candes et al. (2009). In

this chapter, we evaluate the proposed algorithms with synthetic data and two real-

world applications: background detection in surveillance video and shadow removing

from illuminated portraits. In both synthetic experiments and real-world applications,

our proposed algorithms achieve better recovery of X and Y for the RPCA problem

compared to existing convex formulations. In particular, our proposed algorithms can

capture the sparse locations with higher accuracy than existing methods.

2.2 Problem Formulation

In this section, we formulate the RPCA problem as a non-convex minimization

problem via capped norms.

Given a data matrix A, the goal of RPCA is to extract a low rank X and a

sparse component Y from A. Following (2.1), we consider a non-convex formulation

of RPCA with an inequality constraint:

minimizeX,Y

rank(X) + λ∥Y ∥0,

subject to ∥A−X − Y ∥2F ≤ σ2,

(2.4)

where A ∈ Rm×n is the observed matrix, σ2 is the level of Gaussian noise and λ > 0

is a trade-off parameter between the low-rank part and the sparse component. Here,

the ∥ · ∥0 norm is the number of non-zero entries of a matrix. In Aybat et al. (2011);

Tao and Yuan (2011), the trace norm and the ℓ1-norm were used to approximate the

rank function and ℓ0-norm to convert the non-convex problem into a convex one. It is

noteworthy that the convex relaxation may not be the best approximation of (2.4) in

real-world applications. The main motivation of the re-formulation to be presented

in the next part is to reduce the approximation error using non-convex formulations.

16

Page 30: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

2.2.1 Capped Norms

We first introduce the capped norms for matrices and vectors, which are the surro-

gates of the rank function and the ℓ0-norm. Let p = min(m,n). We can approximate

the rank function and ℓ0-norm by:

rank(X) ≈p∑

i=1

min(1,σi(X)

θ1)

=1

θ1

[∥X∥∗ −

p∑i=1

max(σi(X)− θ1, 0)

],

∥Y ∥0 ≈∑ij

min(1,|Yij|θ2

)

=1

θ2

[∥Y ∥1 −

∑i,j

max(|Yij| − θ2, 0)

],

for some small parameters θ1, θ2 > 0. We can observe that if all the singular values

of X are greater than θ1 and all the absolute values of elements in Y are greater than

θ2, the approximation will become equality.

The smaller θ1 and θ2 are, the more accurate the capped norm approximation

would be. we can control the recovery precision via making use of θ1 and θ2. By

carefully choosing θ1 and θ2, we can recovery X and Y more accurately than the

trace norm and ℓ1-norm approximation.

2.2.2 Proposed Non-Convex Formulation

With the aforementioned capped norms, we propose to solve the following non-

convex RPCA formulation:

minimizeX,Y

1

θ1∥X∥∗ +

λ

θ2∥Y ∥1 −

[1

θ1P1(X) +

λ

θ2P2(Y )

],

subject to ∥A−X − Y ∥2F ≤ σ2,

(2.5)

where P1(X) =∑p

i=1max(σi(X)− θ1, 0), and P2(Y ) =∑i,j

max(|Yij| − θ2, 0).

17

Page 31: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Clearly, the new objective function in problem (2.5) is not convex due to the last

two terms being concave functions. However, the intrinsic structure of the formulation

naturally leads to use of Difference of Convex (DC) Programming Tao and An (1997).

DC programming treats non-convex function as the difference of two convex functions,

then iterately solve it on the basis of the combination of the first convex part and

the linear approximation of the second convex part. Obviously, the trace norm and

the ℓ1-norm of are convex, and the summation of maximum is also convex. Thus

problem (2.5) exhibits the DC structure. The details of DC programming to solve

the RPCA problem are presented in the next section.

2.3 A DC Programming-Based Algorithm

In this section, we detail the DC programming framework for solving problem (2.5).

In each iteration of the framework, the first-order appoximation is used to substitute

the non-convex part. To generate the first-order approximation of P1(X) and P2(Y )

in our formulation, we need to compute the subdifferential of a capped trace norm,

as summarized in the following lemma:

Lemma 1. The subdifferential of

P1(X) =

p∑i=1

max(σi(X)− θ1, 0)

is given by:

∂P1(X) ={UDiag(z)V T : z ∈ Z∗} , (2.6)

where U and V are the left and right singular vectors of X, respectively, and

Z∗ =

z ∈ Rp

∣∣∣∣∣∣∣∣∣∣∣zi

= 1 if σi(X) > θ1,

= 0 if σi(X) < θ1,

∈ [0, 1] otherwise.

and p is the rank of X.

18

Page 32: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Proof. First, we can denote X = UΣXVT as the SVD of matrix X where U and V are

unitary matrices. Then we define auxiliary matrices B = UΣBVT and C = UΣθV

T ,

where ΣB = Diag(b), bi ∈ {0, 1} and Σθ = Diag(θ). For simplicity, we denote

σX = (σ1(X), σ2(X), . . . , σp(X))T ,

σB = (σ1(B), σ2(B), . . . , σp(B))T ,

and

σθ = (θ1, θ1, . . . , θ1)T .

Using the notations above, P1(X) can be written as

P1(X) =

p∑i=1

max(σi(X)− θ1, 0)

= maxσB∈E

< σB, σX − σθ >,

where E = {s ∈ Rp : si ∈ {0, 1}}. We can see that the maximum can be achieved if

and only if:

σi(B) ∈

1 if σi(X)− θ1 > 0,

0 if σi(X)− θ1 < 0,

{0, 1} otherwise.

The subdifferential should be the convex hull of B Nesterov (1998):

∂P1(X) = conv{B : B = UΣBVT}

= UΣ∗BV

T ,

where Σ∗B = Diag(σ∗

B) and

σ∗B(i) ∈

1 if σi(X)− θ1 > 0,

0 if σi(X)− θ1 < 0,

[0, 1] otherwise.

This completes the proof of the lemma.

19

Page 33: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

In addition, it can be easily shown that:

∂P2(Y ) =

V ∈ Rm×n

∣∣∣∣∣∣∣∣∣∣∣∣∣∣∣∣∣Vij ∈

{1} if Yij > θ2,

[0, 1] if Yij = θ2,

{0} if |Yij| < θ2,

[−1, 0] if Yij = −θ2,

{−1} if Yij < −θ2.

(2.7)

By denoting U = 1θ1∂P1(X) and V = 1

θ2∂P2(Y ), the formulation (2.5) can be

rewritten as:

minimizeX,Y

1

θ1∥X∥∗ +

λ

θ2∥Y ∥1 − ⟨U,X⟩ − ⟨λV, Y ⟩,

subject to ∥A−X − Y ∥2F ≤ σ2,

(2.8)

where ⟨U,X⟩ =∑m

i=1

∑nj=1 UijXij. Thus, we solve the original non-convex problem by

solving a series of convex problems. In each iteration, we approximate problem (2.5)

by the sub-problem (2.8) at the current Xk and Y k. The key sub-problem is to solve

the convex problem (2.8).

2.3.1 Solving the Sub-Problem

Here, we apply the Augmented Lagrange Method of Multipliers (ADMM) Boyd

et al. (2011) to solve the sub-problem (2.8). ADMM has been applied successfully to

solve many sparse learning problems. We introduce an auxiliary variable S = X and

rewrite the problem (2.8) as:

minimizeX,Y

1

θ1∥S∥∗ +

λ

θ2∥Y ∥1 − ⟨U,X⟩ − ⟨λV, Y ⟩,

subject to ∥A−X − Y ∥2F ≤ σ2,

X − S = 0.

(2.9)

20

Page 34: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

The augmented lagrangian function of (2.9) is:

Lρ(S,X, Y,Λ) =1

θ1∥S∥∗ +

λ

θ2∥Y ∥1 − ⟨U,X⟩ − ⟨λV, Y ⟩

+ ⟨Λ, X − S⟩+ ρ

2∥X − S∥2F ,

subject to ∥A−X − Y ∥2F ≤ σ2,

where Λ is the lagrangian multiplier and ρ is the step size of dual update.

The general approach of ADMM consists of the following iterations:

Sk+1 = argminS

Lρ(S,Xk, Y k,Λk),

{Xk+1, Y k+1} = argminX,Y

Lρ(Sk+1, X, Y,Λk),

Λk+1 = Λk + ρ(Xk+1 − Sk+1).

(2.10)

Next, we present the details for updating each variable in (2.10).

Updating S

The update of S involves the following problem:

Sk+1 = argminS

1

2

∥∥∥∥S −Xk − Λk

ρ

∥∥∥∥2

F

+1

θ1ρ∥S∥∗, (2.11)

which is the proximal operator of the trace norm. It has an analytical solution as

summarized in the following lemma Cai et al. (2008):

Lemma 2. The proximal operator associated with the trace norm, i.e., the minimizer

of the following problem:

minimizeX∈Rm×n

1

2∥X − A∥2F + λ∥X∥∗,

is given by:

proxλ∥·∥∗(X) = USV T ,

S = Diag(max(σ − λ, 0)),

21

Page 35: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

where σ = (σ1, . . . , σp) are the singular values of A and A = UDiag(σ)V T is the SVD

of A.

From Lemma 2, if we denote the SVD of matrix Xk + Λk

ρas UsΣsV

Ts , where

Σs = Diag(σ), and pi = max(σi − 1θ1ρ

, 0), then Sk+1 can be obtained via a closed-

form: Sk+1 = UsDiag(p)VTs .

Updating X and Y

The update of X and Y amounts to solve:

{Xk+1, Y k+1} =argminX,Y

λ

θ2∥Y ∥1 − ⟨Uk, X⟩ − ⟨λV k, Y ⟩

+ ⟨Λk, X − Sk+1⟩+ ρ

2∥X − Sk+1∥2F

s.t. ∥A−X − Y ∥2F ≤ σ2.

(2.12)

By introducing a lagrangian multiplier ν for the inequality constraint, the lagrangian

function is given by:

L =λ

θ2∥Y ∥1 − ⟨Uk, X⟩ − ⟨λV k, Y ⟩+ ⟨Λk, X − Sk+1⟩

2∥X − Sk+1∥2F + ν(∥A−X − Y ∥2F − σ2).

Taking partial derivatives of L with respect to X and Y results in:∂L∂X

= −Uk + Λk + ρ(X − Sk+1) + 2ν(X + Y − A),

∂L∂Y

= λθ2DY − λV k + 2ν(X + Y − A).

where DY is the subdifferential of ∥Y ∥1. Setting both partial derivatives to 0, we

have

ρY +λ(2ν + ρ)

2νθ2DY + C = 0. (2.13)

where C is a constant.

22

Page 36: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

It is easy to verify that (2.13) is precisely the first-order optimality condition

for the problem (2.14), which has a closed form solution summarized below Donoho

(1995).

minimizeY

1

2∥Y +

C

ρ||2F +

λ(2ν + ρ)

2νθ2ρ∥Y ∥1. (2.14)

Lemma 3. The proximal operator associated with the ℓ1-norm, i.e., the minimizer

of the following problem:

minimizeX∈Rm×n

1

2∥X − V ∥2F + λ∥X∥1,

is given by:

proxλ∥·∥1(Vij) = sign(Vij)⊗max(|Vij| − λ, 0),

where the ℓ1-norm of a matrix if given by the summation of the absolute value of all

elements.

From Lemma 3, the solution of problem (2.13) can be obtained by Y k+1 =

sign(Cρ)⊗max{C

ρ− λ(2ν+ρ)

2νθ2ρ, 0} based on ν. And from the relation between X and Y ,

we can get Xk+1. Since there is no direct way to calculate ν, we use the binary search

to find the proper ν.

Thus, we obtain the solution of sub-problem (2.9), which is one iteration of DC

framework. In the DC framework, a series of sub-problem (2.9) are solved iteratively.

The details are summarized in Algorithm 1.

23

Page 37: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Algorithm 1 RobustPCA via DC Programming

Require: X0, Y0, θ1, θ2, σ

Ensure: an optimal solution X and Y

1: while not converge do

2: Calculate P1(X) =∑p

i=1max(σi(X)− θ1, 0) and P2(Y ) =∑i,j

max(|Yij| − θ2, 0)

3: Compute ∂P1 and ∂P2 according to (2.6) and (2.7)

4: Apply ADMM to solve problem (2.9):

5: for j = 1 to MaxIter do

6: update S using (2.11)

7: update X and Y using (2.14)

8: update Λ using (2.10)

9: end for

10: end while

2.4 A Fast Alternating Algorithm

The DC programming is known to have a slow convergence rate. In this section,

we propose an algorithm based on alternating optimization which is practically much

faster than DC programming. Notice that problem (2.5) is equivalent to:

minimizeX,Y

1

θ1

∑i

min{σi(X), θ1}+1

θ2

∑i,j

min{|Yij|, θ2},

subject to ∥A−X − Y ∥2F ≤ σ2.

(2.15)

In the proposed algorithm, we iteratively fix one variable and compute the other one.

24

Page 38: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

2.4.1 Computing the Optimal Y

When X is fixed, computing the optimal Y amounts to solve the following sub-

problem after dropping constants and changing variables:

minimizeY

1

θ2

∑i,j

min{|Yij|, θ2},

subject to ∥Y − Z∥2F ≤ σ2,

(2.16)

where Z = A − X. It is easy to see that we only need to consider the situation

that Z ≥ 0. Moreover, notice that for any feasible solution Y , if Yi > Zi for some i,

by setting Yi = Zi we can always reach another feasible solution with equal or less

objective value. Therefore we can assume that the optimal Y ∗ satisfies Y ≤ Z. In

addition, we may further assume Y and Z are both represented in a vector which is

formed by stacking all columns of the matrix.

It is not hard to see that the objective function of problem (2.16) is non-convex

and it is usually very difficult to find a globally optimal solution. We first present our

method in Algorithm 2. Then we will show that, given any feasible solution of (2.17),

we can always improve it to get a better local solution through our Algorithm 2.

An intuitive example is presented in Figure 2.1. We first initialize Y by Z and

sort Y such that the elements of Y form a non-decreasing sequence. The key idea

behind our algorithm is as follows: among all the solutions of (2.16), there must be

one Y ∗ such that the elements of Y ∗ preserves the order of Z.

Lemma 4. Let Y ∗ be one feasible solution of (2.16) such that there exist indices i

and j satisfying Zi < Zj and Y ∗i > Y ∗

j . There always exists another local solution Y

such that Yi ≤ Yj and Yk = Y ∗k for all k = i, j.

Proof. If there exist indices i and j such that Zi < Zj and Y ∗i > Y ∗

j . Let Yi = Y ∗j ,

25

Page 39: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Algorithm 2 An Approximate Algorithm of (2.16)

Require: Y , Z, θ

Ensure: an optimal solution Y

1: if ∥Z∥2F ≤ σ2 then

2: return 0

3: end if

4: Initialize Y by Z.

5: Sort Y in increasing order

6: i = 1

7: while σ2 > 0 do

8: if σ > Yi then

9: σ =√

σ2 − Y 2i , Yi = 0.

10: else

11: σ = 0, Yi = Yi − σ.

12: end if

13: i = i+ 1

14: end while

15: return Y

Yj = Y ∗i and Yk = Y ∗

k for all k = i, j, then we have:

∥Y ∗ − Z∥2F − ∥Y − Z∥2F

=∥Y ∗ − Z∥2F − (∑k =i,j

(Yk − Zk)2 + (Yi − Zi)

2 + (Yj − Zj)2)

=∥Y ∗ − Z∥2F − (∑k =i,j

(Y ∗k − Zk)

2 + (Y ∗j − Zi)

2 + (Y ∗i − Zj)

2)

=(Y ∗i − Zi)

2 + (Y ∗j − Zj)

2 − (Y ∗j − Zi)

2 − (Y ∗i − Zj)

2

=2Y ∗j Zi + 2Y ∗

i Zj − 2Y ∗j Zj − 2Y ∗

i Zi

=2(Y ∗i − Y ∗

j )(Zj − Zi) ≥ 0.

26

Page 40: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

The above result essentially shows that exchanging Y ∗i and Y ∗

j will not violate the con-

straint and clearly the objective remains unchanged. Therefore we find an alternative

feasible solution that preserves the order of elements in Z.

Lemma 5. Let Y ∗ be one feasible solution of (2.16) such that there exists index i

satisfying 0 < Y ∗i < Y ∗

i+1 < Zi+1. There always exists another feasible solution Y

such that Yk = Y ∗k for all k = i, i+ 1 and either Yi = 0 or Yi+1 = Zi+1 holds.

Lemma 4 essentially states that the optimal solution preserves the order in Z and

Lemma 5 shows that we can obtain an solution such that there exists an index i such

that Y ∗j = 0 for all j < i and Y ∗

j = Zj for all j > i. Then it is straight-forward to

show that our algorithm provides a better local solution.

2.4.2 Computing the Optimal X

When Y is fixed, the optimal X can be obtained via solving the following equiv-

alent problem:

minimizeX

1

θ1

∑i

min{σi(X), θ1},

subject to ∥X − Z∥2F ≤ σ2.

(2.17)

Let Z = UΣV T be the SVD of Z. For any feasible solution X, let UΣV T be its

SVD. We can observe that:

∥X − Z∥2F

=∥X∥2F + ∥Z∥2F − 2Tr(XTZ)

=∥UΣV T∥2F + ∥Z∥2F − 2Tr(XTZ)

≥∥UΣV T∥2F + ∥Z∥2F − 2∑i

σi(X)σi(Z)

=∥UΣV T∥2F + ∥Z∥2F − 2Tr((UΣV T )TUΣV T )

=∥UΣV T − Z∥2F ,

27

Page 41: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Sorted YInitial Y

Reduced YResulted Y

R1R2

R3

2222 321 RRR

Figure 2.1: Illustration of Algorithm 2. Y Is Initialized as Z. We First Sort the

Entries of Y to Form a Non-Decreasing Sequence {Yi}. Then We Reduce Y as Much

as Possible Sequentially within the Threshold σ, Finally We Put Y Back in the

Original Order.

where the second equation follows from the fact that the Frobenius norm is unitary-

invariant and we use the Von Neumann’s trace inequality Mirsky (1975) to obtain the

first inequality. The conclusion above essentially shows that the optimal X shares

the same left and right singular vectors with Z. Using the unitary-invariant property

of the Frobenius norm we can conclude that (2.17) is equivalent to:

minimizeσ(X)

1

θ1

∑i

min{σi(X), θ1},

subject to∑i

(σi(X)− σi(Z))2 ≤ σ2,

(2.18)

which is exactly in the same form as (2.16) and therefore it can also be computed

through Algorithm 2.

28

Page 42: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

2.5 Experimental Results

In this section, we compare our proposed algorithms with ALM Lin et al. (2010)

and NSA Aybat et al. (2011) on synthetic data and real-world data sets. The ALM

algorithm formulates RPCA as a convex problem with an equality constraint and ap-

plied augmented lagrange multiplier method to solve it. The NSA algorithm formu-

lated the RPCA as a convex problem with an inequality constraint, and investigated

a non-smooth augmented lagrange algorithm to solve it.

Our experiments are all executed on a PC with Intel Core2 Quad Q8400 2.66 G

CPU and 8G RAM. The ALM code was downloaded from the Perception and Decision

Laboratory, University of Illinois (http://perception.csl.illinois.edu/), and

the NSA code was kindly provided by the author. We implement both of the DC

framework and fast alternating algorithm in MATLAB.

2.5.1 Synthetic Data

In this experiment, we compare different algorithms on synthetic data. We gener-

ate the rank-r matrix X0 as a product of LRT , where L and R are independent n× r

matrices whose elements are i.i.d. random variables sampled from standard Gaussian

distributions. We generate Y 0 as a sparse matrix whose support is chosen uniformly

at random, and whose non-zero entries are i.i.d. random variables sampled uniformly

in the interval [−100, 100]. We set the standard Gaussian noise level at ρ to simulate

ξ ∼ N (0, ρ2). The matrix A = X0 + Y 0 + ξ is the input to the algorithms.

In our empirical study, we set θ1 = θ2 = 0.01 and σ2 = ρ√n+

√8n as in Tao and

Yuan (2011) and utilize the NSA results to initialize X and Y in our algorithms. Note

that it is very common (in fact it is recommended) to use convex relaxation solutions

as initial solutions for non-convex formulations Fan et al. (2012); Zhang (2010).

29

Page 43: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

There are three key parameters in the process of generating A, namely, n, ρ

and cr: n is the dimension of A; cr represents the ratio between the rank and the

dimension of X0, i.e., r = n × cr; cp denotes the density of non-zero entries in Y 0,

i.e., ∥Y ∥0 = cp × n. To verify the efficacy and robustness of different approaches, we

proceed our experiments in three directions, by varying different parameters:

• We fix ρ and cr, and vary n in the set S1, where S1 = {100, 200, 500} .

• We fix cr and n, and vary ρ in the set S2, where S2 = {0.0001, 0.001, 0.01}.

• We fix n and ρ, and vary cr in the set S3, where S3 = {0.01, 0.02, 0.05, 0.1}.

Following the aforementioned three directions, the experiment results of different

approaches are presented in Table 2.1, Table 2.2 and Table 2.3. In each table, com-

prehensive results of the recovery errors related to X and Y as well as the accuracy

of capturing sparse location are demonstrated. In particular, the computation time

with respect to dimensions is summarized in Table 2.1.

Under all these conditions, our alternating algorithm outperforms ALM and NSA

in terms of the rank of X and the sparsity of Y . In particular, the rank of X produced

by our alternating algorithm is consistent with the value we generated through cr in

the case of ρ = 0.001. Moreover, for capturing the sparse locations in matrix Y , our

DC algorithm and alternating algorithm both perform much better than ALM and

NSA. Especially, in the case of ρ = 0.001, n = 100, compared to ALM (20.35%) and

NSA (52.57%), we have obtained 99.45% (DC) and 98.73% (alternating) accuracy

which nearly capture all of the sparse locations in Y . Furthermore, our alternating

algorithm achieves 81.42% accuracy in the case of ρ = 0.01, showing its robustness

to noise.

It can be observed that the computation time of our DC framework is longer

than ALM and NSA. However, our alternating algorithm has comparable execution

30

Page 44: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

time and is much faster than DC framework. Therefore, in the following real-world

applications, we only focus on our alternating algorithm.

2.5.2 Foreground and Background Separation on Surveillance Video

In this experiment, we apply different approaches on the background separation

for an airport surveillance video Li et al. (2004). The dataset contains a sequence of

201 grayscale frames of size 144×176 during a time period. To form the matrix A, we

stack the columns of each frame into a vector and concatenate the vectors together.

We manually add the Gaussian noise by assuming only impulse noise contained in

the video. The Gaussian noise is set at 20dB signal-to-noise ratio (SNR) and ρ =

∥A∥F√144×176×201×10SNR/20 Aybat et al. (2011). The input A′ is generated through A′ =

A + ρG, where G ∈ R25344×201 and each entry of G is generated independently from

the standard Gaussian distribution.

The results of different algorithms are presented in Table 2.4. In addition, we

show the results of three frames of the video in Figure 2.2. Each of the figures in

Figure 2.2 (from left to right) represents the results of ALM, NSA and our algorithm

respectively. In each figure, the first row represents the 15-th, 150-th and 200-th

frame after adding noise, the second row represents background and the third row

relates to the people in the video. Notice that there is one person who is identified

as part of background; this is due to the fact that he/she did not move at all during

the period we focused on.

From Figure 2.2, we conclude that all of the three algorithms can successfully

extract background from surveillance video. Even though the visual quality of back-

ground and foreground are similar among different algorithms, the numerical mea-

surements in Table 2.4 demonstrate that our alternating approach performs better

than both ALM and NSA in terms of the low rank and sparsity. In particular, the

31

Page 45: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

(a) ALM (b) NSA

(c) Our

Figure 2.2: Background Extraction Results of Different Algorithms. In Each Subfig-

ure, the 15-th, 150-th and 200-th Flames after Adding Noise are Shown in the First

Row. The Low-Rank Recoveries by Certain Algorithm are Shown in the Second Row,

and the Sparse Components Extracted by the Algorithm are Shown in the Last Row.

rank of our X is 67, which is about half of the rank of X computed by ALM or NSA.

2.5.3 Shadows Removing from Face Images

Another interesting application of RPCA is to remove shadows and specularities

from face images Candes et al. (2009). Often times, portraits are taken under different

illuminations which introduces errors to face recognition. If we have enough images

under different illuminations of the same face, we can apply RPCA to extract the

32

Page 46: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

features of the face and remove the illumination errors. We apply different algorithms

on the YaleB face database Georghiades et al. (2001). The images of the dataset are

of size 192× 168, and there are 64 illuminations for each subject. Illuminations vary

from both azimuth and elevation, so the shadows for each subject are different in

both location and intensity. Considering one subject, the matrix A is constructed

by concatenating 64 images under all illuminations together. We add 20dB signal-

to-noise ratio which is similar to surveillance video. We obtain a noisy A′ through

A′ = A + ρG, where G ∈ R32256×64 and each entry of G is generated independently

from the standard Gaussian distribution. In this application, the low rank part would

represent human face, while the sparse component is the shadow induced by different

illumination.

Considering Subject01, Subject05 and Subject10 in the dataset, the results of

all three algorithms are shown in Figure 2.3. For better comparison, each subfigure

represents one subject. And the three columns represent observation image, low-rank

recovery and sparsity illumination from left to right. Though the visual quality of

different algorithms are similar, our algorithm obtains lower rank of X and smaller

ℓ0-norm of Y than ALM and NSA, while keeping the relative error comparable, as

demonstrated in Table 2.5.

2.6 Summary

In this chapter, I investigate a non-convex formulation for Robust Principle Com-

ponent Analysis (RPCA) with the capped trace norm and the capped ℓ1-norm. I de-

velop a DC framework as well as a fast alternating algorithm to solve the non-convex

formulation. In the DC framework, ADMM is applied to solve the sub-problem, while

in our fast alternating algorithm, a greedy and easily implementable algorithm for

solving the sub-problem is developed based on the combinatorial optimization. I have

33

Page 47: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

performed extensive experiments on both synthetic and real-world datasets. Results

show that our proposed approaches perform better in recovering the low-rank part

and the sparse component of a given matrix compared with existing work. The cur-

rent work assumes that the complete data is given, i.e., there is no missing entries in

the data. However, in many real applications the data may come with missing values.

I plan to extend my proposed algorithms to solve the RPCA problem with missing

entries in the future. In addition, I plan to study the theoretical properties of the

proposed non-convex formulation.

The RPCA proposed in this chapter is applied to our Drosophila embryo images.

The sparse component represents the noise inside our images produced from high-

throughput techniques. And low-rank structure corresponds the overall contour for

all images in our database. I will filter the sparse component for each BDGP image

and use low-rank structure as input for classification samples in the next chapters.

34

Page 48: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Tab

le2.1:

Synthetic

ResultsforVaryingnwithFixed

ρ=

0.001an

dc r

=0.05.OurAlgorithmsPerform

Much

Better

Than

ALM

andNSA

inTermsof

Cap

turingSparse

Location

sin

Y.

Dim

ension

Alg

rank(X

)∥Y

∥ 0time(sec)

∥X−X

0∥ F

∥X0∥ F

∥Y−Y

0∥ F

∥Y0∥ F

∥A−X−Y∥ F

∥A∥ F

∑ I(Yij=

Y0 ij)

n=100

ALM

608535

0.3869

2.72e-4

6.46e-5

1.83e-6

0.2035

NSA

575228

0.6135

2.98e-4

5.63e-5

8.93e-6

0.5257

OurD

C100

539

93.8845

4.19e-4

6.79e-5

8.41e-5

0.9945

OurA

L5

612

4.3162

3.30e-3

6.11e-4

8.38e-5

0.9873

n=200

ALM

119

34224

9.1882

1.87e-4

6.10e-5

1.92e-6

0.1930

NSA

116

21704

2.1024

2.12e-4

5.34e-5

5.87e-6

0.5060

OurD

C200

3605

552.8801

2.74e-4

3.78e-5

4.66e-5

0.9586

OurA

L10

7148

5.9051

4.52e-4

1.19e-4

4.71e-5

0.8699

n=500

ALM

299

214001

73.7280

1.16e-4

6.18e-5

1.91e-6

0.1927

NSA

259

150302

19.3789

1.16e-4

6.55e-5

3.51e-6

0.4475

OurD

C500

58745

5259.9872

1.64e-4

2.71e-5

2.25e-5

0.8139

OurA

L50

73997

94.8372

1.25e-4

7.96e-5

2.25e-5

0.7527

35

Page 49: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Tab

le2.2:

Synthetic

ResultsforVaryingρwithFixed

n=

100an

dc r

=0.05.OurAlgorithmsPerform

Much

Betterthan

ALM

andNSA

inTermsof

Cap

turingSparse

Location

sin

Y.

Gau

ssiannoise

Alg

rank(X

)∥Y

∥ 0∥X

−X

0∥ F

∥X0∥ F

∥Y−Y

0∥ F

∥Y0∥ F

∥A−X−Y∥ F

∥A∥ F

∑ I(Yij=

Y0 ij)

ρ=

0.0001

ALM

5493

1.57e-5

1.86e-6

7.20e-6

0.9995

NSA

565252

3.13e-5

5.44e-6

9.00e-7

0.5233

OurD

C100

490

1.32e-4

2.06e-5

2.63e-5

0.9999

OurA

L5

504

1.74e-4

1.24e-5

2.52e-5

0.9984

ρ=

0.001

ALM

608535

2.72e-4

6.46e-5

1.83e-6

0.2035

NSA

575228

2.98e-5

5.63e-5

8.93e-6

0.5257

OurD

C100

539

4.19e-4

6.79e-5

8.41e-5

0.9945

OurA

L5

612

3.30e-3

6.11e-4

8.38e-5

0.9873

ρ=

0.01

ALM

688107

3.00e-3

5.83e-4

7.85e-6

0.2387

NSA

575290

2.90e-3

5.21e-4

9.01e-5

0.5204

OurD

C100

4014

2.71e-3

4.82e-4

2.59e-4

0.6478

OurA

L14

2352

3.40e-3

7.09e-4

2.68e-4

0.8142

36

Page 50: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Tab

le2.3:

Synthetic

ResultsforVaryingc r

withFixed

n=

100an

dρ=

0.001.

OurAlgorithmsPerform

Much

Better

than

ALM

andNSA

inTermsof

Cap

turingSparse

Location

sin

Y.

Ran

kratio

Alg

rank(X

)∥Y

∥ 0∥X

−X

0∥ F

∥X0∥ F

∥Y−Y

0∥ F

∥Y0∥ F

∥A−X−Y∥ F

∥A∥ F

∑ I(Yij=

Y0 ij)

c r=

0.01

ALM

588332

4.38e-4

5.70e-5

6.08e-6

0.2159

NSA

575308

4.90e-4

4.93e-5

9.10e-6

0.5183

OurD

C100

499

6.37e-4

7.28e-5

8.38e-5

0.9991

OurA

L1

635

2.11e-3

1.75e-4

8.55e-5

0.9856

c r=

0.05

ALM

608535

2.72e-4

6.46e-5

1.83e-6

0.2035

NSA

575228

2.98e-5

5.63e-5

8.93e-6

0.5257

OurD

C100

539

4.19e-4

6.79e-5

8.41e-5

0.9945

OurA

L5

612

3.30e-3

6.11e-4

8.38e-5

0.9873

c r=

0.1

ALM

638424

2.28e-4

6.57e-5

1.39e-6

0.2063

NSA

565243

1.30e-3

3.12e-4

8.65e-6

0.5244

OurD

C100

727

3.88e-4

6.40e-5

8.54e-5

0.9758

OurA

L10

621

3.11e-3

7.68e-4

8.13e-5

0.9866

37

Page 51: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Table 2.4: Recovery Results of Airport Surveillance. Our Algorithm Produces X

with a Lower Rank and Y with a Smaller ℓ0-norm, While Keeping the Relative Error

Comparable with ALM and NSA.

Alg rank(X) ∥Y ∥0 ∥X+Y−A∥F∥A∥F

ALM 119 4900870 2.8e-7

NSA 129 4915415 6.8e-4

Our 67 4732014 4.4e-5

Table 2.5: YaleFaceB Recovery Results. Our Algorithm Produces X with a Lower

Rank and Y with a Smaller ℓ0-norm, While Keeping the Relative Error Comparable

with ALM and NSA.

Alg rank(X) ∥Y ∥0 ∥X+Y−A∥F∥A∥F

Subject01

ALM 28 1832343 0.0637

NSA 50 1835919 0.0634

Our 27 1707272 0.0637

Subject05

ALM 29 1823015 0.0637

NSA 49 1828608 0.0634

Our 26 1698480 0.0637

Subject10

ALM 28 1848333 0.0637

NSA 48 1827221 0.0635

Our 26 1703188 0.0637

38

Page 52: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

(a) Subject01 (b) Subject05

(c) Subject10

Figure 2.3: Shadow Removing Results of Different Algorithms (Top: ALM, Middle:

NSA, Bottom: Our Proposed Algorithm). In Each Subfigure, the YaleB Images after

Adding Noise are Shown in the Left Column. The Low-Rank Recoveries of Each

Algorithm are Shown in the Middle Column, and the Sparse Errors Corresponding

to Illumination are Shown in the Right Column.

39

Page 53: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Chapter 3

KEYWORDS PREDICTION FOR DROSOPHILA EMBRYO IMAGES

In this chapter, we propose an image-level approach for the automated annotation of

gene expression pattern images. In this approach, the images are first partitioned into

several local patches; and the SIFT descriptor is used to construct a representation for

each patch. We apply both bag-of-words and the sparse coding approaches to compute

high-level representations from the SIFT descriptors. To obtain image features, a

pooling algorithm is utilized to combine patch-level representations to generate image-

level features. We propose to achieve this by a max pooling algorithm, which takes the

strongest signal of the patches into consideration and is shown to perform comparably

well to other pooling algorithms including average pooling and Sqrt pooling. After the

features are generated, we propose an image-level scheme to perform the annotation.

By assuming each image in the group contain all of the terms in the group, the training

set is constructed from individual images (see Figure 3.2). The model built from the

training set is applied to each image in the test groups. A union operation is applied to

perform prediction for each group, that is, the predicted labels from all images within

one group are properly combined to generate the final prediction for this group (see

Figure 3.3). Note that our image-level representation is different from the method

in Zhou and Peng (2007), as we still treat a group of images as one annotation task,

and our image-level annotation scheme is different from the one in Yuan et al. (2012),

since we build a feature representation for each image in the group. The mechanism of

undersampling is introduced in order to provide a balanced distribution for the binary

classification problem. To reduce the random factor introduced by undersampling,

we repeat the undersampling mechanism for multiple times and use majority vote to

40

Page 54: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

perform the prediction. We test our proposed approach on the BDGP images retrieved

from the FlyExpress database (www.flyexpress.net) Kumar et al. (2011). Annotation

results from our study indicate that the image-level scheme outperforms the group-

level scheme. Results also show that majority vote together with undersampling

produces promising results in dealing with imbalanced data.

3.1 Introduction

In this section, we describe the image representation and keyword annotation

schemes. Given a BDGP image, we apply sparse coding as well as the well-known bag-

of-words (BoW) method to represent the gene expression pattern. We then employ

three different pooling methods to generate the image features. Based on the image

features generated, an image-level scheme and a group-level scheme are developed to

annotate the keywords. The data set exhibits an imbalanced data distribution; we

propose to adopt undersampling to retain the efficacy of the standard SVM classifier.

3.1.1 Image Representation

First, we present our framework for image representation, which is closely re-

lated to the bag-of-words (BoW) framework Sivic and Zisserman (2009). The BoW

approach treats each image as a vector recording the frequency of every presented vi-

sual word detected from the image. The vector representation is then used to classify

images into different categories Mikolajczyk and Schmid (2005).

In our framework, we first apply the SIFT descriptor to generate invariant visual

features from local regions on images. A visual codebook is then constructed by ap-

plying the clustering algorithm on the invariant raw descriptors for a subset of images.

The cluster centers are considered as the visual words of the images (codebook). After

the codebook is generated, each descriptor will be represented by a numerical vector

41

Page 55: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

using either BoW or sparse coding approach. A pooling function is then adopted to

summarize hundreds of representers to form one feature vector for one image. An

overview of our framework is given in Figure 3.1.

images SIFT detection

Bag-of-words

Sparse-coding

Model 1Group-level annotation

Model 2 Image-level annotation

codebook

descriptors

representers

representers

pooling

features

Image processing

KeywordsAnnotation

Figure 3.1: The Proposed Framework of Drosophila Gene Pattern Annotation. Given

an Image, the SIFT Detector is Utilized to Generate Descriptors for Local Patches

of This Image. After SIFT Detection, the BoW or Sparse Coding is Applied to

Transform the Descriptors into Representers. Then the Pooling Functions Map the

Representers into Features. We Then Use These Features to Perform Keywords An-

notation via Different Schemes.

Step 1: Feature Detection and Description

The image feature detection step involves partitioning the original image into multiple

regions that serve as local patches for visual description. The images in the FlyExpress

database have been standardized semi-automatically, including alignment. We use

a series of overlapping circles to generate multiple local patches from each image

and adopt the scale-invariant feature transform (SIFT) Lowe (2004) descriptor to

represent each patch. SIFT converts each patch to a 128-dimensional vector. After

this step, each image is represented as a collection of vectors of the same dimension

(128 for SIFT). The collection of vectors are known as descriptors.

42

Page 56: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

We construct the codebook based on the descriptors by selecting a subset of images

and applying the k-means clustering algorithm. Visual words are then defined as

the centers (or centroids) of the clusters. A visual word can be considered as a

representative of several patches with the same characteristics. The codebook length

is the number of clusters, which can be set manually. In our work, we set this

number to 2000 as in Yuan et al. (2012). After the codebook is constructed, the

descriptor for each patch is mapped to a numerical vector based on its relationship

with the codebook through two different ways: the hard assignment (BoW) and the

soft assignment (sparse coding). We present both approaches in the following section.

Step 2: Descriptor Assignment

The BoW performs hard assignment for descriptors; that is, it chooses the closest

visual word in the codebook to represent each descriptor. Then each image can be

represented by a histogram of the visual word. Assume the number of patches for a

given image is N and the size of the codebook is M . Denote Iij = 1 if the ith patch

is assigned to the jth visual word, and 0 otherwise. Then the given image can be

described as H = [h1, h2, . . . , hM ], where

hj =N∑i=1

eij.

A recent study Yuan et al. (2012) shows that combining BoW and spatial infor-

mation would deliver better performance than using only BoW. We add spatial in-

formation into BoW in our study. The spatial BoW can be obtained via augmenting

the simple BoW representation to a larger vector with extra spatial information. We

implement the spatial BoW by adopting the spatial pyramid matching scheme Lazeb-

nik et al. (2006). Denote H as the histogram of an image generated by a non-spatial

BoW; the spatial histogram bag can be written as Sn = [H1, H2, . . . , Hn] where n is

43

Page 57: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

the number of spatial sections. In our work, we partition patches into 2 by 2 sections

on each image, which enlarges the non-spatial BoW representation of the same image

by a factor of 4.

The BoW approach assigns each patch to the closest visual word in the codebook,

which involves solving the following optimization problem:

minimizex

1

2∥Ax− y∥22,

subject to xi ∈ {0, 1},c∑

i=1

xi = 1.(3.1)

It is clear that BoW is a vector quantization from y to the codebook, which means

only one visual word can be picked to represent the patch represented by y. The

hard assignment ignores the relationship between y and other visual words, while the

soft assignment method overcomes the limitation with assigning each descriptor to

a limited number of visual words with different weights simultaneously. Denote the

codebook matrix as A ∈ Rd×M and the descriptor for a given patch as y ∈ Rd, the

soft assignment can be characterized as the following optimization formulation:

minimizex

1

2∥Ax− y∥22 + λ∥x∥1,

subject to xi ≥ 0, i = 1, . . . , c.

(3.2)

where λ is a parameter that controls the sparsity. This is essentially a linear regression

problem with ℓ1 norm regularization, known as Lasso Tibshirani (1996a) in the ma-

chine learning literature and can be solved efficiently by SLEP Liu et al. (2009). We

also consider the spatial sparse coding, which is expected to produce a more accurate

description of images.

Step 3: Feature Pooling

Next, we apply different pooling methods to transform a collection of representers

to one numerical vector (the image feature). After feature detection and description,

44

Page 58: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

each image is represented by a collection of representers (discrete for BoW, continuous

for sparse coding) of the same dimension. Pooling is used to achieve more compact

representations and better robustness to noise and variance. Let X be the sparse

coding representers of SIFT descriptors. We compute the image features by a pre-

defined pooling function:

z = f(X) (3.3)

where f is applied on each column of X. Recall that each column corresponds to

the responses of all the descriptors to one specific item in the codebook. Therefore,

different pooling functions construct different image statistics. We transform a col-

lection of representers into one vector serving as the feature of the image using three

different pooling functions: average pooling, the max pooling and the square root

(Sqrt) pooling.

Average Pooling

For any image represented by a set of descriptors, we can compute a single feature

vector based on some statistics of the representers. For the BoW, a common choice

is to compute the histogram:

z =1

M

M∑i=1

xi, (3.4)

where xi is the representers generated through BoW and M is the number of patches

we have created. In this method, we have taken the average value of all the BoW

representers. For the more sophisticated spatial BoW, the representation z for one

image is the concatenation of histograms associated with various locations. In this

case, z can be seen as a histogram after normalization. The average pooling is one

commonly used pooling functions Le Cun et al. (1990); LeCun et al. (1998); Pinto

et al. (2008) and it can be applied to sparse coding representers accordingly.

45

Page 59: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Max Pooling

Average pooling can be used on both hard-assignment and soft-assignment repre-

senters. Due to the intrinsic continuity property of soft assignment, there are other

ways to proceed the pooling operations. The max pooling function was introduced

in Riesenhuber and Poggio (1999); Serre et al. (2005); Yang et al. (2009), and it maps

each column of X to its max element:

zj = max{x1j, x2j, . . . , xMj}. (3.5)

The max pooling basically uses the strongest signal among multiple patches to repre-

sent the information of the image, and the characteristics of that image are potentially

well captured by the max pooling. Max pooling has been shown to be particularly

well suited to the sparse features Boureau et al. (2010).

3.1.2 Sqrt Pooling

The P-norm is one of statistics that continuously transitions from average to max

pooling, of which a special case is the square root of mean squared statistics (Sqrt)

pooling (P = 2). Mathematically, it is defined as:

zj =

√√√√ 1

M

M∑i=1

x2ij. (3.6)

It is clear that the Sqrt pooling takes advantage of all the information in X, and

the only difference between the Sqrt pooling and average pooling lies in the statistics

they choose to evaluate the information.

Each of these pooling functions captures one aspect of the statistical property

of representers. We consider all three pooling functions in our empirical studies

for comparison, and the results indicate that all three pooling functions perform

comparably well in our application.

46

Page 60: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

3.1.3 Keywords Annotation Schemes

In this section, we describe two different annotation schemes for the keywords

annotation based on various types of features extracted from the BDGP images, e.g.,

the max-pooling spatial sparse codes, the average-pooling BoW, etc. The group-level

annotation scheme takes each group of images as one sample in the training and

testing and is used in previous studies Ji et al. (2009b); Yuan et al. (2012). It has

been shown to give promising results. We propose the image-level annotation scheme

in this chapter which treats individual image as one sample.

Group-level Annotation

In the current BDGP database, groups of images are manually annotated with a set

of keywords. It is possible that not all images in a group are associated with each

keyword in the set. Following this intrinsic structure of the BDGP data, we first

illustrate the group-level annotation scheme.

Given a group of images and the corresponding keywords, the SIFT descriptors

are generated for each image in the group. We then perform hard assignment as

well as soft assignment from the SIFT descriptors to obtain the representers. By

concatenating all the representers of the group together, various pooling functions

can be applied to produce the feature vector. In the group-level scheme, one group of

images are treated as one sample, and the pooling functions are used to extract the

information from all images within this group. We train our model using the training

samples and the model is used to predict the keywords for the testing samples.

The group-level annotation scheme is built directly from the data structure, where

each group of images is represented by one sample in the training and testing pro-

cedure (see Figure 3.2 and Figure 3.3). Since previous studies Ji et al. (2009b);

47

Page 61: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Yuan et al. (2012) have shown that the group-level scheme gives promising results

in keywords annotation for Drosophila gene expression images, we implement the

group-level scheme and use it as a baseline for comparison.

Samples

Fat body/gonad primordiumSomatic muscle primordium

Fat body/gonad primordiumSomatic muscle primordium

Fat body/gonad primordiumSomatic muscle primordium

Fat body/gonad primordiumSomatic muscle primordium

Labels

Fat body/gonad primordium

Somatic muscle primordium

Group-level training set Image-level training set

Sample Labels

Figure 3.2: The Training Process of the Image-Level Scheme Compared with Group-

Level Scheme. In the Training Process, One Group of Images is Considered as One

Sample in the Group-Level Scheme, While Each Image within a Group is Treated as

a Sample in the Image-Level Scheme.

Image-level Annotation

Different from the group-level scheme, each image serves as one sample in the image-

level scheme and the pooling function is applied on an individual image rather than a

group of images. In the training procedure, we assume that each image in the group

is associated with all the keywords for that group. Hence, a larger number of positive

training samples are generated for training our models. The training procedure is

illustrated in Figure 3.2.

48

Page 62: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

'anterior endoderm anlage' 'head mesoderm primordium P4'

'posterior endoderm primordium P2' 'trunk mesoderm primordium P2'

'anterior endoderm anlage''posterior endoderm primordium P2'

'trunk mesoderm primordium P2'

'trunk mesoderm primordium P2' 'head mesoderm primordium P4'

'anterior endoderm anlage' 'posterior endoderm primordium P2'

Goup-level

Image-level

scheme Test images Predicted keywords Final results

'anterior endoderm anlage' 'head mesoderm primordium P4'

'posterior endoderm primordium P2' 'trunk mesoderm primordium P2'

'anterior endoderm anlage' 'head mesoderm primordium P4'

'posterior endoderm primordium P2' 'trunk mesoderm primordium P2'

Figure 3.3: The Testing Process of the Image-Level Scheme Compared with Group-

Level Scheme. In the Testing Stage, the Group-Level Model Will Provide Prediction

for Each Group of Images, while the Image-Level Model will Predict Keywords for

Each Image within this Group. In Addition, in the Image-Level Scheme, We Make

a Union of All the Predicted Keywords in This Group, Which Will Form the Final

Prediction for This Group of Images.

We train our model from samples representing individual images and apply it

to individual images as well. After obtaining the predicted keywords for individual

images, we make a union of the keywords from the same group (see Figure 3.3). The

evaluation of the scheme is done by comparing the predicted keywords and the ground

truth for that group, which is the same as group-level evaluation. The union operation

is introduced to reduce the noise since not all images within a group are associated

with all the keywords for that group. We also include the image-level scheme without

union for comparison purpose.

Both the group-level annotation scheme and the image-level annotation scheme are

implemented and investigated in our study, and our results indicate that image-level

scheme outperforms the group-level scheme as it captures more information inside

each BDGP image.

49

Page 63: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

3.1.4 Undersampling and Majority Vote

Our image data set is highly imbalanced. For example, there are 1081 groups

of images in stage range 2, while only 90 groups contain the term ‘anlage in statu

nascendi’ (Table 3.1). It has been shown that direct application of commonly used

classification algorithms on an imbalanced data set would usually provide sub-optimal

results and one effective approach is to generate a balanced training set Estabrooks

et al. (2004); Chawla et al. (2004); Han et al. (2005). Intuitively, we may do random

sampling to generate a balanced sample set from the imbalanced data, such as over-

sampling and undersampling. Oversampling adds a set A sampled from the minority

class to the original data. In this way, the total number of samples increases by |A|,

and the class distribution is balanced accordingly. On the contrary, undersampling

removes data from the majority class and reduces the size of the original data set. In

particular, denote the minority class as B, we randomly select a subset of size |B| of

majority class examples while keeping the minority class untouched.

In Yuan et al. (2012), oversampling is utilized to deal with imbalanced gene pattern

images data set. We apply the undersampling in our study as oversampling may cause

overfitting He and Garcia (2009). Compared with the results in Yuan et al. (2012), our

experiments produce better sensitivity in prediction. To reduce the random factor

in undersampling and further improve the performance, we do undersampling for

multiple times and combine multiple predictions by majority vote. In our experiment,

we perform undersampling for 21 times and summarize all the models to evaluate the

system we have built. Majority vote Kuncheva et al. (2003) is adopted to perform the

final prediction in our study. We use mean prediction as the baseline for evaluation

and our results show that majority vote performs better.

50

Page 64: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

3.2 Results and Discussion

We design a series of experiments to compare aforementioned approaches for key-

words annotation, and report and analyze the experimental results. In this section,

we present the comparison in four directions:

• Comparison between spatial and non-spatial features;

• Comparison between group-level and image-level schemes;

• Comparison between BoW and sparse coding;

• Comparison of different pooling methods.

Data Description and Experiment Setup

The Drosophila gene expression images used in our work are obtained from the Fly-

Express database, which contains standardized images from the Berkeley Drosophila

Genome Project (BDGP). The Drosophila embryogenesis is partitioned into 6 stage

ranges (1-3, 4-6, 7-8, 9-10, 11-12, 13-17) in BDGP. We focus on the later 5 stage

ranges as there are very small number of keywords appeared in the first stage range.

The Drosophila embryos are 3D objects Weber et al. (2008), and the FlyExpress

database contains 2D images that are taken from different views (lateral, dorsal, and

lateral-dorsal) Mace et al. (2010). As majority of images in the database are in lateral

view Ji et al. (2009b), we focus on the lateral-view images in our study.

Since most keywords are stage-range specific, we build a codebook for each stage

range. Based on the codebook, the spatial BoW features and the spatial sparse

coding representers are computed. We adopt both group-level and image-level anno-

tation schemes to compare the performance. Various pooling functions are applied to

generate the features for annotation.

51

Page 65: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Given the feature representation, we carry out the undersampling for prediction.

We focus on the most frequent 10 terms for each stage range. The groups that

contain a particular term are selected as our positive samples. We use 80% of the

positive samples as the positive training set Trp, the remaining 20% as testing set

Tstp. We also select a subset of samples Tstn (|Tstn| = |Tstp|) which do not contain

the term. The test set Tst = Tstp + Tstn is kept untouched during our experiment.

As the data distribution is imbalanced, we apply undersamplings for 21 times on the

remaining negative groups to form Trn (|Trn| = |Trp|). Thus, we have 21 training

sets Tr(i) = Trp + Trn(i), i = 1, . . . , 21, which are made up of the same positive set

and different negative sets.

We employ the one-against-rest support vector machines (SVM) Chang and Lin

(2011) to annotate our images, where the SVM builds a decision boundary between

Trp and Trn and produces the predictions for Tst. The average and majority vote

are used to summarize multiple undersamplings (see Figure 3.4). The results of the

experiment verify the superiority of our proposed approach.

Comparison between Spatial and Non-spatial Features

We first carry out the comparison between spatial and non-spatial BoW based on the

group-level schemes. We fix the same settings in both cases, including the training

samples, test samples and the pooling functions. The only difference between the

two cases lies in the length of SIFT descriptors: the spatial descriptors is 5 times

the length of those for non-spatial representation. The extra part of the descriptors

captures the location information of multiple patches for the images. As the positive

samples for most terms are less than the negative samples, we perform 21 under-

samplings to balance the training samples. We choose four measurements: accuracy,

AUC, sensitivity and specificity to evaluate the performance of different approaches.

52

Page 66: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Comparison results for all 5 stage ranges by weighted averaging top 10 terms are

shown in Table 3.2. From the comparison, we conclude that using spatial information

improves the performance. Hence, all of the subsequent empirical studies employ the

spatial representation.

Comparison between Group-level and Image-level Schemes

In this experiment, we compare the group-level scheme and the image-level scheme.

For the group-level scheme, a group of images serve as one sample, and the corre-

sponding keywords act as the labels. For image-level scheme, each image of the group

is treated as one sample, and the keywords of the group are assigned to all the images

within the group (for training). Given a group of images, the group-level models will

predict the terms associated with the whole group; the image-level models will predict

the terms associated with each image within the group and the union of predicted

terms from all images within the group will be used as the final prediction.

To demonstrate the efficacy of our image-level scheme, we implement another

image-level scheme without union which evaluates the prediction on the image level.

The results are reported in Figure 3.5. Under the same condition, the union operation

significantly improves the performance of the image-level scheme over all stage ranges

(see Table 3.3).

Our empirical study also shows that the image-level scheme with union outper-

forms the group-level scheme, which can be seen in Figure 3.6. Thus, in the later

part of this chapter, we focus on the image-level scheme. The comparisons over all

stages are summarized in Table 3.4.

53

Page 67: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Comparison between BoW and Sparse Coding

In this part, we compare the performances of spatial BoW and spatial sparse coding

based on image-level scheme, which produces the best result in the previous study.

The BoW method assigns each SIFT descriptor to its closest visual words, and uses a

histogram to summarize all of the representers. On the other hand, the sparse coding

assigns each SIFT descriptor to multiple visual words with different weights, and

different pooling functions are used to summarize the representers. For the sparse

coding, we use a subset of images to tune λ via cross-validation. We use average

pooling for both BoW and sparse coding.

We compute the weighted average for all of the 10 terms for comparison. Overall,

spatial sparse coding performs better than spatial BoW (see Table 3.5), which is

consistent with previous studies Yuan et al. (2012).

Comparison of Different Pooling Methods

To study the difference of various pooling functions, we conduct another experiment

to compare three different pooling methods based on spatial sparse coding. Recall

that the max pooling takes the max element of each column of X, the average pooling

takes the mean value of each column of X, and the Sqrt pooling take the square root

of ℓ2 norm of each column of X. Different pooling functions utilize different statistics

to summarize the representers.

Overall, the experimental results show that our image-level max-pooling approach

achieves the best performance among different combinations of image representation

methods and keywords annotation schemes. In particular, all three pooling methods

achieve comparable results. Furthermore, the max pooling and the Sqrt pooling

produce slightly better results than the average pooling (see tables 3.6 to 3.8).

54

Page 68: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Comparison with Previous Work

There are three main differences between the approach in reference Yuan et al. (2012)

and our chapter. First, we propose different ways of pooling the representers: the

sum pooling is used in Yuan et al. (2012) while we provide three different poolings

in our study (average pooling, max pooling and Sqrt pooling). Secondly, we apply

different ways of treating imbalanced data: the oversampling is used in Yuan et al.

(2012) while we use undersampling in our chapter. Finally, we adopt a different

classification scheme: the group-level scheme is used in Yuan et al. (2012) while we

propose the image-level scheme in our study. Compared with the results in Yuan

et al. (2012), our approach produces better prediction performance (see Table 3.9).

3.3 Summary

In this chapter, I propose an image-level undersampling scheme for annotating

Drosophila gene expression pattern images. In our study, images are represented

by BoW and spatial sparse codes. To transform the representers to sample features,

different pooling functions are adopted. In addition, an image-level annotation scheme

is presented to boost the performance. The random undersampling mechanics are

applied to deal with the imbalanced data in our study. Results on images from the

BDGP database demonstrate the effectiveness of our proposed approach. The current

approach is only applied to the Drosophila embryo images in lateral view. There are

also images in dorsal view and lateral-dorsal view which are not used in our study.

We plan to extend the proposed method to other views of images in the future.

There are two different image annotation tasks for BDGP images: keywords anno-

tation and stage annotation. I present an end-to-end pipeline for keywords annotation

in this chapter, and the stage annotation task will be the focus of the next chapter.

55

Page 69: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Tab

le3.1:

InThisTab

le,WeShow

theTotal

Number

ofGroupsan

dtheCorrespon

dingGroupswhichCon

tain

theTop

10Terms.

For

Mostof

theTerms,theNumber

ofPositiveSam

ples(G

roupsW

hichCon

tain

theTerm)Is

LessThan

the

Number

ofNegativeSam

ples(G

roupsW

hichDoNot

Con

tain

theTerm).

Stages

grou

ps

Term1

Term2

Term3

Term4

Term5

Term6

Term7

Term8

Term9

Term10

4-6

1081

302

259

231

216

199

195

107

9190

87

7-8

877

390

371

358

342

273

241

162

145

103

71

9-10

1072

472

430

429

413

306

249

224

215

128

103

11-12

2113

936

882

604

568

554

475

284

263

261

232

13-16

2816

1068

811

791

642

564

517

492

389

353

324

56

Page 70: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

0.62

0.64

0.66

0.68

0.7

0.72

0.74

0.76

0.78

0.8

0.82

0.84

embryonic/larvalmuscle system

embryonicmidgut

embryonic brain dorsalprothoracic

pharyngealmuscle

embryonichindgut

embryonic dorsalepidermis

embryonic headepidermis

embryonicventral epidermis

ventral nervecord

embryonicforegut

Stage 13-17

Average

MajVote

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

brain primordium anterior midgutprimordium

hindgut properprimordium

posterior midgutprimordium

dorsal epidermisprimordium

trunk mesodermprimordium

head mesodermprimordium

foregutprimordium

ventral nervecord primordium

ventral epidermisprimordium

Stage 11-12

Average

MajVote

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

foregut anlage ventral ectodermprimordium

procephalicectoderm

primordium

anteriorendoderm

primordium

head mesodermprimordium P2

posteriorendoderm

primordium

trunk mesodermprimordium

inclusive hindgutprimordium

ventral nervecord primordium

P3

foregutprimordium

Stage 9-10

Average

MajVote

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

trunk mesodermprimordium P2

dorsal ectodermprimordium

ventral ectodermprimordium P2

head mesodermprimordium P4

pole cell anteriorendoderm anlage

posteriorendoderm

primordium P2

procephalicectoderm anlage

hindgut anlage foregut anlage

Stage 7-8

Average

MajVote

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

pole cell subset cellularblastoderm

dorsal ectodermanlage in statu

nascendi

ventral ectodermanlage in statu

nascendi

maternal yolk nuclei procephalicectoderm anlage

in statu nascendi

anlage in statunascendi

posteriorendoderm anlage

in statu nascendi

Stage 4-6

Average

MajVote

Figure 3.4: Average VS Majority Vote for All Stage Ranges. The Y-Axis of the Figure

Indicates the AUC.

57

Page 71: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Tab

le3.2:

Non

SpatialVSSpatialGroup-LevelAnnotationforTop

10Termsin

AllStages.

SpatialBOW

Perform

sSligh

tly

Betterthan

Non

SpatialBOW

.

Stages

Non

Spatial

Spatial

Accuracy

AUC

sensitivity

specificity

Accuracy

AUC

sensitivity

specificity

4-6

72.77

0.7573

0.7965

0.7180

73.44

0.7609

0.7965

0.7254

7-8

75.09

0.7533

0.7473

0.7626

75.74

0.7557

0.7532

0.7761

9-10

74.15

0.7499

0.7446

0.7553

74.13

0.7504

0.7429

0.7579

11-12

77.18

0.7827

0.8044

0.7610

77.74

0.7876

0.8074

0.7678

13-17

79.78

0.7988

0.8001

0.7975

80.45

0.8048

0.8042

0.8053

58

Page 72: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

brain primordium anterior midgutprimordium

hindgut properprimordium

posterior midgutprimordium

dorsal epidermisprimordium

trunk mesodermprimordium

head mesodermprimordium

foregutprimordium

ventral nerve cordprimordium

ventral epidermisprimordium

Stage 11-12

without union

with union

0.66

0.68

0.7

0.72

0.74

0.76

0.78

0.8

0.82

0.84

0.86

embryonic/larvalmuscle system

embryonic midgut embryonic brain dorsal prothoracicpharyngeal muscle

embryonic hindgut embryonic dorsalepidermis

embryonic headepidermis

embryonic ventralepidermis

ventral nerve cord embryonic foregut

Stage 13-17

without union

with union

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

foregut anlage ventral ectodermprimordium

procephalicectoderm

primordium

anterior endodermprimordium

head mesodermprimordium P2

posteriorendoderm

primordium

trunk mesodermprimordium

inclusive hindgutprimordium

ventral nerve cordprimordium P3

foregutprimordium

Stage 9-10

without union

with union

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

trunk mesodermprimordium P2

dorsal ectodermprimordium

ventral ectodermprimordium P2

head mesodermprimordium P4

pole cell anterior endodermanlage

posteriorendoderm

primordium P2

procephalicectoderm anlage

hindgut anlage foregut anlage

Stage 7-8

without union

with union

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

pole cell subset cellularblastoderm

dorsal ectodermanlage in statu

nascendi

ventral ectodermanlage in statu

nascendi

maternal yolk nuclei procephalicectoderm anlage

in statu nascendi

anlage in statunascendi

posteriorendoderm anlage

in statu nascendi

Stage 4-6

without union

with union

Figure 3.5: Image-Level Scheme with Union VS Image-Level Scheme without Union.

The Y-Axis of the Figure Indicates the AUC.59

Page 73: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Tab

le3.3:

Image-Level

withUnionvsIm

age-Level

withou

tUnion.Im

age-Level

withUnionPerform

sMuch

Betterthan

Withou

tUnion(Sparse

Coding).

Stages

withou

tunion

withunion

Accuracy

AUC

sensitivity

specificity

Accuracy

AUC

sensitivity

specificity

4-6

69.49

0.6796

0.6623

0.6968

77.35

0.7437

0.7002

0.7872

7-8

75.78

0.7509

0.7260

0.7758

78.42

0.7666

0.7209

0.8123

9-10

74.39

0.7469

0.7373

0.7565

78.98

0.7592

0.7481

0.7702

11-12

74.69

0.7574

0.7753

0.7395

79.84

0.8021

0.8153

0.7889

13-17

78.24

0.7687

0.7433

0.7939

82.04

0.8012

0.7655

0.8359

60

Page 74: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Tab

le3.4:

Group-levelVSIm

age-levelSchem

eforTop

10Termsin

AllStages.

Image-levelSpatialBOW

Perform

sSligh

tly

Betterthan

Group-level

SpatialBOW

.

Stages

Group-level

Image-level

Accuracy

AUC

sensitivity

specificity

Accuracy

AUC

sensitivity

specificity

4-6

73.44

0.7609

0.7965

0.7254

76.27

0.7345

0.6891

0.7798

7-8

75.74

0.7557

0.7532

0.7761

78.47

0.7635

0.7006

0.8264

9-10

74.13

0.7504

0.7429

0.7579

74.28

0.7333

0.6977

0.7669

11-12

77.74

0.7876

0.8074

0.7678

80.14

0.7974

0.7946

0.8002

13-17

80.45

0.8048

0.8042

0.8053

81.02

0.7869

0.7412

0.8325

61

Page 75: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

pole cell subset cellularblastoderm

dorsal ectodermanlage in statu

nascendi

ventral ectodermanlage in statu

nascendi

maternal yolk nuclei procephalicectoderm anlage

in statu nascendi

anlage in statunascendi

posteriorendoderm anlage

in statu nascendi

Stage 4-6

group-level

image-level

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

trunk mesodermprimordium P2

dorsal ectodermprimordium

ventral ectodermprimordium P2

head mesodermprimordium P4

pole cell anteriorendoderm anlage

posteriorendoderm

primordium P2

procephalicectoderm anlage

hindgut anlage foregut anlage

Stage 7-8

group-level

image-level

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

foregut anlage ventral ectodermprimordium

procephalicectoderm

primordium

anteriorendoderm

primordium

head mesodermprimordium P2

posteriorendoderm

primordium

trunk mesodermprimordium

inclusive hindgutprimordium

ventral nervecord primordium

P3

foregutprimordium

Stage 9-10

group-level

image-level

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

brain primordium anterior midgutprimordium

hindgut properprimordium

posterior midgutprimordium

dorsal epidermisprimordium

trunk mesodermprimordium

head mesodermprimordium

foregutprimordium

ventral nervecord primordium

ventral epidermisprimordium

Stage 11-12

group-level

image-level

0.65

0.7

0.75

0.8

0.85

0.9

embryonic/larvalmuscle system

embryonicmidgut

embryonic brain dorsalprothoracic

pharyngealmuscle

embryonichindgut

embryonic dorsalepidermis

embryonic headepidermis

embryonicventral epidermis

ventral nervecord

embryonicforegut

Stage 13-17

group-level

image-level

Figure 3.6: Group-level Scheme VS Image-level Scheme with Union. The Y-Axis of

the Figure Indicates the AUC.

62

Page 76: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Tab

le3.5:

BoW

vsSparse

CodingforTop

10Termsin

AllStages.

SpatialSparse

CodingPerform

sSligh

tlyBetterthan

SpatialBOW

(Image-Level).

Stages

BoW

Sparse

coding

Accuracy

AUC

sensitivity

specificity

Accuracy

AUC

sensitivity

specificity

4-6

76.27

0.7345

0.6891

0.7798

77.35

0.7437

0.7002

0.7872

7-8

78.47

0.7635

0.7006

0.8264

78.42

0.7666

0.7209

0.8123

9-10

74.28

0.7333

0.6977

0.7669

78.98

0.7592

0.7481

0.7702

11-12

80.14

0.7974

0.7946

0.8002

79.84

0.8021

0.8153

0.7889

13-17

81.02

0.7869

0.7412

0.8325

82.04

0.8012

0.7655

0.8359

63

Page 77: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Table 3.6: Max Pooling

StagesMax Pooling

Accuracy AUC sensitivity specificity

4-6 75.43 0.7320 0.7002 0.7637

7-8 77.28 0.7616 0.7311 0.7921

9-10 74.46 0.7457 0.7397 0.7517

11-12 78.51 0.7893 0.8015 0.7771

13-17 80.96 0.7952 0.7664 0.8240

Table 3.7: Average Pooling

StagesAverage Pooling

Accuracy AUC sensitivity specificity

4-6 75.21 0.7315 0.6976 0.7654

7-8 76.80 0.7603 0.7332 0.7874

9-10 74.32 0.7439 0.7397 0.7480

11-12 77.49 0.7828 0.8034 0.7623

13-17 80.12 0.7842 0.7521 0.8164

64

Page 78: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Table 3.8: Sqrt Pooling

StagesSqrt Pooling

Accuracy AUC sensitivity specificity

4-6 75.73 0.7360 0.7031 0.7692

7-8 77.49 0.7664 0.7373 0.7955

9-10 74.58 0.7459 0.7380 0.7538

11-12 78.37 0.7907 0.8094 0.7720

13-17 80.73 0.7936 0.7681 0.8191

Table 3.9: Comparison with Previous Research. Here is a Sample Comparison Be-

tween the Previous Results (Refer to Table 5 in Yuan et al. (2012)) and Our Results

(Refer to Table 5) for Stage Range 11-12.

Measure Yuan et al. (2012) Our approach

Sensitivity 0.6544 0.7946

Specificity 0.8577 0.8002

65

Page 79: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Chapter 4

DEEP TRANSFER LEARNING FOR DROSOPHILA EMBRYO ANNOTATION

4.1 Introduction

Deep Learning has been introduced with the objective of moving Machine Learning

closer to one of its original goals: Artificial Intelligence. Recently, Deep Learning

becomes a hot area of Machine Learning research Girshick et al. (2014); Krizhevsky

et al. (2012); Simonyan and Zisserman (2014). It is about learning multiple levels of

representation and abstraction that help to make sense of data such as images. For

example, we aim to learn the concept from the raw image in Figure 4.1: a man sitting

on the ground. Man and Sitting are two pieces of important information conducted

by this image. We build a multi-layer structure to represent the information inside

the image. Starting from the raw pixels, we transform them into abstract features,

i.e., edges and shapes. Based on these abstract features, we try to further transform

them into high-level features such as concept inside the images.

Deep Learning is inspired by the deep structure of the brain, and researchers

want to build a model to mimic the way of how the brain works. Not many successful

attempts were reported before 2006, as the reported positive experimental results are

typically based on two or three levels (i.e., one or two hidden layers). Training deep

networks consistently yielded poor results. There was one exception: LeNet 5, which

built an 8-layer convolutional neural network to recognize the hand-written digits.

Several products have been brought to the market, including a faxed form reader, a

check reading machine, and an automatic teller machine that can read checks and bills

deposited by bank customers LeCun et al. (1998). In the meantime, Vapnik and his co-

66

Page 80: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Input image

. . .Vector of

pixels

Slightly high level

representation

Etc ...

Very high level representation

Man Sitting

Figure 4.1: Learning Multiple Levels of Representation from Image. Raw Image Is

from Yoshua Bengio.

workers developed the Support Vector Machine Boser et al. (1992), and it got widely

used among different applications. SVM is a so-called shallow architecture compared

to Deep Learning models, in which there are only two layers inside, i.e., kernel and

classifier. The superior performance in many applications and the supporting theory

made SVM a hot research area then. In the 1990’s, many researchers abandoned

67

Page 81: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

neural networks with multiple hidden layers because SVMs worked better, and there

were not many successful attempts to train deep networks.

Breakthrough in deep learning has come in 2006: a greedy training algorithm was

proposed in Hinton et al. (2006); Bengio et al. (2007), and it efficiently improved the

performance of deep models. It is realized that some functions cannot be efficiently

represented by architectures that are too shallow in terms of number of tunable ele-

ments, but deep architectures might be able to represent them. The development of

hardware (i.e., GPU) also boosts the spread of Deep Learning applications. Both the

theoretical and practical breakthrough revoked deep learning from its hibernation.

More and more surprising results demonstrated its advantage in image classification.

In this section, we will apply deep learning on our Drosophila embryo images to tackle

the developmental stage/range annotation problem. In the following part, we will first

introduce the architecture of deep models, and then combine the deep learning with

transfer learning to solve the Drosophila embryo images classification problem.

4.2 Deep Learning

Deep Learning has achieved a lot of success in image annotation and object de-

tection recently. Deep Learning models are a class of multi-level systems that can

act on the raw input images directly to generate high-level representations. One

particular type of deep learning models that have achieved practical success is the

deep convolutional neural networks (CNNs) LeCun et al. (1998). This kind of models

stacks many layers of trainable convolutional filters and pooling operations on top

of each other, thereby computes increasingly abstract representations of the inputs.

CNNs is inspired by the visual system’s structure, and typically has eight layers

(Figure 4.2). Compared to standard feed-forward neural networks with similarly-

sized layers, CNNs have much fewer connections and parameters, so they are easier

68

Page 82: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

to train, while their practical performance is likely to be only slightly worse. Deep

CNNs trained with millions of labeled natural images using supervised learning algo-

rithms have led to dramatic performance improvement in natural image recognition

and detection tasks Girshick et al. (2014); Krizhevsky et al. (2012); Simonyan and

Zisserman (2014).

Figure 4.2: LeNet 5: A Layered Model Composed of Convolution and Subsampling

Operations Followed by a Holistic Representation and Ultimately a Classifier for

Handwritten Digits LeCun et al. (1998)

The basic elements inside a CNN (i.e., LeNet 5) can be categorized into the

following 5 classes:

1. Input layer: It will hold the raw pixel values of the input image. In LeNet 5,

the input is an image of width 32 and height 32.

2. Convolution layer: It will compute the output of nodes that are connected to

local regions in the input, each computing a dot product between their weights

and the region they are connected to in the previous layer. It is usually followed

by a non-linear function to introduce a non-linearity. This may result in a next

layer volume such as 28× 28× 6 with 6 filters.

3. Normalization layer: It will normalize the output of a hidden layer to a range

of (0, 1) or (−1, 1) to prevent large magnitude of data flow, and to ensure the

69

Page 83: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

convergence in the long run. This leaves the size of the volume unchanged

(28× 28× 6).

4. Pooling layer: It will perform a downsampling operation along the spatial di-

mensions (width, height), resulting in a next layer such as 14×14×6 with 2×2

sample window. There are multiple pooling methods for reduce the feature di-

mension. Max pooling is commonly used in CNN. By eliminating non-maximal

values, max pooling reduces computation for later layers.

5. Fully-connected layer: It will compute the class scores, resulting in volume of

size 1× 1× 1000, where each of the 1000 numbers correspond to a class score,

such as the 1000 categories of ILSVRC12. As the name implies, each node in

this layer will be connected to all the nodes in the previous layer.

With the elementary transformations we mentioned above, CNNs transform the

input image layer by layer from the original pixel values to the final class scores. Note

that some layers contain parameters and others don’t. In particular, the Convolu-

tion/FC layers perform transformations that are a function of not only the activiations

in the input volume, but also of the parameters (the weights and biases of the nodes).

On the other hand, the ReLU/Pooling layers will implement a fixed function. The

parameters in the Convolution and Fully-Connected (FC) layers will be trained with

gradient descent so that the class scores that the CNN computes are consistent with

the labels in the training set for each image.

In summary, a CNN architecture is a list of layers that transform the image

volume into an output volume (e.g. holding the class scores). There are a few distinct

types of layers (e.g. Convolution/Normalization/Pooling/Fully-Connected are by far

the most popular). Each layer accepts an input 3D volume and transforms it to an

output 3D volume through a differentiable function. Each Layer may or may not have

70

Page 84: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

parameters, for example, Convolution layers have parameters to learn, and Pooling

layers do not have parameters.

The most appealing success of Deep Learning is reported in Krizhevsky et al.

(2012), which won the Large Scale Visual Recognition Challenge 2012 (ILSVRC2012).

This model (Figure 4.3) uses the same fundamental elements as LeNet 5 (Figure 4.2),

i.e., convolution layer, pooling layer, normalization layer and fully-connected layer.

Figure 4.3: AlexNet: A Layered Model Composed of Convolution, Subsampling, and

Further Operations Followed by a Holistic Representation and All-In-All a Landmark

classifier on ILSVRC12 Krizhevsky et al. (2012)

The difference between AlexNet and LeNet 5 lies in three folds:

1. AlexNet uses Rectified Linear Units (ReLU) to replace the Sigmoid Func-

tion (SF) in LeNet 5 to maintain the non-linearity, which make the deep model

faster to train. ReLU is defined as:

SF in LeNet 5: f(x) =1

1 + e−x

ReLU in AlexNet: f(x) = max(0, x).

2. AlexNet adopts Dropout to further penalize the model and prevent overfitting.

Dropout sets the output of each hidden unit to 0 with probability 0.5. So every

time an input is presented, the neural network samples a different architecture,

71

Page 85: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

but all these architectures share weights. This technique reduces complex co-

adaptations of nodes, and forced to learn more robust features that are useful in

conjunction with many different random subsets of the other nodes. Meanwhile,

Dropout roughly doubles the number of iterations required to converge.

3. AlexNet is implemented in two GPUs, while LeNet 5 is implemented in one

CPU, which improves the efficiency of training process.

The common training algorithm for deep models is called Backpropagation. It

is an abbreviation for backward propagation of errors, and it is usually used in

conjunction with an optimization method such as gradient descent. The method cal-

culates the gradient of a loss function with respect to all the weights in the network.

The gradient is fed to the optimization method which in turn uses it to update the

weights, in an attempt to minimize the loss function. Backpropagation requires a

known, desired output for each input value in order to calculate the loss function

gradient. It is therefore usually considered to be a supervised learning method, al-

though it is also used in some unsupervised networks such as autoencoders Hinton

and Salakhutdinov (2006).

Take a typical three-layer neural network (Figure 4.4) for example, and assume

there is a sigmoid function in each layer to retain the non-linearity, the backpropaga-

tion algorithm works as following:

1. Calculate the errors of output neurons:

δα = Outα(1−Outα)(Targetα −Outα)

δβ = Outβ(1−Outβ)(Targetβ −Outβ)

72

Page 86: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Figure 4.4: A Typical Three-Layer Neural Network.

2. Update the weights for output layer:

W ′Aα = WAα + ηδαOutA W ′

Aβ = WAβ + ηδβOutA

W ′Bα = WBα + ηδαOutB W ′

Bβ = WBβ + ηδβOutB

W ′Cα = WCα + ηδαOutC W ′

Cβ = WCβ + ηδβOutC

3. Calculate the errors of hidden layer:

δA = OutA(1−OutA)(δαWAα + δβWAβ)

δB = OutB(1−OutB)(δαWBα + δβWBβ)

δC = OutC(1−OutC)(δαWCα + δβWCβ)

4. Update the weights for hidden layer:

W ′vA = WvA + ηδAInv W ′

uA = WuA + ηδAInu

W ′vB = WvB + ηδBInv W ′

uB = WuB + ηδBInu

W ′vC = WvC + ηδCInv W ′

uC = WuC + ηδCInu

Where η denotes the stepsize of the update, Out denotes the output, In means

the input, and Target denotes the ground truth.

73

Page 87: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

As it is shown above, backpropagation is a generalization of the delta rule to

multi-layered feed-forward networks, by using the chain rule to iteratively compute

gradients for each layer.

There is another direction of deep learning models called Deep Belief Network

(DBN) Hinton et al. (2006). It is essentially a probabilistic generative model and we

will not focus on it in this section.

In the following part, we will combine deep learning with transfer learning to boost

the performance of the CNN and apply the deep transfer learning on our Drosophila

embryo images.

4.3 Deep Transfer Learning

Learning a deep model usually associates with learning millions of parameters,

and this requires a large number of labeled image samples. This bottleneck currently

prevents the applications of CNNs to many biological problems due to the limited

amount of training data. To overcome this difficulty, we propose to develop generic

and problem-independent feature extraction methods, which involves applying previ-

ously obtained knowledge to solve different problems. This is made possible by the

initial success of transferring features among different natural image data sets Don-

ahue et al. (2013b); Razavian et al. (2014); Zeiler and Fergus (2014); Zhang et al.

(2015). These studies trained the models on the ImageNet data set that contains

millions of labeled natural images with thousands of categories. The learned models

were then applied to other image data sets for feature extraction, since the layers of

the deep models are expected to capture intrinsic characteristics of visual objects.

In practice, very few people train an entire CNN from scratch (with random

initialization), because it is relatively rare to have a dataset of sufficient size for

training deep models. Instead, it is common to pretrain a CNN on a very large

74

Page 88: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

dataset (e.g. ImageNet, which contains 1.2 million images with 1000 categories), and

then use the CNN either as an initialization or a fixed feature extractor for the task

of interest. There are two major deep transfer learning scenarios:

1. Feature Extraction: We treat the deep model as a fixed feature extractor,

like SIFT Lowe (2004) or gabor filter Daugman (1980). Take a CNN pretrained

on ImageNet, remove the last fully-connected layer (this layer’s outputs are the

1000 class scores for a different task like ImageNet), then treat the rest of the

CNN as a fixed feature extractor for the new dataset. In an AlexNet, this would

compute a 4096-dimension vector for every image that contains the activiations

of the hidden layer immediately before the classifier. We call these features

deep features. Based on the deep features, we build linear classifiers to solve

our problem. For a deep model, there are usually multiple hidden layers, thus

how to choose the hidden layer to extract features is also an important issue

for classification performance. Usually, we choose the intermediate layers for

feature extraction as they are not data-specific nor task-specific.

2. Fine-tuning: The second strategy is to fine-tune the weights of the pretrained

network by continuing the backpropagation with labeled target images. It is

straightforward to fine-tune all the layers of the CNN, and it is also possible

to keep some layers fixed (to avoid overfitting) and only fine-tune the other

layers of the network. This is motivated by the observation that the earlier

features of a CNN contain more generic features (e.g. edge detectors or shape

detectors) that should be generic information for other tasks, but later layers

of the CNN becomes progressively more specific to the details of the classes

contained in the original dataset. In the case of ImageNet for example, which

contains many dog breeds, a significant portion of the representational power of

75

Page 89: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

the CNN may be devoted to features that are specific to differentiating between

dog breeds Karpathy (2015).

In the following part of this chapter, we will proceed with each of the two directions

mentioned above to tackle our Drosophila embryo images annotation problem.

Before we start to proceed our ideas on Drosophila embryo images, we would like

to first introduce the software we choose to implement our deep transfer learning ideas.

Throughout this section, we applied Caffe to implement our ideas about deep transfer

learning. Caffe is short for Convolution Architecture For Feature Extraction.

It is a deep learning framework developed by the Berkeley Vision and Learning Center

(BVLC) and by community contributors Donahue et al. (2013b).

There are several reasons we choose Caffe as our weapon for this project:

• Caffe provides expressive architecture, which encourages different applications

as well as personal innovation. Models and optimization are defined by config-

uration without hard-coding.

• Caffe provides easy switch between CPU mode and GPU mode by setting a

single flag.

• Caffe is of high speed compared to most public deep learning toolboxes, which

makes it perfect for research experiments and industry deployment. Caffe can

process over 60M images per day with a single NVIDIA K40 GPU, which indi-

cates 1 ms/image for inference and 4 ms/image for learning.

• Caffe is a software community (Figure 4.5) with a lot of contributors. It already

powered academic research projects, startup prototypes, and even large-scale

industrial applications in vision, speech, and multimedia.

76

Page 90: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Figure 4.5: Caffe Community in Github

In the next four subsections, we will illustrate the deep transfer learning method-

ologies in details. And we will present the results of applying deep transfer learning

algorithms on Drosophila embryo image annotation problem.

4.3.1 Deep Features

Deep models can be treated as a feature extractor with pretrained models, as it

captures high-level concept inside the images, which is similar to SIFT Lowe (2004).

One common strategy between CNNs and SIFT extractors is that multiple convolu-

tional filters are utilized for extracting scale-invariant features. However, the filters

are defined in different ways: the filters are predefined in SIFT, while the filters are

learnt through backpropagation in deep models. Modern CNNs take 2-3 weeks to

train across multiple GPUs on ImageNet, thus it is common to see people release

their final CNN models and weights for the benefit of others who can use the net-

works for further application. This kind of well-trained models is called pretrained

models. For example, the Caffe library provides a Model Zoo where people share

their CNN weights. By directly adopting the pretrained deep model, we can plug

77

Page 91: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

in our image and pull out the output from hidden layers. The output from hidden

layers is one kind of representation for the input image, and we call it deep features.

Various classification models can be applied on the deep features. We extract the

deep features of pretrained deep models for our BDGP images, and visualize the deep

features for different layers here. There are a lot of built-in models in Caffe. Specifi-

cally, we extracted features from Caffe built-in model BVLC Reference CaffeNet

as it is efficient for training and fine-tuning. The structure of this model is similar

to the AlexNet (Figure 4.3), and trained with images in ILSVRC 2012. The detailed

structure and the size of each layer for this reference model is illustrated in Table 4.1.

As we can see from Table 4.1, CaffeNet has an 8-layer structure as AlexNet.

Specifically, Caffe uses a 4D tensor to represent a mini-batch of input images. The

shape of the tensor for image input is as follows: [mini-batch size, number of input

feature maps, image height, image width]. This 4D tensor also corresponds to the

weight matrix W. The shape of the tensor is: [number of feature maps at layer m,

number of feature maps at layer m-1, filter height, filter width]. The height and width

represent the size of image/filter respectively. We will further explain the Batch Size

and Channels/Filters in details in the next part.

Deep model is trained with backpropagation algorithm, which is essentially gradi-

ent descent. Gradient descent is a simple algorithm in which we repeatedly make small

steps downward on an error surface defined by a loss function with some parameters.

For a large amount of data, computing the gradient for all of the training samples

is usually time-consuming. Stochastic gradient descent (SGD) Bottou (2010) works

according to the same principles as ordinary gradient descent, but proceeds more

quickly by estimating the gradient from just a few examples at a time instead of the

entire training set. In its simplest form, we estimate the gradient from just a single

example at a time.

78

Page 92: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Table 4.1: Details of BVLC Reference CaffeNet.

Layer Batch Size Channels/Filters Height Width

data 50 3 227 227

conv1 50 96 55 55

pool1 50 96 27 27

norm1 50 96 27 27

conv2 50 256 27 27

pool2 50 256 13 13

norm2 50 256 13 13

conv3 50 384 13 13

conv4 50 384 13 13

conv5 50 256 13 13

pool5 50 256 6 6

Layer Batch Size Channels × Height × Width

fc6 50 4096

fc7 50 4096

fc8 50 1000

prob 50 1000

The backpropogation algorithm that recommended for deep learning is a further

extension on stochastic gradient descent using so-called minibatch. Minibatch SGD

works identically to SGD, except that more than one training sample are used to

make each estimate of the gradient. This technique reduces variance in the estimate

of the gradient, and often makes better use of the hierarchical memory organization

79

Page 93: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

in modern computers. There is a trade-off in the choice of the minibatch size B. With

large B, time is wasted on reducing the variance of the gradient estimator, that time

would be better spent on additional gradient steps. An optimal B is model-, dataset-,

and hardware-dependent, and can be anywhere from 1 to several hundreds. In most

cases, the choice is application dependent. In our application, we divide our images

into minibatches of size 50 for stochastic gradient descent based on the consideration

of both performance and time complexity.

The Channels represent the primary colors for input images. For color images,

Channels = 3, which means Red, Green, Blue channels. For grayscale images,

Channels = 1. The Filters indicates the number of convolution templates we use

for this particular layer (excluding the input layer). The templates are the weights

we used for convolution. Figure 4.6 shows two layers of a CNN. Layer m− 1 contains

four feature maps. Hidden layer m contains two feature maps (h0 and h1). Pixels

(nodes) in h1 and h2 (outlined as blue and yellow squares) are computed from pixels

of layer m− 1 which fall within their 2× 2 receptive field in the layer below (shown

as colored rectangles).

Notice that the receptive field spans all four input feature maps. The weights W 1

and W 2 of h1 and h2 are thus 3D weight tensors. The leading dimension indexes the

input feature maps, while the other two refer to the pixel coordinates. Putting it all

together, W klij denotes the weight connecting each pixel of the k-th feature map at

layer m, with the pixel at coordinates (i, j) of the l-th feature map of layer (m-1).

For the fully-connected layer, the filters are carefully designed so that the convolution

result is of dimension 1.

We have elaborated the structure of CaffeNet in details in this section, and we will

visualize the deep features extracted from multiple hidden layers inside CaffeNet in

the following part. Based on the CaffeNet model (Table 4.1), we extract the outputs

80

Page 94: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

W1

Layer m-1 Hidden Layer m

W1

W2

Convolution filter

h1

h2

00

00w 00

01w

00

10w 00

11w

03

00w 03

01w

03

10w

01

00w01

01w

01

10w

02

00w 02

01w

02

10w

Figure 4.6: Filters for Hidden Layer

of hidden layers and visualize them in this subsection. The outputs will be utilized

as features for classification in later part of this chapter.

For better understanding the mechanism of deep convolution models, we track

the data flow from input to output. We randomly pick a Drosophila embryo image

(Figure 4.7) as input to our model. And we run the deep model in the forward

direction to generate output and visualize the data flow here. Specifically, we get the

multiple feature maps and concatenate them into one figure for illustration. The first

36 feature maps extracted from the first and the second convolution layers are shown

in Figure 4.8 and Figure 4.9. The total 256 feature maps extracted from the fifth

pooling layer are demonstrated in Figure 4.10. The output of the sixth fully connected

layer is shown in Figure 4.11. We can see that the deep features extracted from the

first two layers usually represent the lower-level features, such as edge, contour, and it

is similar to SIFT feature detectors. The features extracted from later layers represent

high-level features for the image. And the last layer is connected to the class scores.

Based on the structure of CNNs, we can see that the features extracted from early

layers represent generic features of input images; while features extracted from later

81

Page 95: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Figure 4.7: Input Image

Figure 4.8: Output of the First Convolution Layer

Figure 4.9: Output of the Second Convolution Layer

layers represent the task-specific features closely related to labels. Therefore, the

features extracted from the middle layers are usually picked up for transfer learning

purpose, as they are neither data-specific nor task-specific. In the next part, we will

82

Page 96: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Figure 4.10: Output of the Fifth Pooling Layer

Figure 4.11: Output of Fully-Connected Layer

build classification models on top of the deep features and compare the performance

among different layers.

4.3.2 Stage Classification with Deep Features

Based on the deep features we extracted from different hidden layers, we build

SVM classifiers to perform the stage classification. We conduct our experiments with

the following settings:

• There are in total 17 stages in Drosophila embryo images. We treat stage 1 to

stage 3 as one label due to the fact that there is subtle difference in early stages.

83

Page 97: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Thus we have 15 labels for the classification, and essentially it is a multi-label

classification problem.

• As there are different views for the images, for example, lateral view, dorsal

view, ventral view, etc. Among all of the images, 60% of them are of lateral

view. So we first focus on annotating the images of lateral view.

• For each stage, we have roughly 250 images of lateral view labeled by domain

expert. The total number of labeled images is 3721, with which we cannot

afford to train a deep neural network from scratch.

• We randomly split the labeled images into two parts for training a model and

validating it: training data (80%) and test data (20%). In addition, we keep

the training data balanced, i.e., the training data are approximately of the same

size for each class.

The classification performance based on deep features from different hidden layers

is demonstrated in Table 4.2. From the results, we can see that the features from early

layers (pool2, conv3, conv4) give better result compared to later layers (conv5, pool5,

fc6, fc7). The reason may lies in two folds: 1) The early layers have large dimension

of features, which contain more information of the images, and they usually represent

the generic features of the images; 2) The later layers are closer to the label layer of the

nature image classification problem (ILSVRC 2012), thus they contain task-specific

features which may not help in our application.

The results in Table 4.2 is comparable to gabor features Yuan et al. (2012), and

it verifies the efficacy of deep features directly extracted from a pretrained model.

So far, we haven’t changed the deep model, as we treat it as a fixed feature

extractor. In the next section, we try to utilize the label information during feature

84

Page 98: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Table 4.2: Performance of Features Extracted from Pretrained Deep Models.

Layer # features accuracy

pool2 43264 77.69%

conv3 64896 77.01%

conv4 64896 77.42%

conv5 43264 75.94%

pool5 9216 76.48%

fc6 4096 72.18%

fc7 4096 69.62%

generation. Specifically, we try to modify the model with our training data. The

methodology of utilizing a small amount of images to update the pretrained model

by backpropagation is called fine-tuning.

4.3.3 Fine-Tuning

There are a lot of scenarios that we only have a small amount of training data,

but we want to train a robust deep model to solve our classification problem. We

may not afford to train a deep convolution neural network from scratch with small

data size, as are a lot of parameters need to be learnt, and we do not have enough

training data to ensure the quality of the final result. In practice, it is very often

beneficial to initialize the weights of our deep model with that from a pretrained

model. Essentially, this methodology helps us to find a good initialization for non-

convex optimization problem. Based on the pretrained model, we utilize our small

dataset to adjust the model towards our application.

85

Page 99: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Fine-tuning is equivalent to training with less data and fewer iterations based on

pretrained models. We keep the structure of pretrained model unchanged, and only

replace the last layer which connected to original task with our new task. Then we

run the backpropogation with our training images for learning the weights inside the

model. The workflow can be illustrated in Figure 4.12:

Natural images

CNN

1000 ImageNet

Classes

Transfer parameters

.

.

.

.

.

.

15 Stages

Drosophila Embryo images

Figure 4.12: The Workflow of Fine-Tuning

The performance over iterations for the fine-tuning is demonstrated in Figure 4.13.

We can see that in the first 200 iterations, the accuracy jumps from 0.1 to 0.7. Then

the increasing speed slows down in the next 300 iterations. However, fine-tuning

after 500 iterations does not help in improving the performance anymore. The results

indicate that fine-tuning usually takes shorter time to converge compared to training

from scratch, which is another merit for its wide applications.

After we obtained the fine-tuned model, we can conduct feature extraction based

on the new model as we discussed before. Then we build SVM classifier on top of

86

Page 100: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Figure 4.13: Fine-Tuning Convergence

the deep features generated from the fine-tuned CNN. The classification results are

shown in Table 4.3.

Compared to Table 4.2, the results indicate that after fine-tuning, the deep fea-

tures can better capture the features inside images for our task. On the other hand,

the results also demonstrate that even without fine-tuning, the deep features can

serve as general image features for classification task.

Table 4.3: Performance of Features Extracted from Fine-Tuned Deep Models.

Layer # features accuracy

pool2 43264 76.48%

conv3 64896 78.35%

conv4 64896 77.95%

conv5 43264 78.36%

pool5 9216 76.61%

fc6 4096 76.48%

fc7 4096 72.85%

87

Page 101: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

From last section, we conclude that initialization with pretrained model improves

both time complexity and accuracy for classification. However, fully initialization

from pretrained models may introduce bias towards pretrained task. There are re-

search investigate the transferability of different layers inside deep model Yosinski

et al. (2014). Here, we also conduct fine-tuning experiments on selected layers besides

fine-tuning on all layers as mentioned above. Specifically, we initialize the parameters

in early layers (from conv1 to conv3) from pretrained models, and randomly initialize

the parameters in later layers (from conv4 to fc7). This is due to the reason that the

later layers contain more task-specific information compared to early layers. The way

of selective initialization is called Net Surgery. The classification results based on

net surgery are shown in Table 4.4.

Table 4.4: Performance of Features Extracted from Deep Models with Net Surgery.

Layer # features accuracy

pool2 43264 76.61%

conv3 64896 78.36%

conv4 64896 78.90%

conv5 43264 78.90%

pool5 9216 76.75%

fc6 4096 76.08%

fc7 4096 74.60%

We can see that there are small improvements of net surgery compared with fine-

tuning on our application. There are research work demonstrate the superiority of

net surgery Yosinski et al. (2014) in other applications.

88

Page 102: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

4.3.4 Fine-Tuning VS Training-From-Scratch

We have verified the efficacy of deep transfer learning on the stage classification

problem of Drosophila embryo images. The experimental results demonstrate that

deep convolution neural network with fine-tuning works well for our specific lack-of-

data application. In this section, we want to compare the performance of fine-tuning

and training-from-scratch with sufficient training data.

In our application, we have large size of images which are labeled with stage range.

It is possible for us to train a robust deep model from scratch. Furthermore, with this

large amount of training data, we may be able to train a better CNN that captures

detailed information and pattern for our images compared to the previous models. In

this section, we focus on the stage range classification problem, and the settings of

the experiment is summarized as following:

• We have 36, 802 training images with labels, and there are in total 6 classes for

stage range classification.

• We randomly split the labeled images into two parts for training a model and

validating it: training data (80%) and test data (20%). In addition, we keep

the training data balanced, i.e., the training data is approximately of the same

size for each stage range.

• We compare the performance between training-from-scratch and fine-tuning.

We can see from the results (Figure 4.14) that fine-tuning outperforms training-

from-scratch in terms of efficiency. Specifically, Fine-tuning can achieve high accu-

racy (90%) within 2000 iterations, while training-from-scratch only can obtain low

accuracy (60%) at the same time. Fine-tuning converges much faster (within 2000

iterations) than training-from-scratch (with 10000 iterations).

89

Page 103: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Figure 4.14: Comparison between Performance of Training-From-Scratch and Fine-

Tuning for Stage Range Classification.

We conclude that fine-tuning achieves better results within the same training time

compared to training CNN from scratch. The underlying reason is that fine-tuning

transferring the already acquired knowledge. Specifically, fine-tuning borrows the

existed deep models as an initialization, and runs backpropagation with task data.

Basically transfer learning and deep learning are integrated in fine-tuning.

4.4 Summary

In this section, we introduced deep transfer learning techniques and its applica-

tion to our Drosophila embryo images. There are two directions of deep models:

convolution neural networks (CNNs) and deep belief networks. We choose CNNs to

be our research direction as it is proved to achieve success in a lot of applications,

including Large Scale Visual Recognition Challenge (ILSVRC). Besides treating a

deep model as a fixed feature extractor, we applied fine-tuning technique to boost

the performance for image annotation. Specifically, we transfer the knowledge from

the pretrained model as our initialization, and apply backpropogation to update the

parameters inside the deep model.

90

Page 104: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

We verify the deep transfer learning techniques on our Drosophila embryo im-

ages on two tasks: stage classification and stage range classification. Both of the

experimental results demonstrate the efficacy of deep transfer learning techniques. In

addition, fine-tuning boosts the efficiency in training a robust deep model.

91

Page 105: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Chapter 5

A TWO-STAGE TRANSFER LEARNING ALGORITHM

5.1 Introduction

In this chapter we propose a two-stage transfer learning framework which com-

putes weights for the data samples from source domain to reduce marginal probabil-

ity difference between the source and target domains. In the first stage, we compute

weights of the source domain data samples to reduce the marginal probability differ-

ences, using Maximum Mean Discrepancy (MMD) Borgwardt et al. (2006); Huang

et al. (2007) as the measure. A base classifier is learned on the re-weighted source

domain data. The second stage utilizes label information in target domain if avail-

able via penalizing the difference between base classifier and target classifier. A novel

feature of our proposed algorithm is that we combined both instance-based method

and model-based method, and the second stage can serve as an online learning step to

consistently update the classifier. The proposed framework is implemented and veri-

fied in solving the bilingual spam detection challenge at LinkedIn Sun et al. (2015).

And it is easily extendable to the multiple language spam detection problems.

5.2 Proposed Pipeline

In this section, we introduce the proposed pipeline to tackle the multi-lingual

spam detection problem at LinkedIn. Specifically, we design a generic pipeline to do

the spam detection for different languages. Here, we start with the spam detection

in Spanish, as Spanish is the second most widely used language in the US. The

pipeline can be easily generated to any other language than Spanish by substituting

92

Page 106: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

the target language of the machine translator. The pipeline contains four steps,

as shown in Figure 5.1. To make use of labeled English samples in the server, we

collect the English labeled data for translation in the first step. During the collection

step, we first apply the language detector Kurokawa et al. (2009) to make sure that

the selected samples are written in English. After collecting a large set of labeled

English content, we use machine translation tools to translate the English content

into Spanish. We select the Google translate API in the second step as it is one

of the most popular machine translators. The second step results in a large set of

machine translated Spanish content, which can be considered as the training samples

for building a Spanish spam detector. Due to the distribution difference between the

machine translated Spanish and the human written Spanish, we propose a two-step

transfer learning algorithm to generate a Spanish spam classifier. Finally, we evaluate

our model on hunman written Spanish content.

Figure 5.1: Proposed Pipeline for Spanish Spam Detection. Machine Translator is

Employed to Translate the Labeled Content from English to Spanish. We Adopt

Transfer Learning to Build a Spam Detection Model. Model Evaluation is Done on

Human Written Spanish.

5.2.1 Feature Generation

In the aforementioned pipeline, a collection of machine translated Spanish data is

available after the machine translation. We generate tf-idf features to represent both

93

Page 107: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

machine translated Spanish and human written Spanish data. The features are then

used for building transfer learning models.

Term frequency-inverse document frequency (tf-idf) is a numerical statistic that

is intended to reflect how important a word is to a document in a collection of corpus.

The tf-idf value increases proportionally to the number of times a word appears in

the document, but is offset by the frequency of the word in the corpus. The offset is

necessary since some words are generally more common than others.

Tf-idf is the product of two statistics, including term frequency (tf) and inverse

document frequency (idf). Various ways for determining the exact values of both

statistics exist. We apply the Scikit-learn (http://scikit-learn.org/stable/) Pe-

dregosa et al. (2011) package to generate the tf-idf features, where tf, idf and tf-idf

are defined as follows:

• tf(t, d) = 1 if term t occurs in document d.

• idf(t,D) = log N|d∈D:t∈d| where N is total number of documents in the corpus,

and |d ∈ D : t ∈ d| is the number of documents that include term t.

• tfidf(t, d,D) = tf(t, d)× idf(t,D).

A high weight in tf-idf is reached by a high term frequency (in the given document)

and a low document frequency of the term in the whole collection of documents; the

weights hence tend to filter out common terms.

5.2.2 Feature Selection

The vocabulary of the corpus is large, and the data matrix for the tf-idf features is

sparse. To study the distribution of machine translated Spanish and human written

Spanish, we first select the top features using the labels.

94

Page 108: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Here, we apply Lasso Tibshirani (1996b) to do the feature selection. Lasso is

essentially a linear regression problem with an l1 norm regularizer. It is often adopted

as an efficient tool to perform the feature selection and its extension remains as a hot

topic in recent research Xiang et al. (2014); Wang et al. (2013). Denote the feature

matrix as A, the label as y, and the model as x. Lasso can be formulated as the

following optimization problem:

minx

1

2∥Ax− y∥22 + λ∥x∥1 (5.1)

where λ is a parameter that controls the balance between the loss function and the

penalty. The Lasso problem can be solved efficiently by FISTA algorithm Beck and

Teboulle (2009). After we get the solution x, the non-zero entries in x correspond to

the selected features in A.

We apply Lasso to both machine translated Spanish and human written Spanish

to investigate the characteristics of these two domains. The top 20 selected features

are listed in Table 5.1. It can be verified that most of the features relate to jobs. We

can see there are some top features shared by machine translated Spanish documents

and human written Spanish documents, but about half of the features are different

between the two domains. This is not surprising: the machine translated jobs and the

manually constructed jobs were taken from non-overlapping data sources. The for-

mer was translated job posts in English at LinkedIn, while the latter was obtained by

crawling the web and finding job posts written in Spanish. While the top 20 features

in both data sets are related to the broad topic of jobs, e.g. business, information, em-

ployee, opportunity, development, team, management (Table 5.1), we cannot expect

the features to be 100% identical over both data sets. We only expect stop words to

be invariant across different data sets, not nouns, verbs, or adjectives. Furthermore,

some of the top 20 features in the manually constructed and the machine translated

95

Page 109: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

sets are morphological cognates such as buscamos = we look and buscando =

looking; cliente = client and clientes = clients, which further illustrates that the

top 20 features are semantically very close in both data sets.

Despite the above issues, the machine translated content has the same overall

intended meaning as the human written content. In addition, as we select more

unigram features, we find larger overlaps of selected features between the two domains.

Thus, transfer learning is a viable approach to solve the problem. Consider the

machine translated Spanish as the source domain, and the human written Spanish as

the target domain, transfer learning aims to borrow the knowledge learnt from source

domain and apply it to target domain. The details of our two-step transfer learning

algorithm are presented in the next section.

5.3 Transfer Learning

In this chapter, we propose a two-step transfer learning algorithm to learn the

content classifier from machine translated Spanish and apply it to human written

Spanish. To our best knowledge, this is the first work that transfers the knowledge

from machine translation corpus to human written corpus in social network.

This chapter is different from a related two-stage transfer learning algorithm Sun

et al. (2011) in the sense that the number of domains are different. The previous

work Sun et al. (2011) focuses on multiple source domain transfer learning, in which

the second stage utilized the smoothness assumption among hypotheses from multiple

sources. However, this chapter focuses on single source domain transfer learning

where the smoothness assumption cannot be established. The proposed algorithm is

a general algorithm that can be adopted to any other languages as well as any binary

classification problem. Denote the machine translated Spanish as the source domain

T , and the human written Spanish as the target domain H. The distributions of the T

96

Page 110: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Table 5.1: Unigram Features for Two Domains

Machine Translated Spanish Human written Spanish

Spanish Features English translation Spanish Features English translation

anos years buscamos look for

buscando search for clientes customers

cliente customer comercial business

clientes customers conocimientos background

desarrollo development cv cv

empresa company desarrollo development

empresas companies gestion management

equipo team equipo team

estan are espana Spain

experiencia experience experiencia experience

favor please funciones functions

gestion management manejo management

http http millones million

informacion information mas more

mundo world proyectos projects

mas other responsable responsible

oportunidad opportunity seguimiento follow up

posicion position turismo travel

servicios services ventas sales

web web area area

97

Page 111: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

andH could be different while the support can be the same with the increase of feature

dimension. Therefore, we make an assumption that the distribution difference mainly

lies in the marginal probability rather than the conditional probability. Based on the

assumption, we propose our first step of transfer learning: instance re-weighting. The

main idea of instance re-weighting is to calculate the weights for the data in the source

domain, and re-weight the source domain for matching the marginal distribution with

the target domain.

After the first step, we train a logistic regression model on the re-weighted source

data, then apply it to the target domain. In addition, if there are some labels available

in the target domain, we propose an adaptive learning algorithm as the second step to

update our model learnt in the first step. The adaptive learning algorithm penalizes

the mis-classification errors in the target domain while minimizes the modifications of

the model. The second step will also make up the conditional distribution difference

between T and H if our assumption in the first step does not hold.

5.3.1 Instance Re-weighting

The first step in the proposed transfer learning framework is the Kernel Mean

Matching (KMM) algorithm Huang et al. (2006), which is a nonparametric method

to directly infer weights for source domain without distribution estimation. KMM

utilizes the unlabeled data to build a bridge for connecting two domains. In general,

the estimation problem with two different distributions Pr(x, y) and Pr′(x, y) is un-

solvable, as the two terms could be arbitrarily far apart. In particular, for arbitrary

Pr(y|x) and Pr′(y|x), there is no way to infer a good estimator based on the available

training samples in the source domain. Hence a simplified version makes assumption

that Pr(x, y) and Pr′(x, y) only differ via Pr(x) and Pr′(x), i.e., the conditional

probability remains unchanged: Pr(y|x) = Pr′(y|x).

98

Page 112: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

The basic assumption behind KMM approach is that between the two domains,

the key difference lies in the marginal distribution Pr(x) rather than the conditional

distribution Pr(y|x). The idea of KMM is to find the appropriate weights for T

which minimize the discrepancy between the mean values of T and H in the common

projected space. Denote xi as the i-th sample in T and x′i as the i-th sample in H.

The mean differences after projection can be formulated as:

minα

∥∥∥∥∥ 1

m

m∑i=1

αiΦ(xi)−1

m′ ,m′∑i=1

Φ(x′i)

∥∥∥∥∥2

. (5.2)

where αi (i = 1, . . . ,m) is the weight we want to learn, m is the number of samples

in T , m′ is the number of samples in H, and Φ is a mapping which maps the raw

features into a latent space.

The formulation is actually a constrained QP (Quadratic Programming) problem.

After weighting, we still want to ensure αPr(x) is close to a distribution, so there

are two constraints that need to be added: αi ∈ [LB,UB] and | 1m

∑mi=1 αi − 1| ≤ ϵ,

where LB and UB is the lower bound and the upper bound of αi respectively. Denote

Kij = k(xi, xj) and κi =mm′

∑m′

j=1 k(xi, x′j). We can rewrite the formulation (5.2) as:

minα

1

m2αTKα− 2

m2κα,

s.t.

∣∣∣∣∣ 1mm∑i=1

αi − 1

∣∣∣∣∣ ≤ ϵ,

LB ≤ αi ≤ UB.

(5.3)

It can be solved by many optimization algorithms, such as interior point methods or

any other successive optimization procedure. We simply apply the optimize module

in SciPy to calculate the solution.

After we obtaining α, we multiply each sample in T with the corresponding αi,

then build a regularized logistic regression model w0 as follows:

minw0

m∑i=1

log(1 + exp(−yiwT0 αixi)) + λ∥w0∥2. (5.4)

99

Page 113: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

If there is no labeled data available in target domain, we apply the model learnt in

Equation (5.4) to the target domain for Spam detection. There are many different

classification models can be adopted to solve this problem, such as Naive Bayes Kib-

riya et al. (2005) and SVM Yang et al. (2007). The reason of choosing the logistic

regression to build our classification model lies in three folds: (1) Logistic regression

has achieved great success in many real-world applications; (2) The processing time

for an online product should be as small as possible, and logistic regression is a perfect

match; (3) It is easy to be generalized to a distributed version.

5.3.2 Adaptive Learning

KMM targets at the scenario that there is no labeled data available in the target

domain. So far, we have not touched any label information in the target domain.

There are scenarios under which a small portion of labeled target data {x′′i , y

′′i |i =

1, . . . ,m′′} is available. For example, the samples may be labeled by crowdsourcing.

Next, we show how to improve our model w0 learnt from KMM by using additional

labeled instances in H. We propose an adaptive learning strategy, which aims to

learn a perturbation ∆f(x) = ∆wTϕ(x) on the basis of w0 to further compensate for

domain mismatch. Denote the classifier we learnt through KMM as w0, the target

classifier we aim to learn as w, we have the following relation: w = w0 +∆w.

Based on the aforementioned assumption that the difference between the distri-

butions of machine translated Spanish and human written Spanish mainly lies in

the marginal distribution, the intuition of the proposed adaptive learning is that the

difference of w and w0 should be as small as possible. There are related previous

work Duan et al. (2009b); Hoffman et al. (2013), which were proposed to learn a

perturbation of the source hyperplane, by minimizing the classification error on la-

beled examples in the target domain. The pertubation can also be considered as

100

Page 114: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Figure 5.2: Illustration of Adaptive Learning. Model w0 Is Borrowed from the Domain

T , and the Model w for the Domain H Should Be Similar to w0. ∆w is a Small

Perturbation We Aim to Learn.

new feature representations that correct for the domain change. The intuition of our

adaptive learning algorithm is consistent with the online learning, thus this step can

be put into online learning system.

Our adaptive learning step aims to learn a perturbation ∆w such that the dif-

ference between prediction f(x) = wTx and f0(x) = wT0 x is small (Figure 5.2). The

following equation holds: f(x) = f0(x) + ∆f(x) = wT0 ϕ(x) + ∆wTϕ(x), where ∆w is

the difference to be learnt here. Therefore, we formulate the adaptive learning as the

following optimization problem:

min∆w

m′′∑i=1

log(1 + exp(−y′′i wTx′′

i )) + λ∥∆w∥22

=min∆w

m′′∑i=1

log(1 + exp(−y′′i (w0 +∆w)Tx′′i )) + λ∥∆w∥22,

(5.5)

where the logistic loss is utilized with the penalty that encourages the perturbation

∥∆w∥ to be as small as possible. Here λ is a tuning parameter to control the balance

between the loss function and the penalty. Problem (5.5) involves a convex smooth

function, which can be solved efficiently by gradient-based method. Adaptive learning

101

Page 115: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

can make up for the potential issue that there are also conditional differences in the

distributions between the machine translated Spanish and human written Spanish.

It can also serve as an online update algorithm when more and more labeled target

data become available.

Finally, we obtain the model w = w0 + ∆w after the proposed two-step trans-

fer learning algorithm. We summarize our two-step transfer learning algorithm in

Algorithm 3.

Algorithm 3 A Two-Step Transfer Learning Algorithm for Bilingual Spam Detection

Require: a large set of labeled samples in T: {xi, yi}, a large set of unlabeled samples

in H: {x′i, y

′i}, (optional) a small set of labeled samples in H: {x′′

i , y′′i }, ∆w = 0

Ensure: w

1: Calculate weights α by solving

minα

∥∥∥∥∥ 1

m

m∑i=1

αiΦ(xi)−1

m′

m′∑i=1

Φ(x′i)

∥∥∥∥∥2

2: Compute w0 by solving

minw0

m∑i=1

log(1 + exp(−yiwT0 αixi)) + λ∥w0∥2

3: (Optional) Compute ∆w by solving

min∆w

m′′∑i=1

log(1 + exp(−y′′i (w0 +∆w)Tx′′i )) + λ∥∆w∥22

4: return w = w0 +∆w

102

Page 116: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

5.4 Experiments

In this section, we demonstrate the effectiveness of our proposed machine learning

pipeline using production data at LinkedIn.

We collect 12000 pieces of English content in group discussions at LinkedIn. The

content is collected from the groups where job posts are not allowed, i.e., job related

content are considered as spam in these groups. The English content is kept balanced

for the two classes: job-related posts and non-job posts. It is used for translation into

Spanish to form the source domain. We translated all the collected human labeled

content from English to Spanish with Google translate API.

In addition, we also collect 12000 pieces of human written Spanish content (bal-

anced) by crowdsourcing, which serves as the target domain. Inside the target domain,

there are 50 pieces of comments considered as labeled samples used in our adaptive

learning step and the remaining ones are treated as test data for evaluation. We gen-

erate tf-idf features for both machine translated Spanish and human written Spanish,

and it turns out that the feature spaces of the two domains differ from each other, as

shown in Table 5.1.

5.4.1 Experimental Settings

Based on the source domain corpus and the target domain corpus, we first generate

the tf-idf features for each domain, and then use the intersection of selected features

to build the Spanish spam detector.

Scikit-learn (http://scikit-learn.org/stable/) Pedregosa et al. (2011) is a

powerful machine learning package in python which provides feature generation, fea-

ture selection as well as classification functions. We make use of scikit-learn to carry

out the experiment: first, we generate the tf-idf features for both domains by use of

103

Page 117: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

feature extraction.text.TfidfTransformer class; then we build the classification model

via linear model.LogisticRegression class.

In our experiment, it turns out that stemmer would not help in classification

tasks, as it reduced the feature dimension of the problem. There are overlapped

unigram features between machine translated Spanish and human written Spanish,

but the feature spaces of these two domains are not identical. We simply choose the

intersection as our feature space to conduct the transfer learning.

We implement our proposed two-step transfer learning algorithm in Python, and

we solve the optimization problem in both KMM and adaptive learning by calling the

scipy.optimize module (http://docs.scipy.org/doc/scipy/reference/optimize.

html). We increase the tf-idf feature dimension to leverage the loss of information

introduced by performing intersection. To verify the efficacy of the proposed two-step

transfer learning algorithm, we compare it with four other algorithms:

• Baseline: We train a model on T , and directly apply it to the test data in H.

• SLT (Small Labeled dataset for Training): We train a model on a small portion

of labeled data in H (50 samples), and apply it to the test data in H.

• KMM: We calculate the weights for data in T via KMM, then train a model on

weighted data in T , and apply it to the test data in H.

• Adaptive Learning: We apply the adaptive learning (Equation (5.3)) to build a

model on the basis of model generated from Baseline, and apply it to the test

data in H.

We choose the above four algorithms for comparison because they are both straight

forward and practically viable algorithms for an online product. To further test

104

Page 118: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

the robustness of different approaches, we conduct our experiments by varying the

dimension of feature space as follows:

• We generate 600 features for both T and H, and use the intersection (380

features) as the feature space.

• We generate 800 features for both T and H, and use the intersection (504

features) as the feature space.

• We generate 1000 features for both T and H, and use the intersection (629

features) as the feature space.

• We generate 2000 features for both T and H, and use the intersection (1260

features) as the feature space.

From our experience, the more features we generated, the better it will generalize to

unseen data.

5.4.2 Experimental Results

The results of all five algorithms are shown in 5.3 and Table 5.2. To encourage

the model to achieve the desired performance for production, we set the threshold of

logistic regression so as to make the precision fixed at 0.95.

The experimental results exhibit the efficacy of our proposed algorithm in the

following aspects:

• SLT works better than Baseline. The reason is that baseline trains a model

in the source domain and tests it in a different domain, while SLT trains and

tests the data from the same domain. The domain difference makes the baseline

perform worse even with a large amount of training data.

105

Page 119: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Table 5.2: Results for Different Dimensions of Features

# tf-idf #intersection algorithm precision recall f1-score AUC

600 320

baseline 0.95 0.44 0.60 0.87

SLT 0.95 0.42 0.63 0.93

KMM 0.95 0.46 0.64 0.94

adaptive 0.95 0.57 0.71 0.93

twoStage 0.95 0.65 0.77 0.95

800 504

baseline 0.95 0.47 0.63 0.87

SLT 0.95 0.51 0.66 0.93

KMM 0.95 0.66 0.78 0.94

adaptive 0.95 0.57 0.71 0.93

twoStage 0.95 0.65 0.77 0.95

1000 629

baseline 0.95 0.46 0.62 0.88

SLT 0.95 0.52 0.67 0.93

KMM 0.95 0.66 0.78 0.94

adaptive 0.95 0.61 0.74 0.93

twoStage 0.95 0.79 0.86 0.95

2000 1260

baseline 0.95 0.50 0.65 0.89

SLT 0.95 0.58 0.72 0.93

KMM 0.95 0.73 0.82 0.95

adaptive 0.95 0.70 0.80 0.94

twoStage 0.95 0.82 0.88 0.96

• KMM works better than SLT. The reason lies in two folds: KMM trains the

model based on the re-weighted source domain data, which leverages the differ-

106

Page 120: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

(a) AUC Curves for 600 Features (b) AUC Curves for 800 Features

(c) AUC Curves for 1000 Features (d) AUC Curves for 2000 Features

Figure 5.3: Experiment Results for Varied Dimensions of Features: 600, 800, 1000 and

2000. In All These Settings, Our Proposed Algorithm Achieves the Best Performance

Compared with Other Algorithms.

ence between the two domains. In addition, the training data size of KMM is

much larger than that of SLT.

• Adaptive learning works better than SLT, but worse than KMM. Since the

reference model in adaptive learn- ing here comes from baseline, which ignores

the distri- bution difference between the two domains. However, it utilizes more

information from the related source domain, which makes it perform better than

SLT.

107

Page 121: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

• Our proposed two-step algorithm produces the best performance, as it combines

the advantages of KMM and adaptive learning.

From the results, we conclude that each of the two steps (Instant Weighting and

Adaptive Learning) improves the performance compared to the baseline method,

while combining the two steps works the best. As the feature dimension increases,

the performance also improves. To avoid overfitting and ensure the efficiency as an

online product, we stop at 2000 features.

The results also verify our assumption about the distributions of two domains, that

is, the main difference lies in the marginal distribution rather than the conditional

difference. The proposed two-step transfer learning algorithm achieves the required

performance for production.

5.4.3 Use of Less Unlabeled Samples in H

In real applications, we may not have enough labeled samples in H. In addition,

we may not have enough unlabeled samples in H in advance for KMM. As the mean

value of projected distribution for the target domain cannot be captured, KMM may

fail to match the difference between the source domain and target domain.

To verify whether the KMM algorithm works with a small sample size in the target

domain, we further design a series of experiments with varied sizes of samples in the

target domain. Specifically, We vary the amounts of unlabeled samples from 20 to

1000 in human written Spanish to test the KMM algorithm, and the results can be

seen in Table 5.3.

We want to emphasize that even with a small sample size (100) in H, we can get

the desired performance for production. The explanation behind the results lies in

the fact that as long as the target samples are balanced, KMM can capture the mean

value of the distribution in the projected space even with small sample size. This is

108

Page 122: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Table 5.3: KMM Results For a Small Number of Samples in the Target Domain

sample size in H precision recall f1-score AUC

20 0.95 0.28 0.44 0.89

50 0.95 0.56 0.71 0.93

100 0.95 0.62 0.75 0.94

200 0.95 0.68 0.79 0.95

500 0.95 0.82 0.82 0.96

1000 0.95 0.86 0.86 0.97

critical for an online product: as in the real world, we want to ensure the model to

work even with small amounts of human written content. As more and more human

written Spanish data become available, we can apply the second step to incrementally

update the model, which can also be considered as online learning.

5.5 Related Work

Cross-lingual content classification is an important topic in natural language pro-

cessing area, and it gains more attention as the social network becomes global Amini

and Goutte (2010); Platt et al. (2010); Meng et al. (2012); Ling et al. (2008); Pan

et al. (2011); Prettenhofer and Stein (2010); Wan et al. (2011); De Smet et al. (2011).

The transfer learning research on text mining Do and Ng (2005); Jiang (2008) mainly

focus on transferring the knowledge from different topics in the same language. The

work done by Vinokourov Vinokourov et al. (2002) can be considered as the unsuper-

vised cross-lingual learning, where the correlation between two languages are learnt

by Canonical Correlation Analysis. In this work, the high correlation between En-

glish and French words indicates that they represent the same semantic information.

It is different from our work as we deal with supervised learning problems. There

109

Page 123: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

is related research that focus on modeling the multi-view representations in which

each language is considered as a separate source, and a joint loss is minimized while

the consistency between languages is ensured Amini and Goutte (2010). One related

work Pan et al. (2011) also represents the data from two languages views and pro-

poses a non-negative matrix tri-factorization (BNMTF) model for the cross-lingual

sentiment classification problem. These methods aim to utilize the labeled data in

two languages and build a joint model for multi-view learning, which differs from our

work as we do not have labeled data in both languages, and for an online product, it

is not practical to translate every piece of message into the other language.

The most relevant work was done by Ling et al. Ling et al. (2008). In their work,

they translated the Chinese web pages into English, and the common parts of two

languages are extracted and used for classification. Specifically, the KL divergence

between the feature distribution and the label distribution is minimized to generate

the classification model. This work bridges the translated English and label set by

common features, and uses a parametric algorithm to build the classifier. In our

work, we make use of non-parametric transfer learning algorithm to migrate the

information learnt from the machine translated Spanish and apply it to the human

written Spanish. To the best of our knowledge, this is the first attempt to take care of

the distribution difference between the machine translated corpus and human written

corpus in social network. Moreover, we propose an adaptive learning algorithm to

update the model online.

5.6 Summary

In this chapter, we develop a machine learning pipeline to tackle the multilingual

spam detection problem at LinkeIn. We propose to build a model pool in which one

model works for one language, and we start with Spanish as it is widely used in US.

110

Page 124: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Due to the lack of labeled data in Spanish, we first translate the English labeled

content into Spanish content via machine translation. Since the distributions of ma-

chine translated Spanish and human written Spanish are not identical, we propose a

two-step transfer learning algorithm to transfer the knowledge learnt from machine

translated Spanish to human written Spanish. In the first step, KMM is applied in

order to match the marginal distribution difference between the two domains. If there

is a small portion of labeled data available in the human written Spanish, we propose

a second step to incrementally update the model obtained from the first step. Results

show that our proposed pipeline achieves the requirements in Spanish spam detection

for industrial products.

The proposed pipeline is a generic system which can be applied to any other

languages by substituting the target lanaguage in machine translation. We plan to

extend our proposed algorithms to solve the French and Chinese spam detection

problems in the future. In addition, the classification pipeline can deal with different

kinds of classification problems as long as there are labeled samples available. We

also plan to apply the pipeline to other text classification problems at LinkedIn.

Furthermore, the adaptive learning algorithm we propose in the second step can serve

as an online learning model. When more labeled data in Spanish become available,

the model will be adaptively updated.

111

Page 125: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Chapter 6

CONCLUSION

In this chapter, I will summarize the major contributions and make conclusions for

this dissertation. In addition, I will also discuss possible future work along transfer

learning direction.

6.1 Summary of Contributions

Transfer learning is proved to be capable of employing knowledge in relative do-

mains to help the learning tasks in target domain. Compared with the traditional

machine learning, transfer learning shows the advantages of reducing the learning

cost and improving the learning efficiency. The major theme of this dissertation is to

demonstrate that how transfer learning algorithms can help in scientific discoveries

with fewer training data available.

Firstly, I propose deep transfer learning methodologies which integrate transfer

learning and deep learning to tackle our Drosophila embryo image annotation prob-

lem. Due to the lack of stage labels, we cannot afford to train a deep stage model

from scratch. A better way to apply deep learning techniques is to transfer knowl-

edge from pretrained models. There are two ways of transferring knowledge from

pretrained models: 1) we can treat the pretrained models as a feature extractor, with

which we generate deep features for our images and build linear classifier on top of it;

2) we can also upgrade the pretrained model via fine-tuning with our limited train-

ing data, thus the deep model is tuned towards our specific task. To deal with the

noise inside the images, I propose a Robust Principle Component Analysis (RPCA)

algorithm filter the to filter the noise of a given image. Among a stack of images,

112

Page 126: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

the sparse component is considered as the noise produced via the high-throughput

technique, and the low-rank structure will be utilized as input of our transfer learning

pipeline. Moreover, I design and implement an end-to-end system which automati-

cally annotate each embryo image with related keywords (ontology terms). In this

system, the SIFT detector is applied to capture the scale-invariant information from

the image, undersampling is adopted to handle the data imbalance issue, and SVM

is applied for building the multi-class classifier.

Secondly, I present an efficient two-stage transfer learning algorithm for solving

the general domain adaptation problem. The distribution of source domain is mapped

towards the target domain in the first stage; and an adaptive learning model is pro-

posed in the second stage to incorporate label information from the target domain

if it is available. The proposed algorithm is applied on LinkedIn’s group discussion

content for spam detection. The promising results verify the efficacy of our proposed

transfer learning algorithm.

6.2 Future Directions

In the deep transfer learning for Drosophila embryo image annotation problem,

I have conducted transfer learning in two directions: either treating the pretrained

model as a feature extractor or fine-tuning the pretrained model. The structure

(settings of layers, nodes in each layer, etc.) of deep models is not touched in both

of the proposed methodologies. One future direction of deep transfer learning is

to modify the structure of the deep model. For example, stage range layer may

be added before the stage layer to represent the hierarchical information within the

labels. On the other hand, the number of nodes in each layer can also be changed

to fit in our application scenario. To tweak the pretrained model towards the task-

specific structure may help to improve the actual performance. However, there is

113

Page 127: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

no theoretical guideline for modifying the deep structure, and all of the changes on

structures require a lot of empirical study.

There are more ways to add prior knowledge into learning process. Besides adding

one layer in the deep structure to represent the hierarchical structure between stage

range and stage, we may also transfer deep features from deep stage range models

and build stage classification model on top of it. One of the potential underlying

reasons is that there are more training data with stage range labels. The deep model

trained from a large amount of data is capable to capture more information about the

temporal changes among different developmental stages about Drosophila embryo.

The proposed two-stage transfer learning pipeline for bilingual spam detection is a

generic system which can be applied to any other language by substituting the target

lanaguage in machine translation. We plan to extend our proposed algorithms to

solve the French and Chinese spam detection problems in the future. In addition, the

classification pipeline can deal with different kinds of classification problems as long

as there are labeled samples available. Furthermore, the adaptive learning algorithm

we propose in the second step can serve as an online learning model. When more

labeled data in Spanish become available, the model will be adaptively updated.

114

Page 128: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

REFERENCES

Amini, M.-R. and C. Goutte, “A co-classification approach to learning from multilin-gual corpora”, Machine learning 79, 1-2, 105–121 (2010).

Aybat, N., D. Goldfarb and G. Iyengar, “Fast first-order methods for stable principalcomponent pursuit”, arXiv preprint arXiv:1105.2126 (2011).

Aytar, Y., Transfer learning for object category detection, Ph.D. thesis, University ofOxford (2014).

Banea, C., R. Mihalcea and J. Wiebe, “Multilingual subjectivity: are more languagesbetter?”, in “Proceedings of the 23rd International Conference on ComputationalLinguistics”, pp. 28–36 (Association for Computational Linguistics, 2010).

Beck, A. and M. Teboulle, “A fast iterative shrinkage-thresholding algorithm for linearinverse problems”, SIAM Journal on Imaging Sciences 2, 1, 183–202 (2009).

Bengio, Y., P. Lamblin, D. Popovici, H. Larochelle et al., “Greedy layer-wise trainingof deep networks”, Advances in neural information processing systems 19, 153(2007).

Benoıt, L., J. Mairal, F. Bach and J. Ponce, “Sparse image representation with epit-omes”, in “Computer Vision and Pattern Recognition (CVPR), 2011 IEEE Con-ference on”, pp. 2913–2920 (IEEE, 2011).

Blitzer, J., M.Dredze and F. Pereira, “Biographies, bollywood, boom-boxes andblenders: Domain adaptation for sentiment classification”, in “ACL”, (2007).

Borgwardt, K. M., A. Gretton, M. J. Rasch, H. P. Kriegel, B. Scholkopf and A. J.Smola, “Integrating structured biological data by kernel maximum mean discrep-ancy”, in “Bioinformatics”, vol. 22, pp. 49–57 (2006).

Boser, B. E., I. M. Guyon and V. N. Vapnik, “A training algorithm for optimalmargin classifiers”, in “Proceedings of the fifth annual workshop on Computationallearning theory”, pp. 144–152 (ACM, 1992).

Bottou, L., “Large-scale machine learning with stochastic gradient descent”, in “Pro-ceedings of COMPSTAT’2010”, pp. 177–186 (Springer, 2010).

Boureau, Y.-L., J. Ponce and Y. LeCun, “A theoretical analysis of feature poolingin visual recognition”, in “International Conference on Machine Learning”, pp.111–118 (2010).

Boyd, S., N. Parikh, E. Chu, B. Peleato and J. Eckstein, Distributed optimizationand statistical learning via the alternating direction method of multipliers (NowPublishers, 2011).

Cai, J., E. Candes and Z. Shen, “A singular value thresholding algorithm for matrixcompletion”, Arxiv preprint Arxiv:0810.3286 (2008).

115

Page 129: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Cai, J., E. Candes and Z. Shen, “A singular value thresholding algorithm for matrixcompletion”, SIAM Journal on Optimization 20, 4, 1956–1982 (2010).

Candes, E., X. Li, Y. Ma and J. Wright, “Robust principal component analysis?”,Journal of ACM 58, 1, 1–37 (2009).

Chandrasekaran, V., S. Sanghavi, P. Parrilo and A. Willsky, “Rank-sparsity incoher-ence for matrix decomposition”, SIAM Journal on Optimization 21, 2, 572–596(2011).

Chang, C.-C. and C.-J. Lin, “LIBSVM: A library for support vector machines”, ACMTransactions on Intelligent Systems and Technology 2, 27:1–27:27, software avail-able at http://www.csie.ntu.edu.tw/~cjlin/libsvm (2011).

Chawla, N. V., N. Japkowicz and A. Kotcz, “Editorial: special issue on learning fromimbalanced data sets”, ACM SIGKDD Explorations Newsletter 6, 1, 1–6 (2004).

Chen, J., J. Liu and J. Ye, “Learning incoherent sparse and low-rank patterns frommultiple tasks”, ACM Transactions on Knowledge Discovery from Data (TKDD)5, 4, 22 (2012).

Daugman, J. G., “Two-dimensional spectral analysis of cortical receptive field pro-files”, Vision research 20, 10, 847–856 (1980).

Daume III, H., “Frustratingly easy domain adaptation”, arXiv preprintarXiv:0907.1815 (2009).

De Smet, W., J. Tang and M.-F. Moens, “Knowledge transfer across multilingualcorpora via latent topics”, in “Advances in Knowledge Discovery and Data Mining”,pp. 549–560 (Springer, 2011).

Deep Learning, “Deep learning — Wikipedia, the free encyclopedia”, URL http://https://en.wikipedia.org/wiki/Deep_learning (2015).

Do, C. and A. Y. Ng, “Transfer learning for text classification”, in “NIPS”, (2005).

Donahue, J., J. Hoffman, E. Rodner, K. Saenko and T. Darrell, “Semi-superviseddomain adaptation with instance constraints”, in “Computer Vision and PatternRecognition (CVPR)”, (2013a).

Donahue, J., Y. Jia, O. Vinyals, J. Hoffman, N. Zhang, E. Tzeng and T. Darrell,“Decaf: A deep convolutional activation feature for generic visual recognition”,arXiv preprint arXiv:1310.1531 (2013b).

Donoho, D., “De-noising by soft-thresholding”, Information Theory, IEEE Transac-tions on 41, 3, 613–627 (1995).

Duan, L., I. W. Tsang, D. Xu and T. S. Chua, “Domain adaptation from multiplesources via auxiliary classifiers”, in “ICML”, pp. 289–296 (2009a).

116

Page 130: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Duan, L., I. W. Tsang, D. Xu and S. J. Maybank, “Domain transfer svm for videoconcept detection”, in “Computer Vision and Pattern Recognition, 2009. CVPR2009. IEEE Conference on”, pp. 1375–1381 (IEEE, 2009b).

Estabrooks, A., T. Jo and N. Japkowicz, “A multiple resampling method for learningfrom imbalanced data sets”, Computational Intelligence 20, 1, 18–36 (2004).

Fan, J., L. Xue and H. Zou, “Strong oracle optimality of folded concave penalizedestimation”, arXiv preprint arXiv:1210.5992 (2012).

Fei-Fei, L., R. Fergus and P. Perona, “One-shot learning of object categories”, PatternAnalysis and Machine Intelligence, IEEE Transactions on 28, 4, 594–611 (2006).

Fink, M., “Object classification from a single example utilizing class relevance met-rics”, Advances in neural information processing systems 17, 449–456 (2005).

Fowlkes, C. C., C. L. Luengo Hendriks, S. V. Keranen, G. H. Weber, O. Rubel,M.-Y. Huang, S. Chatoor, A. H. DePace, L. Simirenko, C. Henriquez, A. Beaton,R. Weiszmann, S. Celniker, B. Hamann, D. W. Knowles, M. D. Biggin, M. B.Eisen and J. Malik, “A quantitative spatiotemporal atlas of gene expression in theDrosophila blastoderm”, Cell 133, 2, 364–374 (2008).

Gao, T., M. Stark and D. Koller, “What makes a good detector?–structured priorsfor learning from few examples”, in “Computer Vision–ECCV 2012”, pp. 354–367(Springer, 2012).

Georghiades, A., P. Belhumeur and D. Kriegman, “From few to many: Illuminationcone models for face recognition under variable lighting and pose”, IEEE Trans.Pattern Anal. Mach. Intelligence 23, 6, 643–660 (2001).

Girshick, R., J. Donahue, T. Darrell and J. Malik, “Rich feature hierarchies for ac-curate object detection and semantic segmentation”, in “Computer Vision andPattern Recognition (CVPR), 2014 IEEE Conference on”, pp. 580–587 (IEEE,2014).

Grumbling, G., V. Strelets and The FlyBase Consortium, “FlyBase: anatomical data,images and queries”, Nucleic Acids Research 34, D484–488 (2006).

Han, H., W.-Y. Wang and B.-H. Mao, “Borderline-smote: A new over-samplingmethod in imbalanced data sets learning”, in “Advances in intelligent computing”,pp. 878–887 (Springer, 2005).

He, H. and E. A. Garcia, “Learning from imbalanced data”, Knowledge and DataEngineering, IEEE Transactions on 21, 9, 1263–1284 (2009).

Hinton, G. E., S. Osindero and Y.-W. Teh, “A fast learning algorithm for deep beliefnets”, Neural computation 18, 7, 1527–1554 (2006).

Hinton, G. E. and R. R. Salakhutdinov, “Reducing the dimensionality of data withneural networks”, Science 313, 5786, 504–507 (2006).

117

Page 131: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Hoffman, J., E. Rodner, J. Donahue, K. Saenko and T. Darrell, “Efficient learning ofdomain-invariant image representations”, in “International Conference on LearningRepresentations”, (2013).

Huang, J., A. Gretton, K. M. Borgwardt, B. Scholkopf and A. J. Smola, “Correct-ing sample selection bias by unlabeled data”, in “Advances in neural informationprocessing systems”, pp. 601–608 (2006).

Huang, J., A. J. Smola, A. Gretton, K. M. Borgwardt and B. Scholkopf, “Correctingsample selection bias by unlabeled data”, in “NIPS”, vol. 19, p. 601 (2007).

Ji, S., Y.-X. Li, Z.-H. Zhou, S. Kumar and J. Ye, “A bag-of-words approach fordrosophila gene expression pattern annotation”, BMC Bioinformatics 10, 1, 119(2009a).

Ji, S., L. Yuan, Y.-X. Li, Z.-H. Zhou, S. Kumar and J. Ye, “Drosophila gene ex-pression pattern annotation using sparse features and term-term interactions”, in“Proceedings of the Fifteenth ACM SIGKDD International Conference on Knowl-edge Discovery and Data Mining”, pp. 407–416 (2009b).

Jiang, J., Domain adaptation in natural language processing (ProQuest, 2008).

Jiang, J. and C. Zhai, “Instance weighting for domain adaptation in nlp”, in “ACL”,vol. 7, pp. 264–271 (2007).

Karpathy, A., “ Convolutional Neural Networks for Visual Recognition”, http://http://cs231n.github.io/transfer-learning/ (2015).

Kibriya, A. M., E. Frank, B. Pfahringer and G. Holmes, “Multinomial naive bayesfor text categorization revisited”, in “AI 2004: Advances in Artificial Intelligence”,pp. 488–499 (Springer, 2005).

Krizhevsky, A., I. Sutskever and G. E. Hinton, “Imagenet classification with deepconvolutional neural networks”, in “Advances in neural information processing sys-tems”, pp. 1097–1105 (2012).

Kulis, B., K. Saenko and T. Darrell, “What you saw is not what you get: Domainadaptation using asymmetric kernel transforms”, in “Computer Vision and PatternRecognition (CVPR), 2011 IEEE Conference on”, pp. 1785–1792 (IEEE, 2011).

Kumar, S., C. Konikoff, B. Van Emden, C. Busick, K. T. Davis, S. Ji, L.-W.Wu, H. Ramos, T. Brody, S. Panchanathan et al., “Flyexpress: visual mining ofspatiotemporal patterns for genes and publications in drosophila embryogenesis”,Bioinformatics 27, 23, 3319–3320 (2011).

Kuncheva, L. I., C. J. Whitaker, C. A. Shipp and R. P. Duin, “Limits on the majorityvote accuracy in classifier fusion”, Pattern Analysis & Applications 6, 1, 22–31(2003).

118

Page 132: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Kurokawa, D., C. Goutte and P. Isabelle, “Automatic detection of translated text andits impact on machine translation”, Proceedings. MT Summit XII, The TwelfthMachine Translation Summit International Association for Machine TranslationHosted by the Association for Machine Translation in the Americas (2009).

Lazebnik, S., C. Schmid and J. Ponce, “Beyond bags of features: Spatial pyramidmatching for recognizing natural scene categories”, in “Proceedings of the 2006IEEE Computer Society Conference on Computer Vision and Pattern Recognition”,pp. 2169–2178 (IEEE Computer Society, Washington, DC, USA, 2006).

Le Cun, B. B., J. Denker, D. Henderson, R. Howard, W. Hubbard and L. Jackel,“Handwritten digit recognition with a back-propagation network”, in “Advances inneural information processing systems”, (Citeseer, 1990).

LeCun, Y., L. Bottou, Y. Bengio and P. Haffner, “Gradient-based learning appliedto document recognition”, Proceedings of the IEEE 86, 11, 2278–2324 (1998).

Lecuyer, E., H. Yoshida, N. Parthasarathy, C. Alm, T. Babak, T. Cerovina, T. R.Hughes, P. Tomancak and H. M. Krause, “Global analysis of mrna localiza-tion reveals a prominent role in organizing cellular architecture and function”,Cell 131, 1, 174–187, URL http://www.sciencedirect.com/science/article/B6WSN-4PTNRXP-R/2/fb52595ebb5af21b63becf189fbe8c95 (2007).

Lecuyer, E., H. Yoshida, N. Parthasarathy, C. Alm, T. Babak, T. Cerovina, T. R.Hughes, P. Tomancak and H. M. Krause, “Global analysis of mrna localizationreveals a prominent role in organizing cellular architecture and function”, Cell131, 1, 174–187 (2007).

Levine, M. and E. H. Davidson, “Gene regulatory networks for development”, Pro-ceedings of the National Academy of Sciences of the United States of Amer-ica 102, 14, 4936–4942, URL http://www.pnas.org/content/102/14/4936.abstract (2005).

Li, L., W. Huang, I. Gu and Q. Tian, “Statistical modeling of complex backgroundsfor foreground object detection”, Image Processing, IEEE Transactions on 13, 11,1459–1472 (2004).

Li, X., Regularized adaptation: Theory, algorithms and applications, Ph.D. thesis,Citeseer (2007).

Lin, Z., M. Chen and Y. Ma, “The augmented lagrange multiplier method for exactrecovery of corrupted low-rank matrices”, Arxiv preprint arXiv:1009.5055 (2010).

Ling, X., G.-R. Xue, W. Dai, Y. Jiang, Q. Yang and Y. Yu, “Can chinese web pagesbe classified with english data source?”, in “Proceedings of the 17th internationalconference on World Wide Web”, pp. 969–978 (ACM, 2008).

Liu, J., S. Ji and J. Ye, SLEP: Sparse Learning with Efficient Projections, Ari-zona State University, URL http://www.public.asu.edu/~jye02/Software/SLEP (2009).

119

Page 133: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Lowe, D. G., “Distinctive image features from scale-invariant keypoints”, Int. J. Com-put. Vision 60, 2, 91–110 (2004).

Mace, D. L., N. Varnado, W. Zhang, E. Frise and U. Ohler, “Extraction and com-parison of gene expression patterns from 2D RNA in situ hybridization images”,Bioinformatics 26, 6, 761–769 (2010).

Mairal, J., R. Jenatton, G. Obozinski and F. Bach, “Convex and network flow op-timization for structured sparsity”, The Journal of Machine Learning Research999888, 2681–2720 (2011).

Matthews, K. A., T. C. Kaufman and W. M. Gelbart, “Research resources fordrosophila: the expanding universe”, Nat Rev Genet 6, 3, 179–193, URL http://dx.doi.org/10.1038/nrg1554 (2005).

Meng, X., F. Wei, X. Liu, M. Zhou, G. Xu and H. Wang, “Cross-lingual mixturemodel for sentiment classification”, in “Proceedings of the 50th Annual Meetingof the Association for Computational Linguistics: Long Papers-Volume 1”, pp.572–581 (Association for Computational Linguistics, 2012).

Mikolajczyk, K. and C. Schmid, “A performance evaluation of local descriptors”,IEEE Trans. Pattern Anal. Mach. Intell. 27, 10, 1615–1630 (2005).

Mirsky, L., “A trace inequality of john von neumann”, Monatshefte fur Mathematik79, 4, 303–306 (1975).

Nesterov, Y., “Introductory lectures on convex programming”, Lecture Notes pp.119–120 (1998).

Pan, J., G.-R. Xue, Y. Yu and Y. Wang, “Cross-lingual sentiment classification via bi-view non-negative matrix tri-factorization”, in “Advances in Knowledge Discoveryand Data Mining”, pp. 289–300 (Springer, 2011).

Pan, S. J. and Q. Yang, “A survey on transfer learning”, IEEE Transactions onKnowledge and Data Engineering (2009).

Pedregosa, F., G. Varoquaux, A. Gramfort, V. Michel, B. Thirion, O. Grisel, M. Blon-del, P. Prettenhofer, R. Weiss, V. Dubourg, J. Vanderplas, A. Passos, D. Courna-peau, M. Brucher, M. Perrot and E. Duchesnay, “Scikit-learn: Machine learning inPython”, Journal of Machine Learning Research 12, 2825–2830 (2011).

Peng, Y., A. Ganesh, J. Wright, W. Xu and Y. Ma, “Rasl: Robust alignment bysparse and low-rank decomposition for linearly correlated images”, in “ComputerVision and Pattern Recognition (CVPR), 2010 IEEE Conference on”, pp. 763–770(IEEE, 2010).

Pinto, N., D. D. Cox and J. J. DiCarlo, “Why is real-world visual object recognitionhard?”, PLoS computational biology 4, 1, e27 (2008).

120

Page 134: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Platt, J. C., K. Toutanova and W.-t. Yih, “Translingual document representationsfrom discriminative projections”, in “Proceedings of the 2010 Conference on Em-pirical Methods in Natural Language Processing”, pp. 251–261 (Association forComputational Linguistics, 2010).

Prettenhofer, P. and B. Stein, “Cross-language text classification using structural cor-respondence learning”, in “Proceedings of the 48th Annual Meeting of the Associa-tion for Computational Linguistics”, pp. 1118–1127 (Association for ComputationalLinguistics, 2010).

Razavian, A. S., H. Azizpour, J. Sullivan and S. Carlsson, “Cnn features off-the-shelf: an astounding baseline for recognition”, in “Computer Vision and PatternRecognition Workshops (CVPRW), 2014 IEEE Conference on”, pp. 512–519 (IEEE,2014).

Recht, B., M. Fazel and P. Parrilo, “Guaranteed minimum-rank solutions of linearmatrix equations via nuclear norm minimization”, SIAM review 52, 3, 471–501(2010).

Riesenhuber, M. and T. Poggio, “Hierarchical models of object recognition in cortex”,Nature neuroscience 2, 11, 1019–1025 (1999).

Sean Carroll, S. W., Jennifer Grenier, From DNA to diversity : molecular geneticsand the evolution of animal design, vol. 2nd Edition (Wiley-Blackwell, 2005).

Serre, T., L. Wolf and T. Poggio, “Object recognition with features inspired by visualcortex”, in “Computer Vision and Pattern Recognition, 2005. CVPR 2005. IEEEComputer Society Conference on”, vol. 2, pp. 994–1000 (IEEE, 2005).

Simonyan, K. and A. Zisserman, “Very deep convolutional networks for large-scaleimage recognition”, arXiv preprint arXiv:1409.1556 (2014).

Singer, A. and M. Cucuringu, “Uniqueness of low-rank matrix completion by rigiditytheory”, SIAM Journal on Matrix Analysis and Applications 31, 4, 1621–1641(2010).

Sivic, J. and A. Zisserman, “Efficient visual search of videos cast as text retrieval”,IEEE Transactions on Pattern Analysis and Machine Intelligence 31, 591–606(2009).

Sun, Q., M. Amin, B. Yan, C. Martell, V. Markman, A. Bhasin and J. Ye, “Transferlearning for bilingual content classification”, in “Proceedings of the 21th ACMSIGKDD International Conference on Knowledge Discovery and Data Mining”,pp. 2147–2156 (ACM, 2015).

Sun, Q., R. Chattopadhyay, S. Panchanathan and J. Ye, “A two-stage weightingframework for multi-source domain adaptation”, in “Advances in neural informa-tion processing systems”, pp. 505–513 (2011).

121

Page 135: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Sun, Q., S. Muckatira, L. Yuan, S. Ji, S. Newfeld, S. Kumar and J. Ye, “Image-leveland group-level models for drosophila gene expression pattern annotation”, BMCbioinformatics 14, 1, 350 (2013a).

Sun, Q., S. Xiang and J. Ye, “Robust principal component analysis via cappednorms”, in “Proceedings of the 19th ACM SIGKDD international conference onKnowledge discovery and data mining”, pp. 311–319 (ACM, 2013b).

Tang, J., X. Wang, H. Gao, X. Hu and H. Liu, “Enriching short text representationin microblog for clustering”, Frontiers of Computer Science 6, 1, 88–101 (2012).

Tao, M. and X. Yuan, “Recovering low-rank and sparse components of matricesfrom incomplete and noisy observations”, SIAM Journal on Optimization 21, 1,57 (2011).

Tao, P. and L. An, “Convex analysis approach to dc programming: Theory, algorithmsand applications”, Acta Math. Vietnam 22, 1, 289–355 (1997).

Tibshirani, R., “Regression shrinkage and selection via the lasso”, Journal of theRoyal Statistical Society Series B 58, 1, 267–288 (1996a).

Tibshirani, R., “Regression shrinkage and selection via the lasso”, Journal of theRoyal Statistical Society. Series B (Statistical Methodology) pp. 267–288 (1996b).

Toh, K. and S. Yun, “An accelerated proximal gradient algorithm for nuclear normregularized linear least squares problems”, Pacific Journal of Optimization 6, 615-640, 15 (2010).

Tomancak, P., A. Beaton, R. Weiszmann, E. Kwan, S. Shu, S. E. Lewis, S. Richards,M. Ashburner, V. Hartenstein, S. E. Celniker and G. M. Rubin, “Systematic deter-mination of patterns of gene expression during Drosophila embryogenesis”, GenomeBiology 3, 12 (2002).

Tomancak, P., B. Berman, A. Beaton, R. Weiszmann, E. Kwan, V. Hartenstein,S. Celniker and G. Rubin, “Global analysis of patterns of gene expression duringDrosophila embryogenesis”, Genome Biology 8, 7, R145 (2007).

Vinokourov, A., N. Cristianini and J. S. Shawe-taylor, “Inferring a semantic repre-sentation of text via cross-language correlation analysis”, in “Advances in neuralinformation processing systems”, pp. 1473–1480 (2002).

Wan, C., R. Pan and J. Li, “Bi-weighting domain adaptation for cross-language textclassification”, in “IJCAI Proceedings-International Joint Conference on ArtificialIntelligence”, vol. 22, p. 1535 (2011).

Wang, J., J. Zhou, P. Wonka and J. Ye, “Lasso screening rules via dual polytopeprojection”, in “Advances in Neural Information Processing Systems”, pp. 1070–1078 (2013).

122

Page 136: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Wang, P. and C. Domeniconi, “Towards a universal text classifier: Transfer learningusing encyclopedic knowledge”, in “Data Mining Workshops, 2009. ICDMW’09.IEEE International Conference on”, pp. 435–440 (IEEE, 2009).

Weber, G. H., O. Rubel, M.-Y. Huang, A. H. DePace, C. C. Fowlkes, S. V. E. Kera-nen, C. L. Luengo Hendriks, H. Hagen, D. W. Knowles, J. Malik, M. D. Biggin andB. Hamann, “Visual exploration of three-dimensional gene expression using physi-cal views and linked abstract views”, IEEE/ACM Transactions on ComputationalBiology and Bioinformatics 99, 1 (2008).

Wright, J., A. Ganesh, S. Rao, Y. Peng and Y. Ma, “Robust principal componentanalysis: Exact recovery of corrupted low-rank matrices via convex optimization”,submitted to Journal of the ACM (2009).

Xiang, S., X. Shen and J. Ye, “Efficient sparse group feature selection via nonconvexoptimization”, in “Proceedings of The 30th International Conference on MachineLearning”, pp. 284–292 (2012a).

Xiang, S., T. Yang and J. Ye, “Simultaneous feature and feature group selectionthrough hard thresholding”, in “Proceedings of the 20th ACM SIGKDD interna-tional conference on Knowledge discovery and data mining”, pp. 532–541 (ACM,2014).

Xiang, S., Y. Zhu, X. Shen and J. Ye, “Optimal exact least squares rank mini-mization”, in “Proceedings of the 18th ACM SIGKDD international conference onKnowledge discovery and data mining”, pp. 480–488 (ACM, 2012b).

Xu, H., C. Caramanis and S. Sanghavi, “Robust pca via outlier pursuit”, arXivpreprint arXiv:1010.4237 (2010).

Yang, J., R. Yan and A. G. Hauptmann, “Cross-domain video concept detection usingadaptive svms”, in “Proceedings of the 15th international conference on Multime-dia”, pp. 188–197 (ACM, 2007).

Yang, J., K. Yu, Y. Gong and T. Huang, “Linear spatial pyramid matching usingsparse coding for image classification”, in “Computer Vision and Pattern Recogni-tion, 2009. CVPR 2009. IEEE Conference on”, pp. 1794–1801 (IEEE, 2009).

Yao, B., X. Jiang, A. Khosla, A. L. Lin, L. Guibas and L. Fei-Fei, “Human actionrecognition by learning bases of action attributes and parts”, in “Computer Vision(ICCV), 2011 IEEE International Conference on”, pp. 1331–1338 (IEEE, 2011).

Ye, J. and J. Liu, “Sparse methods for biomedical data”, ACM SIGKDD ExplorationsNewsletter 14, 1, 4–15 (2012).

Yosinski, J., J. Clune, Y. Bengio and H. Lipson, “How transferable are features indeep neural networks?”, in “Advances in Neural Information Processing Systems”,pp. 3320–3328 (2014).

123

Page 137: Transfer Learning for BioImaging and Bilingual ... Learning for BioImaging and Bilingual Applications by ... Jiayu Zhou, Sen Yang, Yashu Liu, Zhi Nie, ... YaleB Images after Adding

Yuan, L., C. Pan, S. Ji, M. McCutchan, Z.-H. Zhou, S. J. Newfeld, S. Kumar andJ. Ye, “Automated annotation of developmental stages of drosophila embryos inimages containing spatial patterns of expression”, Bioinformatics 30, 2, 266–273(2014).

Yuan, L., A. Woodard, S. Ji, Y. Jiang, Z.-H. Zhou, S. Kumar and J. Ye, “Learningsparse representations for fruit-fly gene expression pattern image annotation andretrieval”, BMC bioinformatics 13, 1, 107 (2012).

Zeiler, M. D. and R. Fergus, “Visualizing and understanding convolutional networks”,in “Computer Vision–ECCV 2014”, pp. 818–833 (Springer, 2014).

Zhang, T., “Analysis of multi-stage convex relaxation for sparse regularization”,JMLR 11, 1081–1107 (2010).

Zhang, W., R. Li, T. Zeng, Q. Sun, S. Kumar, J. Ye and S. Ji, “Deep model basedtransfer and multi-task learning for biological image analysis”, in “Proceedings ofthe 21th ACM SIGKDD International Conference on Knowledge Discovery andData Mining”, pp. 1475–1484 (ACM, 2015).

Zhou, J. and H. Peng, “Automatic recognition and annotation of gene expressionpatterns of fly embryos”, Bioinformatics 23, 5, 589–596 (2007).

124