edinburgh research explorer€¦ · abide by the legal requirements associated with these rights....

16
Edinburgh Research Explorer Robust Subjective Visual Property Prediction from Crowdsourced Pairwise Labels Citation for published version: Fu, Y, Hospedales, TM, Xiang, T, Xiong, J, Gong, S, Wang, Y & Yao, Y 2016, 'Robust Subjective Visual Property Prediction from Crowdsourced Pairwise Labels', IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 38, no. 3, pp. 563-577. https://doi.org/10.1109/TPAMI.2015.2456887 Digital Object Identifier (DOI): 10.1109/TPAMI.2015.2456887 Link: Link to publication record in Edinburgh Research Explorer Document Version: Peer reviewed version Published In: IEEE Transactions on Pattern Analysis and Machine Intelligence General rights Copyright for the publications made accessible via the Edinburgh Research Explorer is retained by the author(s) and / or other copyright owners and it is a condition of accessing these publications that users recognise and abide by the legal requirements associated with these rights. Take down policy The University of Edinburgh has made every reasonable effort to ensure that Edinburgh Research Explorer content complies with UK legislation. If you believe that the public display of this file breaches copyright please contact [email protected] providing details, and we will remove access to the work immediately and investigate your claim. Download date: 02. May. 2020

Upload: others

Post on 30-Apr-2020

0 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Edinburgh Research Explorer€¦ · abide by the legal requirements associated with these rights. Take down policy The University of Edinburgh has made every reasonable effort to

Edinburgh Research Explorer

Robust Subjective Visual Property Prediction fromCrowdsourced Pairwise Labels

Citation for published version:Fu, Y, Hospedales, TM, Xiang, T, Xiong, J, Gong, S, Wang, Y & Yao, Y 2016, 'Robust Subjective VisualProperty Prediction from Crowdsourced Pairwise Labels', IEEE Transactions on Pattern Analysis andMachine Intelligence, vol. 38, no. 3, pp. 563-577. https://doi.org/10.1109/TPAMI.2015.2456887

Digital Object Identifier (DOI):10.1109/TPAMI.2015.2456887

Link:Link to publication record in Edinburgh Research Explorer

Document Version:Peer reviewed version

Published In:IEEE Transactions on Pattern Analysis and Machine Intelligence

General rightsCopyright for the publications made accessible via the Edinburgh Research Explorer is retained by the author(s)and / or other copyright owners and it is a condition of accessing these publications that users recognise andabide by the legal requirements associated with these rights.

Take down policyThe University of Edinburgh has made every reasonable effort to ensure that Edinburgh Research Explorercontent complies with UK legislation. If you believe that the public display of this file breaches copyright pleasecontact [email protected] providing details, and we will remove access to the work immediately andinvestigate your claim.

Download date: 02. May. 2020

Page 2: Edinburgh Research Explorer€¦ · abide by the legal requirements associated with these rights. Take down policy The University of Edinburgh has made every reasonable effort to

1

Robust Subjective Visual Property Predictionfrom Crowdsourced Pairwise Labels

Yanwei Fu, Timothy M. Hospedales, Tao Xiang, Jiechao Xiong,Shaogang Gong, Yizhou Wang, and Yuan Yao

Abstract—The problem of estimating subjective visual properties from image and video has attracted increasing interest. A subjectivevisual property is useful either on its own (e.g. image and video interestingness) or as an intermediate representation for visualrecognition (e.g. a relative attribute). Due to its ambiguous nature, annotating the value of a subjective visual property for learninga prediction model is challenging. To make the annotation more reliable, recent studies employ crowdsourcing tools to collect pairwisecomparison labels. However, using crowdsourced data also introduces outliers. Existing methods rely on majority voting to prune theannotation outliers/errors. They thus require a large amount of pairwise labels to be collected. More importantly as a local outlierdetection method, majority voting is ineffective in identifying outliers that can cause global ranking inconsistencies. In this paper, wepropose a more principled way to identify annotation outliers by formulating the subjective visual property prediction task as a unifiedrobust learning to rank problem, tackling both the outlier detection and learning to rank jointly. This differs from existing methodsin that (1) the proposed method integrates local pairwise comparison labels together to minimise a cost that corresponds to globalinconsistency of ranking order, and (2) the outlier detection and learning to rank problems are solved jointly. This not only leads tobetter detection of annotation outliers but also enables learning with extremely sparse annotations.

Index Terms—Subjective visual properties, outlier detection, robust ranking, robust learning to rank, regularisation path

F

1 INTRODUCTION

The solutions to many computer vision problems involvethe estimation of some visual properties of an image orvideo, represented as either discrete or continuous vari-ables. For example scene classification aims to estimatethe value of a discrete variable indicating which scenecategory an image belongs to; for object detection thetask is to estimate a binary variable corresponding thepresence/absence of the object of interest and a set ofvariables indicating its whereabouts in the image plane(e.g. four variables if the whereabouts are representedas bounding boxes).Most of these visual properties areobjective; that is, there is no or little ambiguity in theirtrue values to a human annotator.

In comparison, the problem of estimating subjectivevisual properties is much less studied. This class ofcomputer vision problems nevertheless encompass a va-riety of important applications. For example: estimatingattractiveness [1] from faces would interest social mediaor online dating websites; and estimating properties ofconsumer goods such as shininess of shoes [2] improvescustomer experiences on online shopping websites. Re-

• Yanwei Fu, Timothy M. Hospedales, Tao Xiang, and Shaogang Gong arewith the School of Electronic Engineering and Computer Science, QueenMary University of London, E1 4NS, UK.Email: {y.fu,t.hospedales,t.xiang,s.gong}@qmul.ac.uk.

• Jiechao Xiong, and Yuan Yao are with the School of Mathematical Sciences,Peking University, China. Email:[email protected]

• Tao Xiang and Yuan Yao are the corresponding authours.• Yizhou Wang is with the School of Electronics Engineering and Computer

Science, Peking University, China. Email: [email protected]

cently, the problem of automatically predicting if peoplewould find an image or video interesting has started toreceive increasing attention [3], [4], [5]. Interestingnessprediction has a number of real-world applications. Inparticular, since the number of images and videos up-loaded to the Internet is growing explosively, people areincreasingly relying on image/video recommendationtools to select which ones to view. Given a query, rankingthe retrieved data with relevance to the query basedon the predicted interestingness would improve usersatisfaction. Similarly user stickiness can be increasedif a media-sharing website such as YouTube can rec-ommend videos that are both relevant and interesting.Other applications such as web advertising and videosummarisation can also benefit. Subjective visual prop-erties such as the above-mentioned ones are useful ontheir own. But they can also be used as an intermediaterepresentation for other tasks such as visual recognition,e.g., different people can be recognised by how pale theirskin complexions are and how chubby their faces are [6].When used as a semantically meaningful representation,these subjective visual properties often are referred to asrelative attributes [2], [6], [7].

Learning a model for subjective visual property (SVP)prediction is challenging primarily due to the difficultiesin obtaining annotated training data. Specifically, sincemost SVPs can be represented as continuous variables(e.g. an interestingness/aesthetics/shininess score witha value range of 0 to 1 with 1 being most interest-ing/aesthetically appealing/shinning), SVP predictioncan be cast as a regression problem – the low-levelfeature values are regressed to the SVP values given

Page 3: Edinburgh Research Explorer€¦ · abide by the legal requirements associated with these rights. Take down policy The University of Edinburgh has made every reasonable effort to

2

a set of training data annotated with their true SVPvalues. However, since by definition these properties aresubjective, different human annotators often struggle togive an absolute value and as a result the annotations ofdifferent people on the same instance can vary hugely.For example, on a scale of 1 to 10, different people willhave very different ideas on what a scale 5 means for animage, especially without any common reference point.On the other hand, it is noted that humans can in generalmore accurately rank a pair of data points in terms oftheir visual properties [8], [9] , e.g. it is easier to judgewhich of two images is more interesting relatively thangiving an absolute interestingness score to each of them.Most existing studies [2], [1], [9] on SVP prediction thustake a learning to rank approach [10], where annotatorsgive comparative labels about pairs of images/videosand the learned model is a ranking function that predictsthe SVP value as a ranking score.

To annotate these pairwise comparisons, crowdsourc-ing tools such as Amazon Mechanic Turk (AMT) areresorted to, which allow a large number of annotatorsto collaborate at very low cost. Data annotation basedon crowdsourcing is increasingly popular [6], [2], [4], [5]recently for annotating large-scale datasets. However,this brings about two new problems: (1) Outliers –The crowd is not all trustworthy: it is well known thatcrowdsourced data are greatly affected by noise and out-liers [11], [12], [13] which can be caused by a number offactors. Some workers may be lazy or malicious [14], pro-viding random or wrong annotations either carelessly orintentionally; some other outliers are unintentional hu-man errors caused by the ambiguous nature of the data,thus are unavoidable regardless how good the attitudesof the workers are. For example, the pairwise rankingfor Figure 1(a) depends on the cultural/psychologicalbackground of the annotator – whether s/he is morefamiliar/prefers the story of Monkey King or CookieMonster1. When we learn the model from labels collectedfrom many people, we essentially aim to learn the con-sensus, i.e. what most people would agree on. Therefore,if most of the annotators growing up watching SesameStreet thus consciously or subconsciously consider theCookie Monster to be more interesting than the MonkeyKing, their pairwise labels/votes would represent theconsensus. In contrast, one annotator who is familiarwith the stories in Journey to the West may choosethe opposite; his/her label is thus an outlier underthe consensus. (2) Sparsity – the number of pairwisecomparisons required is much bigger than the number ofdata points because n instances define a O(n2) pairwisespace. Consequently, even with crowdsourcing tools,the annotation remains be sparse, i.e. not all pairs arecompared and each pair is only compared a few times.

To deal with the outlier problem in crowdsourceddata, existing studies take a majority voting strategy [6],[2], [4], [15], [16], [17], [18]. That is, a large budget

1. This is also known as Halo Effect in Psychology.

??

Who is smiling more?Who is more interesting?(a) (b)

Figure 1. Examples of pairwise comparisons of subjectivevisual properties.

of 5 − 10 times the number of actual annotated pairsrequired is allocated to obtain multiple annotations foreach pair. These annotations are then averaged over so asto eliminate label noise. However, the effectiveness of themajority voting strategy is often limited by the sparsityproblem – it is typically infeasible to have many annota-tors for each pair. Furthermore, there is no guarantee thatoutliers, particularly those caused by unintentional hu-man errors can be dealt with effectively. This is becausemajority voting is a local consistency detection basedstrategy – when there are contradictory/inconsistentpairwise rankings for a given pair, the pairwise rankingsreceiving minority votes are eliminated as outliers. How-ever, it has been found that when pairwise local rankingsare integrated into a global ranking, it is possible todetect outliers that can cause global inconsistency andyet are locally consistent, i.e. supported by majority votes[19]. Critically, outliers that cause global inconsistencyhave more significant detrimental effects on learning aranking function for SVP prediction and thus should bethe main focus of an outlier detection method.

In this paper we propose a novel approach to sub-jective visual property prediction from sparse and noisypairwise comparison labels collected using crowdsourc-ing tools. Different from existing approaches which firstremove outliers by majority voting, followed by regres-sion [4] or learning to rank [5], we formulate a unified ro-bust learning to rank (URLR) framework to solve jointlyboth the outlier detection and learning to rank problems.Critically, instead of detecting outliers locally and inde-pendently at each pair by majority voting, our outlierdetection method operates globally, integrating all localpairwise comparisons together to minimise a cost thatcorresponds to global inconsistency of ranking order.This enables us to identify those outliers that receivemajority votes but cause large global ranking inconsis-tency and thus should be removed. Furthermore, as aglobal method that aggregates comparisons across dif-ferent pairs, our method can operate with as few as onecomparison per pair, making our method much morerobust against the data sparsity problem compared to theconventional majority voting approach that aggregatescomparisons for each pair in isolation. More specifically,the proposed model generalises a partially penalisedLASSO optimisation or Huber-LASSO formulation [20],

Page 4: Edinburgh Research Explorer€¦ · abide by the legal requirements associated with these rights. Take down policy The University of Edinburgh has made every reasonable effort to

3

[21], [22] from a robust statistical ranking formulationto a robust learning to rank model, making it suitablefor SVP prediction given unseen images/videos. We alsoformulate a regularisation path based solution to solvethis new formulation efficiently. Extensive experimentsare carried out on benchmark datasets including twoimage and video interestingness datasets [4], [5] and tworelative attribute datasets [2]. The results demonstratethat our method significantly outperforms the state-of-the-art alternatives.

2 RELATED WORK

Subjective visual properties Subjective visual prop-erty prediction covers a large variety of computer visionproblems; it is thus beyond the scope of this paper topresent an exhaustive review here. Instead we focusmainly on the image/video interestingness predictionproblem which share many characteristics with otherSVP prediction problem such as image quality [23],memorability [24], and aesthetics [3] prediction.Predicting image and video interestingness Early ef-forts on image interestingness prediction focus on dif-ferent aspects than interestingness as such, includingmemorability [24] and aesthetics [3]. These SVPs arerelated to interestingness but different. For instance, it isfound that memorability can have a low correlation withinterestingness - people often remember things that theyfind uninteresting [4]. The work of Gygli et al [4] is thefirst systematic study of image interestingness. It showsthat three cues contribute the most to interestingness:aesthetics, unusualness/novelty and general preferences,the last of which refers to the fact that people in generalfind certain types of scenes more interesting than oth-ers, for example outdoor-natural vs. indoor-manmade.Different features are then designed to represent thesecues as input to a prediction model. In comparison, videointerestingness has received much less attention, perhapsbecause it is even harder to understand its meaning andcontributing cues. Liu et al. [25] focus on key frames soessentially treats it as an image interestingness problem,whilst [5] is the first work that proposes benchmarkvideo interestingness datasets and evaluates differentfeatures for video interestingness prediction.

Most earlier works cast the aesthetics or interesting-ness prediction problem as a regression problem [23], [3],[24], [25]. However, as discussed before, obtaining an ab-solute value of interestingness for each data point is toosubjective and affected too much by unknown personalpreference/social background to be reliable. Thereforethe most recent two studies on image [4] and video[5] interestingness all collect pairwise comparison databy crowdsourcing. Both use majority voting to removeoutliers first. After that the prediction models differ– [4] converts pairwise comparisons into an absoluteinterestingness values and use a regression model, whilst[5] employs rankSVM [10] to learn a ranking function,with the estimated ranking score of an unseen videoused as the interestingness prediction. We compare with

both approaches in our experiments and demonstratethat our unified robust learning to rank approach issuperior as we can remove outliers more effectively –even if they correspond to comparisons receiving major-ity votes, thanks to its global formulation.

Relative attributes In a broader sense interestingnesscan be considered as one type of relative attribute [6].Attribute-based modelling [26], [27] has gained popu-larity recently as a way to describe instances and classesat an intermediate level of representation. Attributes arethen used for various tasks including N-shot and zero-shot transfer learning. Most previous studies considerbinary attributes [26], [27]. Relative attributes [6] wererecently proposed to learn a ranking function to predictrelative semantic strength of visual attributes. Instead ofthe original class-level attribute comparisons in [6], thispaper focuses on instance-level comparisons due to thehuge intra-class variations in real-world problems. Withinstance-level pairwise comparisons, relative attributeshave been used for interactive image search [2], andsemi-supervised [28] or active learning [29], [30] of visualcategories. However, no previous work addresses theproblem of annotation outliers except [2], which adoptsthe heuristic majority voting strategy.

Learning from noisy paired crowdsourced data Manylarge-scale computer vision problems rely on humanintelligence tasks (HIT) using crowdsourcing services,e.g. AMT (Amazon Mechanical Turk) to collect an-notations. Many studies [14], [31], [32], [13] highlightthe necessity of validating the random or maliciouslabels/workers and give filtering heuristics for datacleaning. However, these are primarily based on majorityvoting which requires a costly volume of redundantannotations, and has no theoretical guarantee of solvingthe outlier and sparsity problems. As a local (per-pair)filtering method, majority voting does not respect globalordering and even risks introducing additional incon-sistency due to the well-known Condorcet’s paradox insocial choice and voting theory [33]. Active learning [34],[29], [30] is an another way to circumvent the O(n2) pair-wise labelling space. It actively poses specific requeststo annotators and learns from their feedback, ratherthan the ‘general’ pairwise comparisons discussed in thiswork. Besides paired crowdsourced data, majority vot-ing is more widely used in crowdsourcing where mul-tiple annotators directly label instances, which attractedlots of attention in the machine learning community [16],[17], [18], [15]. In contrast, our work focuses on pairwisecomparisons which are relatively easier for annotators inevaluating the subjective visual properties [8] .

Statistical ranking and learning to rank Statistical rank-ing has been widely studied in statistics and computerscience [35], [36], [8], [37]. However, statistical rankingonly concerns the ranking of the observed/training data,but not learning to predict unseen data by learningranking functions. To learn ranking functions for ap-plications such as interestingness prediction, a feature

Page 5: Edinburgh Research Explorer€¦ · abide by the legal requirements associated with these rights. Take down policy The University of Edinburgh has made every reasonable effort to

4

representation of the data points must be used as modelinput in addition to the local ranking orders. This isaddressed in learning to rank which is widely studiedin machine learning [38], [39], [40]. However, existinglearning to rank works do not explicitly model andremove outliers for robust learning: a critical issue forlearning from crowdsourced data in practice. In thiswork, for the first time, we study the problem of ro-bust learning to rank given extremely noisy and sparsecrowdsourced pairwise labels. We show both theoreti-cally and experimentally that by solving both the outlierdetection and ranking prediction problems jointly, weachieve better outlier detection than existing statisticalranking methods and better ranking prediction thanexisting learning to rank method such as RankSVMwithout outlier detection.Our contributions are threefold: (1) We propose a novelrobust learning to rank method for subjective visualproperty prediction using noisy and sparse pairwisecomparison/ranking labels as training data. (2) For thefirst time, the problems of detecting outliers and estimat-ing linear ranking models are solved jointly in a unifiedframework. (3) We demonstrate both theoretically andexperimentally that our method is superior to existingmajority voting based methods as well as statistical rank-ing based methods. An earlier and preliminary versionof this work is presented in [41] which focused only onthe image/video interestingness prediction problem.

3 UNIFIED ROBUST LEARNING TO RANK

3.1 Problem definitionWe aim to learn a subjective visual property (SVP)prediction model from a set of sparse and noisy pairwisecomparison labels, each comparison corresponding toa local ranking between a pair of images or videos.Suppose our training set has N data points/instances

represented by a feature matrix Φ =[φTi

]Ni=1∈ RN×d,

where φi is a d-dimensional column low-level featurevector representing instance i. The pairwise comparisonlabels (annotations collected using crowdsourcing tools)can be naturally represented as a directed comparisongraph G = (V,E) with a node set V = {i}Ni=1 corre-sponding to the N instances and an edge set E = {eij}corresponding to the pairwise comparisons.

The pairwise comparison labels can be provided bymultiple annotators. They are dichotomously saved:Suppose annotator α gives a pairwise comparison forinstance i and j (i, j ∈ V ). If α considers that the SVPof instance i is stronger/more than that of j, we save(i, j, yαeij ) and set yαeij = 1. If the opposite is the case,we save (j, i, yαeji) and set yαeji = 1. All the pairwisecomparisons between instances i and j are then aggre-gated over all annotators who have cast a vote on thispair; the results are represented as weij =

∑αJyαeij = 1K

which is the total number of votes on i over j for aspecific SVP, where JK indicates the Iverson’s bracketnotation, and weji which is defined similarly. This gives

an edge weight vector w =[weij

]∈ R|E| where |E| is the

number of edges. Now the edge set can be representedas E =

{eij |weij > 0

}and weij ∈ R is the weight for

the edge eij . In other words, an edge eij : i→ j exists ifweij > 0. The topology of the graph is denoted by a flagindicator vector y =

[yeij]∈ R|E| where each indicator

yeij = 1 indicates that there is an edge between instancesi to j regardless how many votes it carries. Note that allthe elements in y have the value 1, and their index eijgives the corresponding nodes in the graph.

Given the training data consisting of the feature matrixΦ and the annotation graph G, there are two tasks:

1) Detecting and removing the outliers in the edge setE of G. To this end, we introduce a set of unknownvariables γ =

[γeij

]∈ R|E| where each variable

γeij indicates whether the edge eij is an outlier.The outlier detection problem thus becomes theproblem of estimating γ.

2) Estimating a prediction function for SVP. In thiswork a linear model is considered due to its lowcomputational complexity, that is, given the low-level feature φx of a test instance x we use a linearfunction f(x) = βTφx to predict its SVP, whereβ is the coefficient weight vector of the low-levelfeature φx. Note that all formulations can be easilyupdated to use a non-linear function.

So far in the introduced notations three vectors shareindices: the flag indicator vector y, the outlier variablevector γ and the edge weight vector w. For notationconvenience, from now on we use yij , γij and wij toreplace yeij , γeij and weij respectively. As in most graphbased model formulations, we define C ∈ R|E|×N as theincident matrix of the directed graph G, where Ceiji =−1/1 if the edge eij enters/leaves vertex i.

Note that in an ideal case, one hopes that the votesreceived on each pair are unanimous, e.g. wij > 0 andwji = 0; but often there are disagreements, i.e. we haveboth wij > 0 and wji > 0. Assuming both cannot be truesimultaneously, one of them will be an outlier. In thiscase, one is the majority and the other minority whichwill be pruned by the majority voting method. This iswhy majority voting is a local outlier detection methodand requires as many votes per pair as possible to beeffective (the wisdom of a crowd).

3.2 Framework formulation

In contrast to majority voting, we propose to pruneoutliers globally and jointly with learning the SVP pre-diction function. To this end, the outlier variables γij foroutlier detection and the coefficient weight vector β forSVP prediction are estimated in a unified framework.Specifically, for each edge eij ∈ E, its corresponding flagindicator yij is modelled as

yij = βTφi − βTφj + γij + εij , (1)

where εij ∼ N (0, σ2) is the Gaussian noise with zeromean and a variance σ, and the outlier variable γij ∈ R

Page 6: Edinburgh Research Explorer€¦ · abide by the legal requirements associated with these rights. Take down policy The University of Edinburgh has made every reasonable effort to

5

is assumed to have a higher magnitude than σ. For anedge eij , if yij is not an outlier, we expect βTφi −β

Tφjshould be approximately equal to yij , therefore we haveγij = 0. On the contrary, when the prediction of βTφi −βTφj differs greatly from yij , we can explain yij as anoutlier and compensate for the discrepancy between theprediction and the annotation with a nonzero value ofγij . The only prior knowledge we have on γij is that itis a sparse variable, i.e. in most cases γij = 0.

For the whole training set, Eq (1) can be re-written inits matrix form

y = CΦβ + γ + ε (2)

where y = [yij ] ∈ R|E|, γ = [γij ] ∈ R|E|, ε = [εij ] ∈ R|E|and C ∈ R|E|×N is the incident matrix of the annotationgraph G.

In order to estimate the |E|+ d unknown parameters(|E| for γ and d for β), we aim to minimise the dis-crepancy between the annotation y and our predictionCΦβ + γ, as well as keeping the outlier estimation γsparse. Note that y only contains information aboutwhich pairs of instances have received votes, but nothow many. The discrepancy thus needs to weighted bythe number of votes received, represented by the edgeweight vector w = [wij ] ∈ R|E|. To that end, we puta weighted l2−loss on the discrepancy and a sparsityenhancing penalty on the outlier variables. This gives usthe following cost function:

L(β,γ) =1

2‖y − CΦβ − γ‖22,w + pλ(γ) (3)

where

‖y−CΦβ−γ‖22,w =∑eij∈E

wij(yij − γij −βTφi +βTφj)2,

and pλ(γ) is the sparsity constraint on γ. With this costfunction, our Unified Robust Learning to Rank (URLR)framework identifies outliers globally by integrating alllocal pairwise comparisons together. Note that in Eq(3), the noise term ε has been removed because thediscrepancy is mainly caused by outliers due to theirlarger magnitude.

Ideally the sparsity enhancing penalty term pλ(γ)should be a l0 regularisation term. However, for atractable solution, a l1 regularisation term is used:pλ(γ) = λ‖γ‖1,w = λ

∑eijwij |γij |, where λ is a free

parameter corresponding to the weight for the regulari-sation term. With this l1 penalty term, the cost functionbecomes convex:

L(β,γ) =1

2‖√W (y − γ)−Xβ‖22 + λ‖γ‖1,w, (4)

where X =√WCΦ, W = diag(w) is the diagonal matrix

of w and√W = diag(

√w).

Setting ∂L∂β = 0, the problem of minimisation of the cost

function in (4) can be decomposed into the following twosubproblems:

1) Estimating the parameters β of the prediction func-tion f(x):

β = (XTX)†XT√W (y − γ), (5)

Mathematically, the Moore-Penrose pseudo-inverse of XTX is defined as (XTX)† =limµ→0

((XTX)T (XTX) + µI)−1(XTX)T , where

I is the identity matrix. The scalar variable µ isintroduced to avoid numerical instability [42], andtypically assumes a small value2. With the theintroduction of µ, Eq (5) becomes:

β = (XTX + µI)−1XT√W (y − γ). (6)

A standard solver for Eq (6) has a O(|E|d2) com-putational complexity, which is almost linear withrespect to the size of the graph |E| if d � n.Faster algorithms based on the Krylov iterative andalgebraic multi-grid methods [43] can also be used.

2) Outlier detection:

γ = arg minγ12‖(I −H)

√W (y − γ)‖22 + λ‖γ‖1,w(7)

= arg minγ12‖y − Xγ)‖22 + λ‖γ‖1,w (8)

where H = X(XTX)†XT is the hat matrix, X =(I − H)

√W and y = Xy. Eq (7) is obtained by

plugging the solution β back into Eq (4).

3.3 Outlier detection by regularisation path

From the formulations described above, it is clear thatoutlier detection by solving Eq (8) is the key – once theoutliers are identified, the estimated γ can be used tosubstitute γ in Eq (5) and the estimation of the predictionfunction parameter β becomes straightforward. Now letus focus on solving Eq (8) for outlier detection.

Note that solving Eq (8) is essentially a LASSO (LeastAbsolute Shrinkage and Selection Operator) [20] prob-lem. For a LASSO problem, tuning the regularisationparameter λ is notoriously difficult [44], [45], [46], [47]. Inparticular, in our URLR framework, the λ value directlydecides the ratio of outliers in the training set which isunknown. A number of methods for determining λ exist,but none is suitable for our formulation:

1) Some heuristics rules on setting the value of λ suchas λ = 2.5σ are popular in existing robust rankingmodels such as the M-estimator [44], where σ is aGaussian variance set manually based on humanprior knowledge. However setting a constant λvalue independent of dataset is far from optimalbecause the ratio of outliers may vary for differentcrowdsourced datasets.

2) Cross validation is also not applicable here becauseeach edge eij is associated with a γij variableand any held-out edge eij also has an associatedunknown variable γij . As a result, cross validationcan only optimise part of the sparse variables while

2. In this work, µ is set to 0.001.

Page 7: Edinburgh Research Explorer€¦ · abide by the legal requirements associated with these rights. Take down policy The University of Edinburgh has made every reasonable effort to

6

leaving those for the held-out validation set unde-termined.

3) Data adaptive techniques such as Scaled LASSO[45] and Square-Root LASSO [46] typically generateover-estimates on the support set of outliers. More-over, they rely on the homogeneous Gaussian noiseassumption which is often not valid in practice.

4) The other alternatives e.g. Akaike information cri-terion (AIC) and Bayesian information criterion(BIC) are often unstable in outlier detection LASSOproblems [47]3.

This inspires us to sequentially consider all availablesolutions for all sparse variables along the RegularisationPath (RP) by gradually decreasing the value of theregularisation parameter λ from ∞ to 0. Specifically,based on the piecewise-linearity property of LASSO, aregularisation path can be efficiently computed by the R-package “glmnet” [48]4. When λ =∞, the regularisationparameter will strongly penalise outlier detection: if anyannotation is taken as an outlier, it will greatly increasethe value of the cost function in Eq (8). When λ ischanged from ∞ to 0, LASSO5 will first select the vari-able subset accounting for the highest deviations to theobservations X in Eq (8). These high deviations shouldbe assigned higher priority to represent the nonzeroelements6 of γ of Eq (2), because γ compensates thediscrepancy between annotation and prediction. Basedon this idea, we can order the edge set E according towhich nonzero γij appears first when λ is decreasedfrom ∞ to 0. In other words, if an edge eij whoseassociated outlier variable γij becomes nonzero at alarger λ value, it has a higher probability to be an outlier.Following this order, we identify the top p% edge set Λpas the annotation outliers. And its complementary setΛ1−p = E \ Λp are the inliers. Therefore, the outcome ofestimating γ using Eq (8) is a binary outlier indicatorvector f =

[feij]:

feij =

{1 eij ∈ Λ1−p

0 eij ∈ Λp

where each element feij indicates whether the corre-sponding edge eij is an outlier or not.

Now with the outlier indicator vector f estimatedusing regularisation path, instead of estimating β bysubstituting γ in Eq (5) with an estimated γ, β can becomputed as

β = (XTFX + µI)−1XT√WFy (9)

3. We found empirically that the model automatically selected byBIC or AIC failed to detect any meaningful outliers in our experi-ments. For details of the experiments and a discussion on the issueof determining the outlier ratio, please visit the project webpage athttp://www.eecs.qmul.ac.uk/∼yf300/ranking/index.html

4. http://cran.r-project.org/web/packages/glmnet/glmnet.pdf5. For a thorough discussion from a statistical perspective, please see

[49], [50], [51], [47].6. This is related with LASSO for covariate selection in a graph.

Please see [52] for more details.

Algorithm 1 Learning a unified robust learning to rank(URLR) model for SVP predictionInput: A training dataset consisting of the feature matrixΦ and the pairwise annotation graph G, and an outlierpruning rate p%.Output: Detected outliers f and prediction model pa-rameter β.

1) Solve Eq (8) using Regularisation Path;2) Take the top p% pairs as outliers to obtain the

outlier indicator vector f ;3) Compute β using Eq (9).

where F = diag(f), that is, we use f to ‘clean up’ ybefore estimating β.

The pseudo-code of learning our URLR model is sum-marised in Algorithm 1.

3.4 Discussions3.4.1 Advantage over majority votingThe proposed URLR framework identifies outliers glob-ally by integrating all local pairwise comparisons to-gether, in contrast to the local aggregation based majorityvoting. Figure 2(a) illustrates why our URLR frameworkis advantageous over the local majority voting methodfor outlier detection. Assume there are five images A−Ewith five pairs of them compared three times each, andthe correct global ranking order of these 5 images interms of a specific SVP is A < B < C < D < E.Figure 2(a) shows that among the five compared pairs,majority voting can successfully identify four outliercases: A > B, B > C, C > D, and D > E, but notthe fifth one E < A. However when considered globally,it is clear that E < A is an outlier because if we haveA < B < C < D < E, we can deduce A < E. Ourformulation can detect this tricky outlier. More specifi-cally, if the estimated β makes βTφA−β

TφE > 0, it has asmall local inconsistency cost for that minority vote edgeA → E. However, such β value will be ‘propagated’ toother images by using the voting edges B → A, C → B,D → C, and E → D, which are accumulated into amuch bigger global inconsistency with the annotation.This enables our model to detect E → A as an outlier,contrary to the majority voting decision. In particular,the majority voting will introduce a loop comparisonA < B < C < D < E < A which is the well-knownCondorcet’s paradox [33], [19].

We further give two more extreme cases in Figures 2(b)and (c). Due to the Condorcet’s paradox, in Figure 2(b)the estimated β from majority voting, which removesA → E, is even worse than that from all annotationpairs which at least save the correct annotation A→ E.Furthermore, Figure 2(c) shows that when each pair onlyreceives votes in one direction, majority voting will ceaseto work altogether, but our URLR can still detect outliersby examining the global cost. This example thus high-lights the capability of URLR in coping with extremely

Page 8: Edinburgh Research Explorer€¦ · abide by the legal requirements associated with these rights. Take down policy The University of Edinburgh has made every reasonable effort to

7

2

1 2

1

21

2

1

2 1

D

E

CB

A

(a)

2

1 2

2

2

2

D

E

CB

A

(b)

2

2

1

2

2

D

E

CB

A

(c)

Figure 2. Better outlier detection can be achieved usingour URLR framework than majority voting. Green ar-rows/edges indicate correct annotations, while red arrowsare outliers. The numbers indicate the number of votesreceived by each edge.

sparse pairwise comparison labels. In our experiments(see Section 4), the advantage of URLR over majority isvalidated on various SVP prediction problems.

3.4.2 Advantage over robust statistical rankingOur framework is closely related to Huber’s theory ofrobust regression [44], which has been used for robuststatistical ranking [53]. In contrast to learning to rank,robust statistical ranking is only concerned with rankinga set of training instances by integrating their (noisy)pairwise rankings. No low-level feature representationof the instances is used as robust ranking does not aimto learn a ranking prediction function that can be appliedto unseen test data. To see the connection between URLRwith robust ranking, consider the Huber M-estimator[44] which aims to estimate the optimal global rankingfor a set of training instances by minimising the follow-ing cost function:

minθ

∑i,j

wijρλ((θi − θj)− yij) (10)

where θ = [θi] ∈ R|E| is the ranking score vector storingthe global ranking score of each training instance i. TheHuber’s loss function ρλ(x) is defined as

ρλ(x) =

{x2/2, if |x| ≤ λλ|x| − λ2/2, if |x| > λ. (11)

Using this loss function, when |(θi − θj) − yij | < λ, thecomparison is taken as a “good” one and penalised byan l2−loss for Gaussian noise. Otherwise, it is regardedas a sparse outlier and penalised by an l1−loss. It canbe shown [53] that robust ranking with Huber’s loss isequivalent to a LASSO problem, which can been appliedto joint robust ranking and outlier detection [47]. Specif-ically, the global ranking of the training instances andthe outliers in the pairwise rankings can be estimated as{

θ, γ}

= minθ,γ

1

2‖y − Cθ − γ‖22,w + λ ‖ γ ‖1,w (12)

= minθ,γ

∑eij∈E

wij

[‖ 1

2(yij − γij − (θi − θj) ‖2 +λ|γij |

](13)

The optimisation problem (12) is designed for solvingthe robust ranking problem with Huber’s loss function,hence called Huber-LASSO [53].

Our URLR can be considered as a generalisation ofthe Huber-LASSO based robust ranking problem above.Comparing Eq (12) with Eq (3), it can be seen thatthe main difference between URLR and conventionalrobust ranking is that in URLR the cost function has thelow-level feature matrix Φ computed from the traininginstances, and the prediction function parameter β, suchthat θ = Φβ. This is because the objective of URLR isto predict SVP for unseen test data. However, URLRand robust ranking do share one thing in common –the ability to detect outliers in the training data basedon a Huber-LASSO formulation. This means that, asopposed to our unified framework with feature Φ, onecould design a two-step approach for learning to rankby first identifying and removing outliers using Eq (12),followed by introducing the low-level feature matrix Φand prediction model parameter β and estimating βusing Eq (9). We call this approach Huber-LASSO-FLbased learning to rank which differs from URLR mainlyin the way outliers are detected without considering lowlevel features.

Next we show that there is a critical theoretical ad-vantage of URLR over conventional Huber-LASSO indetecting outliers from the training instances. This is dueto the difference in the projection space for estimatingγ which is denoted as Γ. To explain this point, wedecompose X in Eq (8) by Singular Value Decomposition(SVD),

X = UΣVT (14)

where U = [U1,U2] with U1 being an orthogonal basisof the column space of X and U2 an orthogonal basisof its complement. Therefore, due to the orthogonalityUTU = I and UT2 X = 0, we can simplify Eq (8) into

γ = arg minγ‖UT2 y − UT2 γ‖22,w + λ‖γ‖1,w. (15)

The SVD orthogonally projects y onto the columnspace of X and its complement, while U1 is an orthogo-nal basis of the column space X and U2 is the orthogonalbasis of its complement Γ (i.e. the kernel space of XT ).With the SVD, we can now compute the outliers γby solving Eq (15) which again is a LASSO problem[42], where outliers provide sparse approximations ofprojection UT2 y. We can thus compare dimensions of theprojection spaces of URLR and Huber-LASSO-FL:• Robust ranking based on the featureless Huber-

LASSO-FL7: to see the dimension of the projectionspace Γ, i.e. the space of cyclic rankings [19], [53],we can perform a similar SVD operation and rewriteEq (12) in the same form as Eq (15), but this timewe have X =

√WC, U1 ∈ R|E|×(|V |−1) and U2 ∈

R|E|×(|E|−|V |+1). So the dimension of Γ for Huber-LASSO-FL is dim(Γ) = |E| − |V |+ 1.

• URLR: in contrast we have X =√WCΦ, U1 ∈

R|E|×d and U2 ∈ R|E|×(|E|−d). So the dimension of Γfor URLR is dim(Γ) = |E| − d.

7. We assume that the graph is connected, that is, |E| ≥ |V | − 1; wethus have rank(C) = |V | − 1.

Page 9: Edinburgh Research Explorer€¦ · abide by the legal requirements associated with these rights. Take down policy The University of Edinburgh has made every reasonable effort to

8

From the above analysis we can see that given a verysparse graph with |E| ∼ |V |, the projection space Γ forHuber-LASSO-FL will have a dimension (|E| − |V | + 1)too small to be effective for detecting outliers. In contrast,by exploiting a low dimensional (d � |V |) featurerepresentation of the original node space, URLR canenlarge the projection space to that of dimension |E|−d.Our URLR is thus able to enlarges its outlier detectionprojection space Γ. As a result our URLR can betteridentify outliers, especially for sparse pairwise anno-tation graphs. In general, this advantage exists whenthe feature dimension d is smaller than the number oftraining instance |V | = N , and the smaller the valueof d, the bigger the advantage over Huber-LASSO. Inpractice, given a large training set we typically haved � |V |. On the other hand, when the number ofinstances is small, and each instance is represented by ahigh-dimensional feature vector, we can always reducethe feature dimension using techniques such as PCAto make sure that d � |V |. This theoretical advantageof URLR over conventional Huber-LASSO in outlierdetection is validated experimentally in Section 4.

3.4.3 Regularisation on β

It is worth mentioning that in the cost function of URLR(Eq (3)), there are two sets of variables to be estimated,γ and β, but only one l1 regularisation term on γ toenforce sparsity. When the dimensionality of β (i.e. d) ishigh, one would expect to see a l2 regularisation termon β (e.g. ridge regression) due to the fact that thecoefficients of highly correlated low-level features canbe poorly estimated and exhibit high variance withoutimposing a proper size constraint on the coefficients [42].The reason we do not include such a regularisation termis because, as mentioned above, using URLR we need tomake sure the low-level feature space dimensionality dis low, which means that the dimensionality of β is alsolow, making the regularisation term β redundant. Thisleads to the applicability of much simpler solvers andwe show empirically in the next section that satisfactoryresults can be obtained with this simplification.

4 EXPERIMENTS

Experiments were carried out on five benchmarkdatasets (see Table 1) which fall into three categories: (1)experiments on estimating subjective visual properties(SVPs) that are useful on their own including image(Section 4.1) and video interestingness (Section 4.2), (2)experiments on estimating SVPs as relative attributesfor visual recognition (Section 4.3), and (3) experimentson human age estimation from face images (Section4.4). The third set of experiments can be considered assynthetic experiments – human age is not a subjectivevisual property although it is ambiguous and poses aproblem even for humans [56]. However, as ground truthis available, this set of experiments are designed to gaininsights into how different SVP prediction models work.

4.1 Image interestingness prediction

Datasets The image interestingness dataset was firstintroduced in [24] for studying memorability. It was laterre-annotated as an image interestingness dataset by [4].It consists of 2222 images. Each was represented as a915 dimensional attribute8 feature vector [24], [4] such ascentral object, unusual scene and so on. 16000 pairwisecomparisons were collected by [4] using AMT and usedas annotation. On average, each image is viewed andcompared with 11.9 other images, resulting a total of16000 pairwise labels9.Settings 1000 images were randomly selected fortraining and the remaining 1222 for testing. All the ex-periments were repeated 10 times with different randomtraining/test splits to reduce variance. The pruning ratep was set to 20%. We also varied the number of annotatedpairs used to test how well each compared method copeswith increasing annotation sparsity.Evaluation metrics For both image and video inter-estingness prediction, Kendall tau rank distance wasemployed to measure the percentage of pairwise mis-matches between the predicted ranking order for eachpair of test data using their prediction/ranking functionscores, and the ground truth ranking provided by [4] and[5] respectively. Larger Kendall tau rank distance meanslower quality of the ranking order predicted.Competitors We compare our method (URLR) withfour competitors.

1) Maj-Vot-1 [5]: this method uses majority voting foroutlier pruning and rankSVM for learning to rank.

2) Maj-Vot-2 [4]: this method also first removes out-liers by majority voting. After that, the fraction ofselections by the pairwise comparisons for eachdata point is used as an absolute interestingnessscore and a regression model is then learned forprediction. Note that Maj-Vot-2 was only comparedin the experiments on image and video interesting-ness prediction, since only these two datasets haveenough dense annotations for Maj-Vot-2.

3) Huber-LASSO-FL: robust statistical ranking thatperforms outlier detection using the conventionalfeatureless Huber-LASSO as described in Section3.4.2, followed by estimating β using Eq (9).

4) Raw: our URLR model without outlier detection,that is, all annotations are used to estimate β.

Comparative results The interestingness predictionperformance of the various models are evaluated whilevarying the amount of pairwise annotation used. Theresults are shown in Figure 3 (left). It shows clearly thatour URLR significantly outperforms the four alternativesfor a wide range of annotation density. This validatesthe effectiveness of our method. In particular, it can

8. We delete 8 attribute features from the original feature vector in[24], [4] such as “attractive” because they are highly correlated withimage interestingness.

9. On average, for each labelled pair, around 80% of the annotationsagree with one ranking order and 20% the other.

Page 10: Edinburgh Research Explorer€¦ · abide by the legal requirements associated with these rights. Take down policy The University of Edinburgh has made every reasonable effort to

9

Dataset No. pairs No. img/video Feature Dim. No. classesImage Interestingness [24] 16000 2222 932 (150) 1Video Interestingness [5] 60000 420 1000 (60) 14

PubFig [54], [2] 2616 772 557 (100) 8Scene [55], [2] 1378 2688 512 (100) 8

FG-Net Face Age Dataset [56] – 1002 55 –

Table 1Dataset summary. We use the original features to learn the ranking model (Eq (9)) and reduce the feature dimension(values in brackets) using Kernel PCA [57] to improve outlier detection (Eq (8)) by enlarging the projection space of γ.

30 40 50 60 70 80 90 100 1100.11

0.12

0.13

0.14

0.15

0.16

0.17

Percentage of total available pairs (%)

Ken

dal

l tau

dis

tan

ce

Image Interestingness

Raw

Maj−Vot−1

Maj−Vot−2

Huber−LASSO−FL

URLR

−20 0 20 40 60 800.11

0.12

0.13

0.14

0.15

Pruning rate(%)

Ken

dal

l tau

dis

tan

ce

Pruning rate using all pairs

Huber−LASSO

URLR

Figure 3. Image interestingness prediction comparative evaluation. Smaller Kendall tau distance means betterperformance. The mean and standard deviation of each method over 10 trials are shown in the plots.

Sucess cases Failure cases

Figure 4. Qualitative examples of outliers detected by URLR. In each box, there are two images. The left image wasannotated as more interesting than the right. Success cases (green boxes) show true positive outliers detected byURLR (i.e. right images are more interesting according to the ground truth). Two failure cases are shown in red boxes(URLR thinks the images on the right are more interesting but the ground truth agrees with the annotation).

be observed that: (1) The improvement over Maj-Vot-1[5] and Maj-Vot-2 [4] demonstrates the superior outlierdetection ability of URLR due to global rather than localoutlier detection. (2) URLR is superior to Huber-LASSO-FL because the joint outlier detection and ranking pre-diction framework of URLR enables the enlargement ofthe projection space Γ for γ (see Section 3.4.2), resultingin better outlier detection performance. (3) The perfor-mance of Maj-Vot-2 [4] is the worst among all methodscompared, particularly so given sparser annotation. Thisis not surprising – in order to get an reliable absoluteinterestingness value, dozens or even hundreds of com-parisons per image are required, a condition not met bythis dataset. (4) The performance of Huber-LASSO-FL isalso better than Maj-Vot-1 and Maj-Vot-2 suggesting evena weaker global outlier detection approach is better thenthe majority voting based local one. (5) Interestingly eventhe baseline method Raw gives a comparable result toMaj-Vot-1 and Maj-Vot-2 which suggests that just usingall annotations without discrimination in a global cost

function (Eq (4)) is as effective as majority voting10.Figure 3 (right) evaluates how the performances of

URLR and Huber-LASSO-FL are affected by the pruningrate p. It can be seen that the performance of URLR isimproving with an increasing pruning rate. This meansthat our URLR can keep on detecting true positiveoutliers. The gap between URLR and Huber-LASSO-FLgets bigger when more comparisons are pruned showingHuber-LASSO-FL stops detecting outliers much earlieron. However, when the pruning rate is over 55%, sincemost outliers have been removed, inliers start to bepruned, leading to poorer performance.Qualitative Results Some examples of outlier detec-tion using URLR are shown in Figure 4. It can be seen

10. One intuitive explanation for this is that given a pair of data withmultiple contradictory votes, using Raw, both the correct and incorrectvotes contribute to the learned model. In contrast, with Maj-Vot, oneof them is eliminated, effectively amplifying the other’s contributionin comparison to Raw. When the ratio of outliers gets higher, Maj-Votwill make more mistakes in eliminating the correct votes. As a result,its performance drops to that of Raw, and eventually falls below it.

Page 11: Edinburgh Research Explorer€¦ · abide by the legal requirements associated with these rights. Take down policy The University of Edinburgh has made every reasonable effort to

10

that those in the green boxes are clearly outliers andare detected correctly by our URLR. The failure casesare interesting. For example, in the bottom case, groundtruth indicates that the woman sitting on a bench is moreinteresting than the nice beach image, whilst our URLRpredicts otherwise. The odd facial appearance on thatwoman or the fact that she is holding a camera couldbe the reason why this image is considered to be moreinteresting than the otherwise more visually appealingbeach image. However, it is unlikely that the featuresused by URLR are powerful enough to describe suchfine appearance details.

4.2 Video interestingness prediction

Datasets The video interestingness dataset is theYouTube interestingness dataset introduced in [5]. It con-tains 14 categories of advertisement videos (e.g. ‘food’and ‘digital products’), each of which has 30 videos.10 ∼ 15 annotators were asked to give complete inter-esting comparisons for all the videos in each category.So the original annotations are noisy but not sparse. Weused bag-of-words of Scale Invariant Feature Transform(SIFT) and Mel-Frequency Cepstral Coefficient (MFCC)as the feature representation which were shown to beeffective in [5] for predicting video interestingness.Experimental settings Because comparing videosacross different categories is not very meaningful, wefollowed the same settings as in [5] and only comparedthe interestingness of videos within the same category.Specifically, from each category we used 20 videos andtheir paired comparisons for training and the remaining10 videos for testing. The experiments were repeatedfor 10 rounds and the averaged results are reported.

Since MFCC and SIFT are bag-of-words features, weemployed χ2 kernel to compute and combine the fea-tures. To facilitate the computation, the χ2 kernel isapproximated by additive kernel of explicit feature map-ping [58]. To make the results of this dataset morecomparable to those in [5], we used rankSVM model toreplace Eq (9) as the ranking model. As in the imageinterestingness experiments, we used Kendal tau rankdistance as the evaluation metric, while we find that thesame results can be obtained if the prediction accuracyin [5] is used. The pruning rate was again set to 20%.Comparative Results Figure 5(a) compares the inter-estingness prediction methods given varying amounts ofannotation, and Figure 5(b) shows the per category per-formance. The results show that all the observations wehad for the image interestingness prediction experimentstill hold here, and across all categories. However in gen-eral the gaps between our URLR and the alternatives aresmaller as this dataset is densely annotated. In particularthe performance of Huber-LASSO-FL is much closer toour URLR now. This is because the advantage of URLRover Huber-LASSO-FL is stronger when |E| is close to|V |. In this experiment, |E| (1000s) is much greater than|V | (20) and the advantage of enlarging the projectionspace Γ for γ (see Section 3.4.2) diminishes.

Qualitative Results Some outlier detection examplesare shown in Figure 6. In the two successful detectionexamples, the bottom videos are clearly more interestingthan the top ones, because they (1) have a plot, some-times with a twist, and (2) are accompanied by popularsongs in the background and/or conversations. Notethat in both cases, majority voting would consider theminliners. The failure case is a hard one: both videos havecartoon characters, some plot, some conversation, andsimilar music in the background. This thus correspondsto a truly ambiguous case which can go either way.

4.3 Relative attributes prediction

Datasets The PubFig [54] and Scene [55] datasets aretwo relative attribute datasets. PubFig contains 772 im-ages from 8 people and 11 attributes (‘smiling’, ‘roundface’, etc.). Scene [55] consists of 2688 images from 8categories and 6 attributes (‘openness’, ‘natrual’ etc.).Pairwise attribute annotation was collected by AmazonMechanical Turk [2]. Each pair was labelled by 5 workersand majority vote was used in [2] to average the com-parisons for each pair11. A total of 241 and 240 trainingimages for PubFig and Scene respectively were labelled(i.e. compared with at least another image). The averagenumber of compared pairs per attribute were 418 and426 respectively, meaning most images were only com-pared with one or two other images. The annotationsfor both datasets were thus extremely sparse. GIST andcolour histogram features were used for PubFig, andGIST alone for Scene. Each image also belongs to a class(different celebrities or scene types). These datasets weredesigned for classification, with the predicted relativeattribute scores used as image representation.Experimental Settings We evaluated two different im-age classification tasks: multi-class classification wheresamples from all classes were available for training andzero-shot transfer learning where one class was held outduring training (a different class was used in each trialwith the result averaged). Our experiment setting wassimilar to that in [6], except that image-level, rather thanclass-level pairwise comparisons were used. Two settingswere used with different amounts of annotation noise:• Orig: This was the original setting with the pairwise

annotations used as they were.• Orig+synth: By visual inspection, there were lim-

ited annotation outliers in these datasets, perhapsbecause these relative attributes are less subjec-tive compared to interestingness. To simulate morechallenging situations, we added 150 random com-parisons for each attribute, many of which wouldcorrespond to outliers. This will lead to around 20%extra outliers.

The pruning rate was set to 7% for the original datasets(Orig) and 27% for the dataset with additional outliersinserted for all attributes of both datasets (Orig+synth).

11. Thanks to the authors of [2] we have all the the raw pairs databefore majority voting.

Page 12: Edinburgh Research Explorer€¦ · abide by the legal requirements associated with these rights. Take down policy The University of Edinburgh has made every reasonable effort to

11

0.15

0.2

0.25

Ken

dal

l tau

dis

tan

ce

fooddrin

k

clothingshoe

accessoriess

personalcare

houseapplication

housewarefurnitures

hygienicproduct

digitalproductsphone

computerwebsite

transporta

tion

medicine

(a) (b)

30 40 50 60 70 80 90 100 110

0.18

0.2

0.22

0.24

Percentage of total available pairs (%)K

end

all t

au d

ista

nce

Raw

Maj-Vot-1

Maj-Vot-2

Huber−LASSO-FL

URLR

Figure 5. Video interestingness prediction comparative evaluation.

Figure 6. Qualitative examples of video interestingness outlier detection. For each pair, the top video was annotatedas more interesting than the bottom. Green boxes indicate the annotations are correctly detected as outliers by ourURLR and red box indicates a failure case (false positive). All 6 videos are from the ‘food’ category.

Evaluation metrics For Scene and Pubfig datasets,relative attributes were very sparsely collected and theirprediction performance is thus evaluated indirectly byimage classification accuracy with the predicted relativeattributes as image representation. Note that for imageclassification there is ground truth and its accuracy isclearly dependent on the relative attribute predictionaccuracy. For both datasets, we employed the methodin [6] to compute the image classification accuracy.

Comparative Results Without the ground truth ofrelative attribute values, different models were evalu-ated indirectly via image classification accuracy in Fig-ure 7. The following observations can be made: (1) OurURLR always outperforms Huber-LASSO-FL, Maj-Vot-1and Raw for all experiment settings. The improvementis more significant when the data contain more errors(Orig+synth). (2) The performance of other methods is ingeneral consistent to what we observed in the image andvideo interestingness experiments: Huber-LASSO-FL isbetter than Maj-Vot-1 and Raw often gives better resultsthan majority voting. (3) For PubFig, Maj-Vot-1 [5] isbetter than Raw given more outliers, but it is not thecase for Scene. This is probably because the annotatorswere more familiar with the celebrity faces in PubFig andhence their attributes than those in Scene. Consequentlythere should be more subjective/intentional errors for

Scene, causing majority voting to choose wrong localranking orders (e.g. some people are unsure how to com-pare the relative values of the ‘diagonal plane’ attributefor two images). These majority voting + outlier casescan only be rectified by using a global approach such asour URLR, and Huber-LASSO-FL to a lesser extent.Qualitative Results Figure 8 gives some examples ofthe pruned pairs for both datasets using URLR. In thesuccess cases, the left images were (incorrectly) anno-tated to have more of the attribute than the right ones.However, they are either wrong or too ambiguous togive consistent answers, and as such are detrimental tolearning to rank. A number of failure cases (false positivepairs identified by URLR) are also shown. Some of themare caused by unique view point (e.g. Hugh Laurie’smouth is not visible, so it is hard to tell who smilesmore; the building and the street scene are too zoomedin compared to most other samples); others are causedby the weak feature representation, e.g. in the ‘male’attribute example, the colour and GIST features are notdiscriminative enough for judging which of the two menhas more ‘male’ attribute.Running Cost Our algorithm is very efficient with aunified framework where all outliers are pruned simulta-neously and the ranking function estimation has a closedform solution. Using URLR on PubFig, it took only 1

Page 13: Edinburgh Research Explorer€¦ · abide by the legal requirements associated with these rights. Take down policy The University of Edinburgh has made every reasonable effort to

12

Orig Orig+synth.0.4

0.45

0.5

PubFig: Multi-class learning via attribute representation

Acc

ura

cy

RawHuber-LASSO-FLURLRMaj-Vot-1

Orig Orig+synth.0.35

0.4

PubFig: Zero-shot transfer learning

Acc

ura

cy

Orig Orig+synth.0.6

0.62

0.64

0.66

0.68A

ccu

racy

Scene: Multi-class learning via attribute representation

Orig Orig+synth.0.55

0.56

0.57

0.58

0.59

0.6

0.61

Scene: Zero-shot transfer learning

Acc

ura

cy

Figure 7. Relative attribute performance evaluated indirectly as image classification rate (chance = 0.125).

Smiling Smiling

ChubbyYoung

Diagonal Plane Nature

Natural Size-large

Smiling

Male

Nature

Size-large

Failure casesSucess cases

Figure 8. Qualitative results on image relative attribute prediction.

minutes to prune 240 images with 10722 comparisonsand learn the ranking function for attribute predictionon a PC with four 3.3GHz CPU cores and 8GB memory.

4.4 Human age prediction from face imagesIn this experiment, we consider age as a subjective visualproperty of a face. This is partially true – for manypeople, given a face image predicting the person’s agecan be subjective. The key difference between this andthe other SVPs evaluated so far is that we do have theground truth, i.e. the person’s age when the picture wastaken. This enables us to perform in-depth evaluation ofthe significance of our URLR framework over the alter-natives on various factors such as annotation sparsity,and outlier ratio (we now know the exact ratio). Outlierdetection accuracy can also now be measured directly.Dataset The FG-NET image age dataset12 was em-ployed which contains 1002 images of 82 individualslabelled with ground truth ages ranging from 0 to 69.The training set is composed of the images of 41 ran-domly selected people and the rest used as the test set.All experiments were repeated 10 times with differenttraining/testing splits to reduce variability. Each imagewas represented by a 55 dimension vector extracted byactive appearance models (AAM) [56].Crowdsourcing errors We used the ground truth ageto generate the pairwise comparisons without any er-ror. Errors were then synthesised according to human

12. http://www.fgnet.rsunit.com/

error patterns estimated by data collected by an onlinepilot study13: 4000 pairwise image comparisons from 20willingly participating “good” workers were collected asunintentional errors. So we assume they are not contribut-ing random or malicious annotations. Thus the errors ofthese pairwise comparisons come from the natural dataambiguity. The human unintentional age error patternwas built by fitting the error rate against true age differ-ence between collected pairs. As expected, humans aremore error-prone for smaller age difference. Specifically,we fit quadratic polynomial function to model relationof age difference of two samples towards the chanceof making an unintentional error. We then used thiserror pattern to generate unintentional errors. Intentionalerrors were introduced by ‘bad’ workers who providedrandom pairwise labels. This was easily simulated byadding random comparisons. In practice, human errorsin crowdsourcing experiments can be a mixture of bothtypes. Thus two settings were considered: Unint.: errorswere generated following the estimated human unin-tentional error model resulting in around 10% errors.Unint.+Int.: random comparisons were added on topof Unint., giving an error ratio of around 25%, unlessotherwise stated. Since the ground-truth age of each faceimage is known to us, we can give an upper bound forall the compared methods by using ground-truth age oftraining data to generate a set of pairwise comparisons.

13. http://www.eecs.qmul.ac.uk/∼yf300/survey4/

Page 14: Edinburgh Research Explorer€¦ · abide by the legal requirements associated with these rights. Take down policy The University of Edinburgh has made every reasonable effort to

13

0 0.1 0.2 0.3 0.4 0.50.5

0.52

0.54

0.56

0.58

0.6

0.62

0.64

0.66

0.68

0.7

Pruning rate

Rank

Cor

rela

tion

Unint.

Huber-LASSO-FL

URLR

Raw

GT

0 0.1 0.2 0.3 0.4 0.50.5

0.52

0.54

0.56

0.58

0.6

0.62

0.64

0.66

0.68

0.7

Pruning rate

Rank

Cor

rela

tion

Unint.+Int.

Figure 9. Comparing URLR and Huber-LASSO-FL onranking prediction under two error settings. Note that theranking prediction accuracy is measured using Kendalltau rank correlation which is very similar to Kendall taudistance (see [59]). With rank correlation, the higher thevalue the better the performance.

0 0.1 0.2 0.3 0.4 0.50.5

0.55

0.6

0.65

0.7

Pruning rate

Ran

k C

orr

elat

ion

Huber-LASSO-FL

URLR

Raw

GT

Maj-Vot-1

Figure 10. Comparing URLR and Huber-LASSO-FLagainst majority voting (5 comparisons per pair).

This outlier-free dataset is then used to learn a kernelridge regression with Gaussian kernel. This ground-truthdata trained model is denoted as GT.Quantitative results Four experiments were conductedusing different settings to show the effectiveness of ourURLR method quantitatively.(1) URLR vs. Huber-LASSO-FL. In this experiment, 300training images and 600 unique comparisons were ran-domly sampled from the training set. Figure 9 shows thatURLR and Huber-LASSO-FL improve over Raw indicat-ing that outliers are effectively pruned using both globaloutlier detection methods. Both methods are robust tolow error rate (Figure 9 Left: 10% in Unint.) and arefairly close to GT, whilst the performance of URLR issignificantly better than that of Huber-LASSO-FL givenhigh error ratio (Figure 9 Right: 25% in Unint.+Int.)because of the using low-level feature representation toincrease the dimension of projection space dimension forγ from 301 for Huber-LASSO-FL to 546 for URLR (seeSection 3.4.2). This result again validates our analysisthat higher dim(Γ) leads to better chance of identifyingoutliers correctly. It is noted that in Figure 9(Right), given25% outliers, the result indeed peaks when p is around25; importantly, it stays flat when up to 50% of theannotations are pruned.(2) Comparison with Maj-Vot-1. Given the same data buteach pair compared by 5 workers (instead of 1) under theUnint.+Int. error condition, Figure 10 shows that Maj-Vot-1 beats Raw. This shows that for relative dense graph,majority voting is still a good strategy of removing

15 20 25 30 35 400.5

0.6

0.7

0.8

0.9

1

Error Ratio

AU

C v

alu

es

15 20 25 30 35 400.35

0.4

0.45

0.5

0.55

0.6

0.65

Error Ratio

Ran

k C

orr

elat

ion

Huber-LASSO-FL

URLRRaw

GT

Figure 11. Effect of error ratio. Left: outlier detectionperformance measured by area under ROC curve (AUC).Right: rank prediction performance measured by rankcorrelation.

0

5

10

15

20

25

30

35

40

45

λ

Ag

e d

iffe

ren

ce

λmaxλmin

Figure 12. Relationship between the pruning order andactual age difference for URLR.

some outliers and improves the prediction accuracy.However, URLR outperforms Maj-Vot-1 after the pruningrate passes 10%. This demonstrates that aggregating allpaired comparisons globally for outlier pruning is moreeffective than aggregating them locally for each edge asdone by majority voting.

(3) Effects of error ratio. We used the Unint.+Int. errormodel to vary the amount of random comparisonsand simulate different amounts of errors in 10 sam-pled graphs from 300 training images and 2000 uniquesampled pairs from the training images. The pruningrate was fixed at 25%. Figure 11 shows that URLRremains effective even when the true error ratio reachesas high as 35%. This demonstrates that although a sparseoutlier model is assumed, our model can deal with non-sparse outliers. It also shows that URLR consistentlyoutperforms the alternative models especially when theerror/outlier ratio is high.

What are pruned and in what order? The effectivenessof the employed regularisation path method for outlierdetection can be examined as λ decreases to producea ranked list for all pairwise comparisons according tothe outlier probability. Figure 12 shows the relationshipbetween the pruning order (i.e. which pair is prunedfirst) and ground truth age difference and illustrated byexamples. It can be seen that overall outliers with largerage difference tend to be pruned first. This means thateven with a conservative pruning rate, obvious outliers(potentially causing more performance degradation inlearning) can be reliably pruned by our model.

Page 15: Edinburgh Research Explorer€¦ · abide by the legal requirements associated with these rights. Take down policy The University of Edinburgh has made every reasonable effort to

14

5 CONCLUSIONS AND FUTURE WORK

We have proposed a novel unified robust learning torank (URLR) framework for predicting subjective visualproperties from images and videos. The key advantageof our method over the existing majority voting basedapproaches is that we can detect outliers globally byminimising a global ranking inconsistency cost. Thejoint outlier detection and feature based rank predictionformulation also provides our model with an advantageover the conventional robust ranking methods withoutfeatures for outlier detection: it can be applied with alarge number of candidates in comparison but a sparsesampling in crowdsourcing. The effectiveness of ourmodel in comparison with state-of-the-art alternativeshas been validated on the tasks of image and video inter-estingness prediction and predicting relative attributesfor visual recognition. Its effectiveness for outlier detec-tion has also been evaluated in depth in the human ageestimation experiments.

By definition subjective visual properties (SVPs) areperson-dependent. When our model is learned usingpairwise labels collected from many people, we areessentially learning consensus – given a new data pointthe model aims to predict its SVP value that can beagreed upon by most people. However, the predictedconsensual SVP value could be meaningless for a specificperson when his/her taste/understanding of the SVP iscompletely different to that of most others. How to learna person-specific SVP prediction model is thus part ofthe on-going work. Note that our model is only one ofthe possible solutions to inferring global ranking frompairwise comparisons. Other models exist. In particular,one widely studied alternative is the (Bradley-Terry-Luce (BTL) model [60], [61], [62]), which aggregates theranking scores of pairwise comparisons to infer a globalranking by maximum likelihood estimation. The BTLmodel is introduced to describe the probabilities of thepossible outcomes when individuals are judged againstone another in pairs [60]. It is primarily designed toincorporate contextual information in the global rank-ing model. We found that directly applying the BTLmodel to our SVP prediction task leads to much inferiorperformance because it does not explicitly detect andremove outliers. However, it is possible to integrate itinto our framework to make it more robust againstoutliers and sparse labels whilst preserving its abilityto take advantage of contextual information. Other newdirections include extending the presented work to otherapplications where noisy pairwise labels exist, both invision such as image denoising [63], iterative search andactive learning of visual categories [30], and in otherfields such as statistics and economics [19].

REFERENCES

[1] J. Donahue and K. Grauman, “Annotator rationales for visualrecognition,” in ICCV, 2011.

[2] A. Kovashka, D. Parikh, and K. Grauman, “Whittlesearch: Imagesearch with relative attribute feedback,” in CVPR, 2012.

[3] S. Dhar, V. Ordonez, and T. L. Berg, “High level describableattributes for predicting aesthetics and interestingness,” in CVPR,2011.

[4] M. Gygli, H. Grabner, H. Riemenschneider, F. Nater, and L. V.Gool, “The interestingness of images,” in ICCV, 2013.

[5] Y.-G. Jiang, YanranWang, R. Feng, X. Xue, Y. Zheng, and H. Yang,“Understanding and predicting interestingness of videos,” inAAAI, 2013.

[6] D. Parikh and K. Grauman, “Relative attributes,” in ICCV, 2011.[7] Z. Zhang, C. Wang, B. Xiao, W. Zhou, and S. Liu, “Robust relative

attributes for human action recognition,” Pattern Analysis andApplications, 2013.

[8] K. Chen, C. Wu, Y. Chang, and C. Lei, “Crowdsourceable QoEevalutation framework for multimedia content,” in ACM MM,2009.

[9] Y. Ma, T. Xiong, Y. Zou, and K. Wang, “Person-specific ageestimation under ranking framework,” in ACM ICMR, 2011.

[10] O. Chapelle and S. S. Keerthi, “Efficient algorithms for rankingwith svms,” Inf. Retr., 2010.

[11] X. Chen and P. N. Bennett, “Pairwise ranking aggregation in acrowdsourced setting,” in ACM International Conference on WebSearch and Data Mining, 2013.

[12] O. Wu, W. Hu, and J. Gao, “Learning to rank under multipleannotators,” in IJCAI, 2011.

[13] C. Long, G. Hua, and A. Kapoor, “Active visual recognition withexpertise estimation in crowdsourcing,” in ICCV, 2013.

[14] A. Kittur, E. H. Chi, and B. Suh., “Crowdsourcing user studieswith mechanical turk,” in ACM CHI, 2008.

[15] A. Kovashka and K. Grauman, “Attribute adaptation for per-sonalized image search,” in The IEEE International Conference onComputer Vision (ICCV), December 2013.

[16] P. Welinder, S. Branson, S. Belongie, and P. Perona, “The multidi-mensional wisdom of crowds,” in NIPS, pp. 2424–2432, 2010.

[17] V. C. Raykar, S. Yu, L. H. Zhao, A. Jerebko, C. Florin, G. H.Valadez, L. Bogoni, and L. Moy, “Supervised learning from mul-tiple experts: Whom to trust when everyone lies a bit,” in ICML,pp. 889–896, 2009.

[18] J. Whitehill, T. fan Wu, J. Bergsma, J. R. Movellan, and P. L.Ruvolo, “Whose vote should count more: Optimal integration oflabels from labelers of unknown expertise,” in NIPS, 2009.

[19] X. Jiang, L.-H. Lim, Y. Yao, and Y. Ye, “Statistical ranking andcombinatorial hodge theory,” Math. Program., 2011.

[20] R. Tibshirani, “Regression shrinkage and selection via the lasso,”J. of the Royal Statistical Society, Series B, 1996.

[21] I. Gannaz, “Robust estimation and wavelet thresholding in partiallinear models,” Stat. Comput., vol. 17, pp. 293–310, 2007.

[22] F. L. Wauthier, N. Jojic, and M. I. Jordan, “A comparative frame-work for preconditioned lasso algorithms,” in Neural InformationProcessing Systems, 2013.

[23] Y. Ke, X. Tang, , and F. Jing, “The design of high-level featuresfor photo quality assessment,” in CVPR, 2006.

[24] P. Isola, J. Xiao, A. Torralba, and A. Oliva, “What makes an imagememorable?,” in CVPR, 2011.

[25] F. Liu, Y. Niu, and M. Gleicher, “Using web photos for measuringvideo frame interestingness,” in IJCAI, 2009.

[26] C. H. Lampert, H. Nickisch, and S. Harmeling, “Learning todetect unseen object classes by between-class attribute transfer,”in CVPR, 2009.

[27] A. Farhadi, I. Endres, D. Hoiem, and D. Forsyth, “Describingobjects by their attributes,” in CVPR, 2009.

[28] A. Shrivastava, S. Singh, and A. Gupta, “Constrained semi-supervised learning via attributes and comparative attributes,”in ECCV, 2012.

[29] A. Parkash and D. Parikh, “Attributes for classifier feedback,” inECCV, 2012.

[30] A. Biswas and D. Parikh, “Simultaneous active learning of clas-sifiers and attributes via relative feedback,” in CVPR, 2013.

[31] A. Sorokin and D. Forsyth, “Utility data annotation with amazonmechanical turk,” in CVPR Workshops, 2008.

[32] G. Patterson and J. Hays, “Sun attribute database: Discovering,annotating, and recognizing scene attributes.,” in CVPR, 2012.

[33] W. V. Gehrlein, “Condorcet’s paradox,” Theory and Decision, 1983.[34] L. Liang and K. Grauman, “Beyond comparing image pairs:

Setwise active learning for relative attributes,” in CVPR, 2014.[35] Q. Xu, Q. Huang, T. Jiang, B. Yan, W. Lin, and Y. Yao, “Hodgerank

on random graphs for subjective video quality assessment,” IEEETMM, 2012.

Page 16: Edinburgh Research Explorer€¦ · abide by the legal requirements associated with these rights. Take down policy The University of Edinburgh has made every reasonable effort to

15

[36] Q. Xu, Q. Huang, and Y. Yao, “Online crowdsourcing subjectiveimage quality assessment,” in ACM MM, 2012.

[37] M. Maire, S. X. Yu, and P. Perona, “Object detection and segmen-tation from joint embedding of parts and pixels,” in ICCV, 2011.

[38] Z. Cao, T. Qin, T.-Y. Liu, M.-F. Tsai, and H. Li, “Learning to rank:From pairwise approach to listwise approach,” in ICML, 2007.

[39] Y. Liu, B. Gao, T.-Y. Liu, Y. Zhang, Z. Ma, S. He, and H. Li,“Browserank: letting web users vote for page importance,” inACM SIGIR, 2008.

[40] Z. Sun, T. Qin, Q. Tao, and J. Wang, “Robust sparse rank learningfor non-smooth ranking measures,” in ACM SIGIR, 2009.

[41] Y. Fu, T. M. Hospedales, T. Xiang, S. Gong, and Y. Yao, “Interest-ingness prediction by robust learning to rank,” in ECCV, 2014.

[42] T. Hastie, R. Tibshirani, and J. Friedman, The elements of statisticallearning:Data Mining, Inference, and Prediction (2nd). Springer, 2009.

[43] A. N. Hirani, K. Kalyanaraman, and S. Watts, “Least squaresranking on graphs,” arXiv:1011.1716, 2010.

[44] P. J. Huber, Robust Statistics. New York: Wiley, 1981.[45] T. Sun and C.-H. Zhang, “Scaled sparse linear regression,”

Biometrika, vol. 99, no. 4, pp. 879–898, 2012.[46] A. Belloni, V. Chernozhukov, and L. Wang, “Pivotal recovery

of sparse signals via conic programming,” Biometrika, vol. 98,pp. 791–806, 2011.

[47] Y. She and A. B. Owen, “Outlier detection using nonconvexpenalized regression,” Journal of American Statistical Association,2011.

[48] J. Friedman, T. Hastie, and R. Tibshirani, “Regularization pathsfor generalized linear models via coordinate descent,” Journal ofStatistical Software, vol. 33, no. 1, pp. 1–22, 2010.

[49] J. Fan and R. Li, “Variable selection via nonconcave penalizedlikelihood and its oracle properties,” JASA, 2001.

[50] J. Fan, R. Tang, and X. Shi, “Partial consistency with sparseincidental parameters,” arXiv:1210.6950, 2012.

[51] B. Efron, T. Hastie, I. Johnstone, and R. Tibshirani, “Least angleregression,” Annals of Statistics, 2004.

[52] N. Meinshausen and P. Buhlmann, “High-dimensional graphsand variable selection with the lasso,” Ann. Statist., 2006.

[53] Q. Xu, J. Xiong, Q. Huang, and Y. Yao, “Robust evaluation forquality of experience in crowdsourcing,” in ACM MM, 2013.

[54] N. Kumar, A. C. Berg, P. N. Belhumeur, and S. K. Nayar, “At-tribute and simile classifiers for face verification,” in ICCV, 2009.

[55] A. Oliva and A. Torralba., “Modeling the shape of the scene:Aholistic representation of the spatial envelope,” IJCV, vol. 42,2001.

[56] Y. Fu, G. Guo, and T. Huang, “Age synthesis and estimation viafaces: A survey,” TPAMI, 2010.

[57] S. Mika, B. Scholkopf, A. Smola, K.-R. Muller, M. Scholz, andG. Ratsch, “Kernel PCA and de-noising in feature spaces,” inNIPS, pp. 536–542, 1999.

[58] A. Vedaldi and A. Zisserman, “Efficient additive kernels viaexplicit feature maps,” in IEEE TPAMI, 2011.

[59] B. Carterette, “On rank correlation and the distance betweenrankings,” in ACM SIGIR, 2009.

[60] R. Hunter, “Mm algorithms for generalized bradley-terry mod-els,” The Annals of Statistics, vol. 32, p. 2004, 2004.

[61] H. Azari Soufiani, W. Chen, D. C. Parkes, and L. Xia, “GeneralizedMethod-of-Moments for Rank Aggregation,” in NIPS, 2013.

[62] F. Caron and A. Doucet, “Efficient bayesian inference for gener-alized bradley-terry models,” 2012.

[63] S. X. Yu, “Angular embedding: A robust quadratic criterion,”TPAMI, 2012.

Yanwei Fu received the PhD degree fromQueen Mary University of London in 2014, andthe MEng degree in the Department of Com-puter Science & Technology at Nanjing Univer-sity in 2011, China. He is currently a Post-docof Disney Research, Pittsburgh. His researchinterest is image and video understanding.

Timothy M. Hospedales received the PhD de-gree in neuroinformatics from the University ofEdinburgh in 2008. He is currently a lecturer (as-sistant professor) of computer science at QueenMary University of London. His research inter-ests include probabilistic modelling and machinelearning applied variously to problems in com-puter vision, data mining, interactive learning,and neuroscience. He has published more than30 papers in major international journals andconferences. He is a member of the IEEE.

Tao Xiang received the PhD degree in electri-cal and computer engineering from the NationalUniversity of Singapore in 2002. He is currentlya reader (associate professor) in the School ofElectronic Engineering and Computer Science,Queen Mary University of London. His researchinterests include computer vision and machinelearning. He has published over 100 papers ininternational journals and conferences and co-authored a book, Visual Analysis of Behaviour:From Pixels to Semantics.

Jiechao Xiong is currently pursuing the Ph.D.degree in statistics in BICMR & School ofMathematical Science, Peking University, Bei-jing, China. His research interests include sta-tistical learning, data science and topologicaland geometric methods for high-dimension dataanalysis.

Shaogang Gong is Professor of Visual Compu-tation at Queen Mary University of London, aFellow of the Institution of Electrical Engineersand a Fellow of the British Computer Society.He received his D.Phil in computer vision fromKeble College, Oxford University in 1989. Hisresearch interests include computer vision, ma-chine learning and video analysis.

Yizhou Wang Dr. Yizhou Wang is a Professorof Computer Science Department at Peking Uni-versity, Beijing, China. He is a vice director of In-stitute of Digital Media at Peking University, andthe director of New Media Lab of National En-gineering Lab of Video Technology. He receivedhis Ph.D. in Computer Science from Universityof California at Los Angeles (UCLA) in 2005. Dr.Wang’s research interests include computationalvision, statistical modeling and learning, patternanalysis, and digital visual arts.

Yuan Yao received his Ph.D. in mathematicsfrom the University of California, Berkeley, in2006. Since then he has been with StanfordUniversity and in 2009, he joined the School ofMathematical Sciences, Peking University, Bei-jing, China, as a professor of statistics. His cur-rent research interests include topological andgeometric methods for high dimensional dataanalysis and statistical machine learning, withapplications in computational biology, computervision, and information retrieval.