recognition of ongoing complex activities by sequence prediction

9
Recognition of Ongoing Complex Activities by Sequence Prediction over a Hierarchical Label Space Wenbin Li Mario Fritz Max Planck Institute for Informatics wenbinli,[email protected] Abstract Human activity recognition from full video sequence has been extensively studied. Recently, there has been increas- ing interest in early recognition or recognition from partial observation. However, from a small fraction of the video, it might be demanding if not even impossible to make a fine grained prediction of the activity that is taking place. Therefore, we propose the first method to predict ongoing activities over a hierarchical label space. We approach this task as a sequence prediction problem in a recurrent neural network where we predict over a hierarchical label space of activities. Our model learns to realize accuracy-specificity trade-offs over time by starting with coarse labels and pro- ceeding to more fine grained recognition as more evidence becomes available in order to meet a prescribed target ac- curacy. In order to study this task we have collected a large video dataset of complex activities with long duration. The activities are annotated in a hierarchical label space from coarse to fine. By directly training a sequence predictor over the hierarchical label space, our method outperforms several baselines including prior work on accuracy speci- ficity tradeoffs originally developed for object recognition. 1. Introduction Strong progress has been achieved in the area of hu- man activity recognition over the last decade ranging from coarse actions [34] to fine-grained activities [30]. The ma- jority of these techniques focus on what has happened. However, many application in surveillance system, crime prevention, assistive technology for daily living, human computer interaction interface, human robot interaction would like to infer which activity is currently going on in or- der to enable a system to behave proactive or context-aware. Current system that address early recognition and also activity recognition in general have a relatively short tem- poral reach. Typically the investigated activities are on the Pottery Crafting Teapot Full Observation Observation Observation Observation Doing Figure 1: An overview of our approach to inferring goal during the recognition process of a complex activity by pre- dicting with semantic abstraction. order of a minute – with a few exceptions like cooking ac- tivities. This limits the systems in the way they can act and assist to a more immediate, anticipatory response. In order to increase the temporal reach of our systems we have to study a richer set of longer term activities. While this goal is desirable, it raises the question how much can be inferred about a complex activity from a very short observation? The natural concern is that in many cases the information might be quite limited until a more sub- stantial fraction is being observed. Therefore we argue that recognition of complex activities from partial observation has to go hand in hand with a mechanism to predict at dif- ferent semantic granularity in order to deliver a useful pre- diction at anytime. In this sense, we have to bring coarse and fine-grained recognition together in a single approach. In order to study this challenging problem, we propose the first dataset of complex activities that have a substan- tial time span and that are organized in a semantic hier- archy. We study how prediction can be performed from partial observation of ongoing activities in this setting by employing of a semantic back-off strategy. We introduce a new representation that models the uncertainty over differ- ent time frames across different levels of the semantic hier- archy. More specifically, we propose a recurrent neural net- work formulation that learns to directly predict on this hier-

Upload: vuthien

Post on 06-Feb-2017

219 views

Category:

Documents


2 download

TRANSCRIPT

Page 1: Recognition of Ongoing Complex Activities by Sequence Prediction

Recognition of Ongoing Complex Activities bySequence Prediction over a Hierarchical Label Space

Wenbin Li Mario FritzMax Planck Institute for Informaticswenbinli,[email protected]

Abstract

Human activity recognition from full video sequence hasbeen extensively studied. Recently, there has been increas-ing interest in early recognition or recognition from partialobservation. However, from a small fraction of the video,it might be demanding if not even impossible to make afine grained prediction of the activity that is taking place.Therefore, we propose the first method to predict ongoingactivities over a hierarchical label space. We approach thistask as a sequence prediction problem in a recurrent neuralnetwork where we predict over a hierarchical label space ofactivities. Our model learns to realize accuracy-specificitytrade-offs over time by starting with coarse labels and pro-ceeding to more fine grained recognition as more evidencebecomes available in order to meet a prescribed target ac-curacy. In order to study this task we have collected a largevideo dataset of complex activities with long duration. Theactivities are annotated in a hierarchical label space fromcoarse to fine. By directly training a sequence predictorover the hierarchical label space, our method outperformsseveral baselines including prior work on accuracy speci-ficity tradeoffs originally developed for object recognition.

1. Introduction

Strong progress has been achieved in the area of hu-man activity recognition over the last decade ranging fromcoarse actions [34] to fine-grained activities [30]. The ma-jority of these techniques focus on what has happened.However, many application in surveillance system, crimeprevention, assistive technology for daily living, humancomputer interaction interface, human robot interactionwould like to infer which activity is currently going on in or-der to enable a system to behave proactive or context-aware.

Current system that address early recognition and alsoactivity recognition in general have a relatively short tem-poral reach. Typically the investigated activities are on the

Pottery

Crafting

Teapot

Full Observation

Observation

Observation

Observation

Doing………

………

………

………

Figure 1: An overview of our approach to inferring goalduring the recognition process of a complex activity by pre-dicting with semantic abstraction.

order of a minute – with a few exceptions like cooking ac-tivities. This limits the systems in the way they can act andassist to a more immediate, anticipatory response. In orderto increase the temporal reach of our systems we have tostudy a richer set of longer term activities.

While this goal is desirable, it raises the question howmuch can be inferred about a complex activity from a veryshort observation? The natural concern is that in many casesthe information might be quite limited until a more sub-stantial fraction is being observed. Therefore we argue thatrecognition of complex activities from partial observationhas to go hand in hand with a mechanism to predict at dif-ferent semantic granularity in order to deliver a useful pre-diction at anytime. In this sense, we have to bring coarseand fine-grained recognition together in a single approach.

In order to study this challenging problem, we proposethe first dataset of complex activities that have a substan-tial time span and that are organized in a semantic hier-archy. We study how prediction can be performed frompartial observation of ongoing activities in this setting byemploying of a semantic back-off strategy. We introduce anew representation that models the uncertainty over differ-ent time frames across different levels of the semantic hier-archy. More specifically, we propose a recurrent neural net-work formulation that learns to directly predict on this hier-

Page 2: Recognition of Ongoing Complex Activities by Sequence Prediction

archical label space from coarse to fine in order to achievean accuracy specificity trade off. A performance measureis proposed that evaluates the gain in specificity over timeas more observations become available. Our new methodoutperforms several baselines as well as techniques that weadopt from the literature on accuracy specificity tradeoffsfor object recognition.

2. Related Work

Early computer vision research in recognizing humanmotion and activities can be date back in early 1990’s [1].Researchers have explored various approaches to tackle thetask and excellent surveys are available [25, 28, 1, 40].In this section, we first review related datasets for activityrecognition and then discuss some related approaches foranalysis from partial observation.

Activity Datasets There is a large number of datasets pro-posed for activity recognition or detection with various lev-els of complexity. Early efforts to construct such datasetsinclude the KTH dataset [34] and the Weizmann dataset[3] which feature simple activities like walking and jump-ing. More complex datasets are introduced later with var-ious backgrounds, slightly longer duration and more im-portantly, with additional interaction of objects or peoples.Typical examples are answering phone (human-object in-teraction) in the Hollywood2 [23], golf swinging (human-object interaction) in the UCF Sports [29] and hand shak-ing (human-human interaction) in the UT-Interaction [33].However, activities covered in these datasets usually consistonly very few types of (in most cases only one) interactions,hence their structure are still relatively simple and individ-ual video’s length is rather short. In contrast to this, thereare also long-duration datasets collected from surveillancevideo like VIRAT [26] or daily living recordings like TUMKitchen [38], CMU-MMAC [8], URDAL [24], TRECVID[27], ego-centric datasets [11, 36], and MPII Cooking [30].Videos from these datasets show complex activities thatcomprise multiple interactions of different type. Consid-ering the example of making a dish in the cooking dataset[30], one needs to interact with different tools and ingredi-ents to complete a recipe.

One limitation about current long-duration datasets isthat they are mostly used for recognition of elementary ac-tivities within the long sequence instead of the recogniz-ing the overall process. In addition, they are limited to aparticular domain and do not provide a hierarchical labelspace. Our dataset exactly aims to fill this gap by provid-ing dataset of complex activities across multiple domainsthat are organized in a semantic label hierarchy. A relateddataset was built in [10] where activities are grouped withrespect to social interactions and places where the activity

Dataset #Class Avg. #Frames per Seq.KTH 6 91UT interaction 6 84UCF sports 9 58HMDB51 51 93MPI cooking 65 157Youtube action 11 163Olympics sports 16 234Hollywood2 12 285VIRAT 12 357KSCGR 8 786Our dataset 48 8.7K

Table 1: Action/activity datasets.

usually takes places. There are several recent works collect-ing videos from YouTube [22, 19, 35]. We also collect ourdataset from web sources, as it is a practical way to collecta sizable dataset of realistic videos.

Analysis from Partial Observation There are a few re-cent works focusing on analysis from partial observation.Early recognition [32] aims to recognize activity from theforepart of videos. The authors use accumulated histogramwith respect to the observed frames to represent the activityof interest, and performs sequence matching with templatesaveraged from training dataset. Following the similar ideaof sequence matching, [5] relaxes the location of observedpart and addresses recognition from partial observation andapplies sparse coding for better representation of the match-ing likelihood. A slightly different line of work like [13]focuses on early detection, deciding the temporal locationof the activity, yet as shown in [4] the algorithm does notwork so well in practical recognition task. A more recentwork [20] builds the early activity prediction on the humandetection and uses the resulted tracks to form a hierarchicalrepresentation, yet in real world scenario, the detection andtracking are usually expensive for complex scene involvingmultiple people and can often be unreliable, not mention insituations like ego-centric video where it is impossible toobtain a detection result for the person doing the activity.

Beyond predicting a label for the sequence, [17] posesa more challenging task to predict activity where a MDP isused for the distribution over all possible human navigationtrajectories. The walking path addressed in [17] is still rel-atively simple, hence [18] further explores indoor activitieswhich involves more complex interactions between humanand object. They represent the distribution over incomingstates with a set of particles sampled from an augmentedCRF. The activities exploited in [18], like taking medicine,already meets our definition of complexities, yet it only ap-plies to a very limited number of objects in very simplescenes under controlled lab conditions. It is not clear how itscale up to real world application and longer time scales.

Page 3: Recognition of Ongoing Complex Activities by Sequence Prediction

Peel Peel Cut Mix! ?

?(Making a Salad)

Action Prediction

Early Recognition with

accuracy-specificity trade-off

Observed

Early Recognition

?(Making a Garden Salad)

?(Making a Salad)

?(Cooking)

Unobserved

Figure 2: Example to show the difference between actionprediction and early recognition. Given a streaming videosequence, (top) action prediction aims to give a label to a in-coming local temporal segment (mix or cut); (middle) stan-dard early recognition gives a label for the global video se-quence (make a salad); (bottom) our proposed framework topredict in the hierarchy with accuracy-specificity trade-off.The solid arrows mark when the decision is made.

Prediction vs. Early Recognition Here we want todistinguish two different types of problems addressed in re-lated literature, i.e. prediction for unseen action/activity orearly recognition on the current observation. This is illus-trated in Figure 2 with the example of making a salad whichinvolves multiple activities including peeling, cutting andmixing. As one watches the activity moving on, predictionis to forecast what is the next move after mixing whereasearly recognition is to tell the main theme of the activity ascooking regardless of only partial observation of the wholeprocess. [17, 18] are examples of prediction and [32, 5, 20]are for early recognition.

Indeed predicting real world activity is a very difficulttask, it generally requires various components including butnot limited to pose estimation, object detection and tempo-ral segmentation where each task along is challenging al-ready in real world application not to mention it is nontriv-ial to combine them for a reasonable solution. Our workfocuses on early recognition which due to the long tempo-ral duration of our investigated activities, also extend to thefuture.

Activity Recognition with Recurrent Neural Networks[2] applies a different variant of RNN to action recogni-tion task. Our work differs from it in that we focus onthe early recognition and prediction over a hierarchical label

space with a learned accuracy-specificity tradeoff. Our re-sults show the joint training of the sequence model w.r.t. tothe accuracy specificity tradeoff is key to our performanceimprovements.

3. MethodIn this section, we first discuss different ways of video

representation and formulate the task of early recognitionin video sequence. Our goal is to predict at any point in thesequence over a hierarchical label space in order to realizeaccuracy-specificity trade-offs. In order to have a temporalintegration of information over time and learn classifier andthe hierarchical tradeoff jointly, we employ recurrent neuralnetworks for modeling videos and extend it to labeling ina hierarchy, and learning accuracy-specificity trade-off inearly activity recognition in videos.

3.1. Video Representation

Given a specific descriptor x, a video sequence canthen be represented as [x1, x2, ..., xT ], where T denotes theframe index. There are several ways to obtain a compactrepresentation for sequences as shown in Figure 3: (1)pooling over the whole sequence; (2) partition the video intoseparated temporal segments, and combine the representa-tion from the segments into a temporal sequence represen-tation [x1:t, xt+1:2t, ...xT−t+1:T ], where t is the number offrames for each segment. This approach has been applied in[4, 37]; (3) sample clips (also temporal segments) from thevideo to represent the video, each clip is trained as a singleinstance in the video class, i.e. [xt1:t1+t] ∪ [xt2:t2+t] ∪ ...,where t1, t2, ... are the time stamps when the clips are sam-pled, and final prediction on the video is based on the ma-jority vote from the sampled clips. This approach is appliedin the large-scale video recognition work [16].

We use improved dense trajectory feature [39] as thebasic feature to encode spatio-temporal information acrossframes in the videos with a code book of size 4000 obtainedvia k-means clustering. It uses a dense representation ofextracted trajectories and combines with trajectory-alignedfeatures, including HOG [6], HOF [21], motion boundaryhistograms (MBF) [7]. The descriptor itself has been shownto achieved the state-of-art performances on several publicdatasets.

3.2. Early Recognition

Given a video sequence [x1, x2, ..., xT ], where T is thelast frame index of the video. Early recognition generallyrefers to the task of classification:

y = f([x1, x2, ..., xt]) , t ≤ T

The standard full video classification can be seen as aspecial case of this formulation where t = T . It is impor-

Page 4: Recognition of Ongoing Complex Activities by Sequence Prediction

Temporal Segment Sampled ClipsFull Sequence

Figure 3: Different ways to represent individual video se-quence: full sequence (left), use one representation for thefull sequence; temporal segment (middle), partition full se-quence into temporal segments and combine representationfor individual segments into one representation; sampledclips (right), sample video clips from the sequence to formseveral representations.

tant to point out the difference between early recognitionof activity [32] and human action prediction/anticipation[17, 18, 20] whereas the former is essentially a sequenceclassification problem that maps sequence in a single label

X : [x1, x2, ..., xT ] 7→ y

and the latter is a temporal classification problem thatmaps a input sequence into a target sequence.

X : [x1, x2, ..., xT ] 7→ [y1, y2, ..., yL]

A specific example is shown in Figure 2 for ongoing videostream of making salad. Assume we already observe theperson in video peeled the cucumber, cut it into slice, mix itwith other ingredients, early recognition generally aims tooutput a global label for the sequence based on the availableobservation, in this case, the label should be ‘making salad’while the action prediction tries to predict the specific localaction label for the incoming video frames, in this case, thecorrect label would be ‘put the salad in plate’.

3.3. Recurrent Neural Networks

A recurrent neural network (RNN) is a class of artifi-cial neural network that allows connections between unitsto form a directed cycle. We consider an architecture withone self-connected hidden layer, which can be unrolled intime as shown in Figure 4. One notable merit for RNN isthat the recurrent connections allow a ‘memory’ of previ-ous inputs to persist in the network’s internal state whichcan then be used to influence the network output [12]. Thischaracteristic makes it suitable for sequence analysis, espe-cially in our case, with long duration and complex videosequences.

Given a video sequence composed of temporal segmentsx1, x2, ..., xT , each in Rd, the network computes a se-quence of hidden states h1, h2, ..., hT , each in Rm and pre-dictions y1, y2, ..., yT , each in Rk. The hidden unit in-tegrates the information from the arrived observation and

those propagated from previous blocks of the networks:

xi = Wxhxi +Whhhi−1 + bh

hi = tanh(xi)

Each prediction unit represents the class label up to the cur-rent observation and is activated by the hidden unit via asoftmax function:

hi = Whyhi + by

yi = softmax(hi)

where Wxh,Whh,Why are the weight matrices and bh, byare the biases. The training is done by BackPropagationThrough Time (BPTT) [31] and the parameter is learned byconjugate gradient descent method.

3.4. Early Recognition in a Semantic Hierarchy

Deng et al. [9] first introduce the concept of optimizingan accuracy-specificity tradeoff for hierarchical image clas-sification with the DARTS algorithm. Here we briefly recapthe formulation of the concept and discuss how we extendit to our settings.

Optimizing Accuracy-Specificity Trade-Offs The keyidea behind the accuracy-specificity trade-off is to makecost-sensitive prediction over the hierarchy, where predic-tions at upper level (fine-grained level categories) of the hi-erarchy get penalized more than those at lower level (coarselevel categories). By optimizing over both the specificityand accuracy an optimal tradeoff is found. More formally,given a classifier f : X 7→ Y , with accuracy Φ(f) definedas,

Φ(f) = E[f(X) ∈ π(Y )]

where π(Y ) is the set of all possible correct predictions,[P ] is the Iverson bracket, i.e., 1 if P is true and 0 otherwise.The preference for specific class labels over general classlabels at each node v, i.e. the specificity is encoded withinformation gain (decrease in entropy) by

rv = log2|Y | − log2∑y∈Y

[v ∈ π(y)]

The total reward for the classifier f is hence defined as

R(f) = E(rf (X)[f(X) ∈ π(Y )])

The optimal trade-off between accuracy and specificityis then formulated as maximizing the reward given an accu-racy guarantee 1− ε ∈ (0, 1]:

minimizef

R(f)

subject to Φ(f) ≥ 1− ε

Page 5: Recognition of Ongoing Complex Activities by Sequence Prediction

xt

ht

yt

(a)

xt-1 xt xt+1

ht-1 ht ht+1

yt-1 yt yt+1

… …

(b) xt-1 xt xt+1

ht-1 ht ht+1

yt-1,L3 yt,L3 yt+1,L3

… …

yt-1,L2 yt,L2 yt+1,L2

yt-1,L1 yt,L1 yt+1,L1

(c)

Figure 4: Graphical model for recurrent neural networks. (a) The recurrent neural networks with a single, self-connectedhidden layer. (b) The unrolled model with respect to discrete time steps from (a). (c) The recurrent network with structuraloutput over the hierarchy.

Accuracy-specificity Trade-off Over Time Motivatedby ideas from incremental feature computation and anytimerecognition [14, 15], we extend this concept with a tempo-ral dimension to early recognition and model the decisionprocess when analyzing an ongoing video stream. At eachtime step, we have to predict a label in the hierarchy andhereby trade-off between accuracy and specificity.

R(f, t) = E(rf (Xt)[f(Xt) ∈ π(Y )])

The intuition behind this is that when only observing a smallportion of the video, we have little evidence to accuratelypredict at a fine-grained level but may still be able to givea sensible coarse-level class label. By considering the totalcost of possible wrong prediction at fine-grained level andprobably correct prediction at coarse-level, together withour preference to predict at different levels, we are likelyto give prediction at coarse-level given little observed data.When observing more and more parts of the video, we be-come more certain about our prediction at the fine-grainedlevel, and by the same mechanism, we start to predict at amore fine-grained level. Figure 2 shows a concrete exam-ple. By summing up the term over time T ,

T∑t

R(f, t) =

T∑t

E(rf (Xt)[f(Xt) ∈ π(Y )])

we can evaluate the efficiency for accuracy-specificity froma classifier f .

Structured Output RNN for Predictions Over Hierar-chical Label Spaces We propose an RNN optimizing theobjective from above by predicting over a structured out-put space – our hierarchy H . We denote labels at top-layer,middle-layer and bottom layer in the hierarchy as Y1, Y2, Y3(coarse to fine). As shown in Figure 4, our RNN model di-rectly predicts an output layer y3,i representing the posterior

probabilities over the fine grained labels in the bottom layer.

y3,i = p(Y3 = i|x), i = 1, ...,K3

where K3 is the number of classes within the layer. On topof this layer, we introduce an additional layer to representthe middle layer, where the connections between these twolayers are defined according to the hierarchy H , i.e. if classi in bottom layer belongs to class j in middle layer, node iand j are connected or (i, j) ∈ H . Accordingly, the middlelayer activations are defined as follows:

y2,j = p(Y2 = j|x), j = 1, ...,K2

=∑i

p(Y2 = j|Y3 = i)p(Y3 = i|x)

Similarly we define another layer above this layer to repre-sent top layer prediction (coarse labels):

y1,k = p(Y1 = k|x), k = 1, ...,K1

=∑j

p(Y1 = k|Y2 = j)p(Y2 = j|x)

=∑j

∑i

p(Y1 = k|Y2 = j)p(Y2 = j|Y3 = i)p(Y3 = i|x)

The complete model is shown in Figure 4. Based on the thisprediction over the hierarchy we define a structured loss inorder to optimize for the desired accuracy specificity trade-off:

Loss(θ,D) = −∑i

log p(Y1 = y(i)1 |x(i), θ)

+ log p(Y2 = y(i)2 |x(i), θ)

+ log p(Y3 = y(i)3 |x(i), θ)

where D denotes the training set{X,Y1, Y2, Y3}, and i in-dexes the i-th data instance in D.

Page 6: Recognition of Ongoing Complex Activities by Sequence Prediction

Figure 5: 3-layer hierarchy defined in our dataset.

4. ExperimentWe first present our new dataset of complex activities

with a hierarchical label space and afterwards perform aquantitative comparison of our proposed method and com-pare to several baselines.

4.1. Datasets

We explore various complex activities composed of mul-tiple interactions. Recording videos for a large set of diverseclasses is difficult, considering we need to find experts indifferent fields to perform the activities and capture multiplevideos for a single class. In addition, this would eliminatethe challenge of different capture devices. Hence, we buildour dataset on videos from web to create our dataset. In thefollowing part, we briefly discuss how we collect videos,pre-process the data and tackle the associated challenges ofbuilding such a dataset.

Data Collection We begin by defining a 3-layer seman-tic hierarchy (we count the root node of “Doing something”as layer 0) with 3 nodes in the first layer (cooking, craft-ing, repairing), then for each node, we select 4 specific de-rived categories as nodes to form the following layer, andfor each node in the middle layer, we select 4 more specificclasses as leaf nodes to form the bottom layer. For exam-ple, we include making “pizza”, “soup”, “salad” and “sand-wich” as the second layer for “cooking”, and for “salad”, weconsider making 4 different kinds of salads, namely “eggsalad”, “garden salad”, “chicken salad” and “greek salad”.Overall, we obtain a tree for complex activities with total3× 4× 4 = 48 activity classes as leaf nodes. The full hier-archy is shown in Figure 5. 10 videos are collected for eachleave node from YouTube and eHow, which sum up to 480video clips with total length of more than 41 hours or morethan 4.18 million frames. The dataset is available online1.Sample frames of the videos are shown in Figure 6. We use

1http://www.mpii.de/ongoing-activity

Figure 6: Some sample frames from our dataset.From topto bottom: make neapolitan pizza, make cheese steak sand-wich, repair bike brake, change car oil, make vase, buildchair.

half number of videos for training and the rest for testing.These videos from the web differ from the ones recorded

from lab: while the latter record the whole process for eachactivity with good controlled conditions, the former are of-ten edited (adding head leader, tail leader, titles, flashback,etc) from the uploaders under various conditions (differentpoint of view, shooting skills, etc). Hence such data is verynoisy and exposes many of the aforementioned challengesof realistic videos. We rescale video into 360p for furtherprocessing (640× 360).

4.2. Full Video Recognition

First we consider a setting where we classify full videosequences. This is particularly interesting since our col-lected videos are significant more complex than previousdatasets on both the temporal scale and internal structure.This provides a relative measure of difficulty for activityrecognition on our database. We train and predict on fullsequences with a SVM classifier. While a χ2 kernel isusually applied to integrate different descriptors in a multi-channel fashion as in [39], we find out that a linear kernelgets slightly better results on our dataset. Therefore we usethe linear SVM for all our experiments. As can be seen fromTable 2, the performance reaches 25.7% at layer−3 (fine-grained), which suggests that we have established indeed avery difficult task at this detailed level.

An alternative to training on full sequence is to use sam-pled clips to represent the whole video [16]. Accordingly,we randomly sample 20% of each video in the training setfor training, and test on the full video sequences in the testset. Note here we predict directly on the entire video se-quence instead of the average on the prediction of sampledclips from each test sequence. As shown in 2, this approachis also valid on our dataset and is only slightly worse thantraining on full video sequence.

Page 7: Recognition of Ongoing Complex Activities by Sequence Prediction

Layer Clips-training(%) Full-training(%)bottom-layer 25.3 25.7middle-layer 52.3 59.1top-layer 76.4 78.1

Table 2: Classification results on full video sequence

4.3. Recognition from Partial Observation

We proceed by examining the case where the video isonly partially observed.We simulate two types of video seg-ments to represent incremental arrival of video frames, (i)frames from the beginning with different observation ratios[10%, 20%, ..., 100%] where we explore different strategiesfor early recognition (ii) continuous frames taking up 10%of the full observation starting at different temporal loca-tion. As a result a video is represented as a temporal se-quence of these frame segments [0−10%, 10−20%, ..., 90−100%, ]. This setting simulates the practical setting wherethe observer starts from arbitrary position and perceive apart of videos and wants to infer the current activity class.We refer to this as online recognition.

Early Recognition We evaluate the following strategiesfor early recognition: (1) train single model on full videoand predict at different observation ratios; (2) train withaugmented data i.e the combination of full video and videosegments of different observation ratios and make predic-tions; (3) train on sampled video segments and test on par-tial observation, this is inspired by the result from our full-video recognition experiment and we would like to inves-tigate how models trained on sampled clips can discrimi-nate activity classes based on partial observation. We usethe same pipeline as in the full video recognition and alsotest on differ layers in the label hierarchy. The results areshown in Figure 7. As we see from the plots, while thetraining on sampled clips gets slightly worse results than theother two settings, i.e. training on full sequence and on bothfull sequence and augmented data is still feasible. Compar-ing training on full sequence with and without the augmentdataset, we observe improvement on upper two layer. Atthe lower-level, it helps when the observation ratio is below60% but degrades the performance slightly afterwards.

In addition to early recognition setting, we also evaluatethe accuracy-specificity trade-off over time as shown in Fig-ure 8. We compare to our adaption of the DARTS algorithmas described above as a baseline. As we can see from theexpected information gain at fixed target accuracy, trainingwith full sequence and augment data (red curve) most of thetime achieve the best reward over the other two, i.e. trainingon full sequence (green curve) and training on clips (bluecurve). To help better understand the concept, Figure 8 alsoshows examples of prediction a distribution over the hierar-chy and observation ratios at several target accuracy. Theproportion of predictions at lower level grows with time,

which means the classifier gets more certain about the ac-tivity class over time. When specifying a lower target ac-curacy, there are more predictions at lower levels in the hi-erarchy, with higher target accuracy the other way around.In order to reach the target accuracy, the prediction has tomove to higher layer that has better confidence.

0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9

Target Accuracy1.0

1.5

2.0

2.5

3.0

3.5

Sum

/Avg

. Inf

orm

atio

n G

ain

DARTSRNNstructRNN

Figure 9: Results for online recognition.Online Recognition We evaluate online recognition us-ing our proposed RNN model that performs a structuredprediction over the label hierarchy (structRNN) and com-pare it to the DARTS algorithm and a plain RNN. For RNNand struct RNN, we use 50 units for the hidden layer anduse a L2 regularizer. Parameters were selected based onthe validation set. We perform dimensionality reduction byusing the decision value from linear classifiers trained overdifferent layers in order to reduce the raw feature vector of20K dimension. In a preliminary study we have observedthat this generally gives better performance on the expectedreward for accuracy-specificity trade-off. The final resultsare show in Figure 9. Compared to the baseline DARTS al-gorithm, the RNN achieves better performance. This is dueto the connectivity between the hidden units that improvethe propagation of information along time. The structRNNfurther improves the RNN performance as we enforce thestructural loss with respect to the hierarchy. In addition, wenote that our structRNN even outperforms the previouslyinvestigated early recognition settings. Figure 10 showssome example predictions for videos.

5. ConclusionIn this paper, we proposed a new challenging dataset

with hierarchical labels to study the recognition oflong-duration, complex activities and temporal accuracy-specificity trade-offs. We propose a new method based onrecurrent neural networks that learns to predicts over thishierarchy and realizes accuracy specificity tradeoffs. Ourmethod outperforms several baselines on this new challengeincluding an adaptation of hierarchical prediction from ob-ject recognition.

Acknowledgements This work was partially supportedby the DFG collaborative research center SFB 1223.

Page 8: Recognition of Ongoing Complex Activities by Sequence Prediction

0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0

Observation Ratio0.0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1.0

Acc

urac

y

Train on full observationbottom layer(48-class)middle layer(12-class)top layer(3-class)

0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0

Observation Ratio0.0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1.0

Acc

urac

y

Train on full observationbottom layer(48-class)middle layer(12-class)top layer(3-class)

0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0

Observation Ratio0.0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1.0

Acc

urac

y

Train on full observationbottom layer(48-class)middle layer(12-class)top layer(3-class)

Figure 7: (Left) train SVM model from full video and predict at different observation ratios; (middle) train model from fullvideo sequences and augmented dataset; (right) train model from sampled clips.

0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9

Target Accuracy1.0

1.5

2.0

2.5

3.0

3.5

Sum

/Avg

. Inf

orm

atio

n G

ain

full+aug2segfull2segseg2part

10 20 30 40 50 60 70 80 90 100Observation Ratio

0.0

0.2

0.4

0.6

0.8

1.0

Dis

tribu

tion

Guaranteed accuracy @ 20.0%

rootlayer1layer2layer3

10 20 30 40 50 60 70 80 90 100Observation Ratio

0.0

0.2

0.4

0.6

0.8

1.0

Dis

tribu

tion

Guaranteed accuracy @ 50.0%

rootlayer1layer2layer3

10 20 30 40 50 60 70 80 90 100Observation Ratio

0.0

0.2

0.4

0.6

0.8

1.0

Dis

tribu

tion

Guaranteed accuracy @ 70.0%

rootlayer1layer2layer3

Figure 8: (Left)the expected information over time given specific accuracy.(right) Example of prediction distribution overtime based on the optimization over accuracy-specificity trade-off for train on full and augment segments.

DARTS doing doing doing doingRNN doing doing doing doing

structRNN doing repairing changing filter changing filter

change fuel filter

change CPU

make a cup

DARTS doing repairing computer repairing repairingRNN repairing changing repairing computer repairing computer

structRNN repairing computer changing CPU changing CPU changing CPU

DARTS doing making a cup making a cup doingRNN doing doing pottery doing pottery making a cup

structRNN doing making a cup making a cup making a cup

Figure 10: Example prediction over three activities in the video, from top to bottom: change car filter, change a CPU andmake a cup.

Page 9: Recognition of Ongoing Complex Activities by Sequence Prediction

References[1] J. Aggarwal and M. S. Ryoo. Human activity analysis: A

review. ACM Computing Surveys (CSUR), 2011.[2] M. Baccouche, F. Mamalet, C. Wolf, C. Garcia, and

A. Baskurt. Sequential deep learning for human actionrecognition. In Human Behavior Understanding. Springer,2011.

[3] M. Blank, L. Gorelick, E. Shechtman, M. Irani, and R. Basri.Actions as space-time shapes. In ICCV, 2005.

[4] L. Cao, Y. Mu, A. Natsev, S.-F. Chang, G. Hua, and J. R.Smith. Scene aligned pooling for complex video recognition.In ECCV 2012. 2012.

[5] Y. Cao, D. Barrett, A. Barbu, S. Narayanaswamy, H. Yu,A. Michaux, Y. Lin, S. Dickinson, J. M. Siskind, andS. Wang. Recognizing human activities from partially ob-served videos. In CVPR, 2013.

[6] N. Dalal and B. Triggs. Histograms of oriented gradients forhuman detection. In CVPR, 2005.

[7] N. Dalal, B. Triggs, and C. Schmid. Human detection usingoriented histograms of flow and appearance. In ECCV. 2006.

[8] F. De la Torre, J. Hodgins, A. Bargteil, X. Martin, J. Macey,A. Collado, and P. Beltran. Guide to the carnegie mellonuniversity multimodal activity (cmu-mmac) database. 2008.

[9] J. Deng, J. Krause, A. C. Berg, and L. Fei-Fei. Hedgingyour bets: Optimizing accuracy-specificity trade-offs in largescale visual recognition. In CVPR, 2012.

[10] B. G. Fabian Caba Heilbron, Victor Escorcia and J. C.Niebles. Activitynet: A large-scale video benchmark for hu-man activity understanding. In CVPR, 2015.

[11] A. Fathi, A. Farhadi, and J. M. Rehg. Understanding ego-centric activities. In ICCV, 2011.

[12] A. Graves. Supervised sequence labelling with recurrentneural networks. Springer, 2012.

[13] M. Hoai and F. De la Torre. Max-margin early event detec-tors. In CVPR, 2012.

[14] S. Karayev, T. Baumgartner, M. Fritz, and T. Darrell. Timelyobject recognition. In NIPS, 2012.

[15] S. Karayev, M. Fritz, and T. Darrell. Anytime recognition ofobjects and scenes. In CVPR, 2014.

[16] A. Karpathy, G. Toderici, S. Shetty, T. Leung, R. Sukthankar,and L. Fei-Fei. Large-scale video classification with convo-lutional neural networks. In CVPR, 2014.

[17] K. M. Kitani, B. D. Ziebart, J. A. Bagnell, and M. Hebert.Activity forecasting. In ECCV. 2012.

[18] H. S. Koppula and A. Saxena. Anticipating human activitiesusing object affordances for reactive robotic response. RSS,2013.

[19] H. Kuehne, H. Jhuang, E. Garrote, T. Poggio, and T. Serre.Hmdb: a large video database for human motion recognition.In ICCV, 2011.

[20] T. Lan, T.-C. Chen, and S. Savarese. A hierarchical repre-sentation for future action prediction. In ECCV. 2014.

[21] I. Laptev, M. Marszalek, C. Schmid, and B. Rozenfeld.Learning realistic human actions from movies. In CVPR,2008.

[22] J. Liu, J. Luo, and M. Shah. Recognizing realistic actionsfrom videos in the wild. In CVPR, 2009.

[23] M. Marszalek, I. Laptev, and C. Schmid. Actions in context.In CVPR, 2009.

[24] R. Messing, C. Pal, and H. Kautz. Activity recognition usingthe velocity histories of tracked keypoints. In ICCV, 2009.

[25] T. B. Moeslund, A. Hilton, and V. Kruger. A survey of ad-vances in vision-based human motion capture and analysis.Computer vision and image understanding, 2006.

[26] S. Oh, A. Hoogs, A. Perera, N. Cuntoor, C.-C. Chen, J. T.Lee, S. Mukherjee, J. Aggarwal, H. Lee, L. Davis, et al.A large-scale benchmark dataset for event recognition insurveillance video. In CVPR, 2011.

[27] P. Over, G. Awad, J. G. Fiscus, B. Antonishek, M. Michel,W. Kraaij, A. F. Smeaton, and G. Quenot. Trecvid 2010 -an overview of the goals, tasks, data, evaluation mechanismsand metrics. In TRECVID, 2010.

[28] R. Poppe. A survey on vision-based human action recogni-tion. Image and vision computing, 2010.

[29] M. D. Rodriguez, J. Ahmed, and M. Shah. Action macha spatio-temporal maximum average correlation height filterfor action recognition. In CVPR, 2008.

[30] M. Rohrbach, S. Amin, M. Andriluka, and B. Schiele. Adatabase for fine grained activity detection of cooking activ-ities. In CVPR, 2012.

[31] D. E. Rumelhart, G. E. Hinton, and R. J. Williams. Learn-ing internal representations by error propagation. Technicalreport, DTIC Document, 1985.

[32] M. Ryoo. Human activity prediction: Early recognition ofongoing activities from streaming videos. In ICCV, 2011.

[33] M. Ryoo and J. Aggarwal. Ut-interaction dataset, icpr con-test on semantic description of human activities(sdha), 2010.

[34] C. Schuldt, I. Laptev, and B. Caputo. Recognizing humanactions: a local svm approach. In ICPR, 2004.

[35] G. Sergio, N. Krishnamoorthy, G. Malkarnenkar, T. Darrell,R. Mooney, and K. Saenko. Youtube2text: Recognizing anddescribing arbitrary activities using semantic hierarchies andzero-shoot recognition. In ICCV, 2013.

[36] S. Stein and S. J. McKenna. Combining embedded ac-celerometers with computer vision for recognizing foodpreparation activities. In Proceedings of the 2013 ACM inter-national joint conference on Pervasive and ubiquitous com-puting. ACM, 2013.

[37] K. Tang, L. Fei-Fei, and D. Koller. Learning latent temporalstructure for complex event detection. In CVPR, 2012.

[38] M. Tenorth, J. Bandouch, and M. Beetz. The tum kitchendata set of everyday manipulation activities for motion track-ing and action recognition. In ICCV Workshops, 2009.

[39] H. Wang, A. Klaser, C. Schmid, and C.-L. Liu. Action recog-nition by dense trajectories. In CVPR, 2011.

[40] D. Weinland, R. Ronfard, and E. Boyer. A survey of vision-based methods for action representation, segmentation andrecognition. Computer Vision and Image Understanding,2011.