kaihua zhang lei zhang (polyu, hong kong) ming-hsuan yang (uc merced, california, u.s.a. ) real-time...
TRANSCRIPT
Kaihua Zhang
Lei Zhang
(PolyU, Hong Kong)
Ming-Hsuan Yang
(UC Merced, California, U.S.A. )
Real-Time Compressive Tracking
European Conference on Computer Vision, 2012 (ECCV 2012)
Outline
• Introduction
•Random projection
•Classifier construction and update
• Experiments
•Conclusion
2
Outline
• Introduction
•Random projection
•Classifier construction and update
• Experiments
•Conclusion
3
Introduction
• Propose an effective and efficient tracking algorithm with an appearance model based on features extracted in the compressed domain.
• Our appearance model employs non-adaptive random projections that preserve the structure of the image feature space of objects
• Compress samples of foreground targets and the background using the same sparse measurement matrix
• The tracking task is formulated as a binary classification via a naive Bayes classifier with online update in the compressed domain
4
Compressed sensing
• Compressed sensing is a signal processing technique for efficiently acquiring and reconstructing a signal, by finding solutions to underdetermined linear systems (Emmanuel Candès, Terence Tao, and David Donoho)
• There are two conditions under which recovery is possible
• The first one is sparsity which requires the signal to be sparse in some domain
• The second one is incoherence which is applied through the restricted isometry property (RIP) which is sufficient for sparse signals.
5
Random Measurement Matrix
high-dimensional space
lower-dimensional space
Main components of compressive tracking
6
Frame(t) Samples
Classifier
Sparse projection
matrix
Compressed vectors
Image features
Main components of compressive tracking(2)
7
Sample with maximal classifier response
Frame(t+1)
Classifier
Frame(t+1)Sparse measurement
matrix
Multiscale Image features
Compressed vectors
Outline
• Introduction
•Random projection
•Classifier construction and update
• Experiments
•Conclusion
8
Random projection
9
A matrix R whose rows have unit length projects data from the high-dimensional image space x to a lower-dimensional space v
v = Rx where nm
Random projection(2)
10
Intuitive interpretation of JL lemma linear map f: ->
𝑹𝒏
Random projection(3)
• We write the Euclidean distance between two data vectors and in the original large-dimensional space as
• After the random projection, approximated by the scaled Euclidean distance of these vectors in the reduced space:
11
where d is the original and k the reduced dimensionality of the data set
Random projection(4)
12
(Johnson-Lindenstrauss lemma)
Given 0<<1, a set X of d points in and number n > 8ln(m)/,there is a linear map f: -> such that
The Johnson-Lindenstrauss Lemma
13
• Selecting matrix A that provide the desired result are:
Random Measurement Matrix
14
𝑹𝒎
𝑹𝒏
Construct random matrix R such that JL lemma
We adopt a very sparse random matrix with entries defined as
This type of matrix with s =3 satisfies the JL-lemma by Achlioptas
Random Measurement Matrix(2)
• In this work, we set s = m/4 which makes a very sparse random matrix
• For each row of R, only about c, c ≤ 4, entries need to be computed
• Therefore, the computational complexity is only O(cn) which is very low
• Furthermore, we only need to store the nonzero entries of R which makes the memory requirement also very light.
15
Multiscale filter
16
Haar-like features
• Haar-like features are digital image features used in object recognition
• A simple rectangular Haar-like feature can be defined as the difference of the sum of pixels of areas inside the rectangle
17
Dimensionality Reduction
18
Graphical representation of compressing a high-dimensional vector x to a low-dimensional vector v x х = 𝑣 𝑖=∑
𝑗
𝑟 𝑖𝑗 𝑥 𝑗
Negative entry Positive entry
Zero entry
Analysis of Low-Dimensional Compressive Features
19
Outline
• Introduction
•Random projection
•Classifier construction and update
• Experiments
•Conclusion
20
Classifier Construction
21
For each sample z its low-dimensional representaion
Assume all elements in v are independently distributed model them with a naïve Bayes classifier
=
Classifier Update
22
Contitional distributions and in the classifier H(v) are assumed to Gaussian distributed with four parameters
) )
The scalar parameters are incrementally updated:
+
where
Algorithm : Compressive Tracking
23
Outline
• Introduction
•Random projection
•Classifier construction and update
•Experiments
•Conclusion
24
Experiments
25
http://www4.comp.polyu.edu.hk/~cslzhang/CT/CT.htm
Experiments(2)
26
Experiments(3)
27
Outline
• Introduction
•Random projection
•Classifier construction and update
• Experiments
•Conclusion
28
Conclusion
• In this paper, we proposed a simple yet robust tracking algorithm with an ap- pearance model based on non-adaptive random projections that preserve the structure of original image space
• Numerous experiments with state-of-the-art algorithms on challenging sequences demonstrated that the proposed algorithm performs well in terms of accuracy, robustness, and speed.
29