bit-planes: dense subpixel alignment of binary descriptors · bit-planes: dense subpixel alignment...

10
Bit-Planes: Dense Subpixel Alignment of Binary Descriptors Hatem Alismail Robotics Institute Carnegie Mellon University [email protected] Brett Browning Robotics Institute Carnegie Mellon University [email protected] Simon Lucey Robotics Institute Carnegie Mellon University [email protected] Abstract Binary descriptors have been instrumental in the recent evolution of computationally efficient sparse image align- ment algorithms. Increasingly, however, the vision commu- nity is interested in dense image alignment methods, which are more suitable for estimating correspondences from high frame rate cameras as they do not rely on exhaustive search. However, classic dense alignment approaches are sensitive to illumination change. In this paper, we propose an easy to implement and low complexity dense binary descriptor, which we refer to as bit-planes, that can be seamlessly inte- grated within a multi-channel Lucas & Kanade framework. This novel approach combines the robustness of binary descriptors with the speed and accuracy of dense align- ment methods. The approach is demonstrated on a tem- plate tracking problem achieving state-of-the-art robustness and faster than real-time performance on consumer laptops (400+ fps on a single core Intel i7) and hand-held mobile devices (100+ fps on an iPad Air 2). 1. Introduction Binary descriptors such as BRIEF [10] & BRISK [24] are powerful tools for solving sparse image alignment prob- lems due to their discriminative power, robustness to illumi- nation change, and low complexity [21, 16, 20, 39]. Match- ing binary descriptors is typically performed by exhaustive search [7, 23] using the Hamming distance. Exhaustive search, however, is inefficient when dense correspondences are required in real-time [19, 30]. A classical way of speeding up the task of image align- ment is to linearize pixel intensities of an image with re- spect to geometric displacement. The most notable example of this strategy can be found in the seminal work of Lucas & Kanade [27]. The Lucas & Kanade (LK) algorithm at- tempts to establish an approximate linear relationship be- tween appearance and geometric displacement. Efficient linear solvers can then be employed for finding the best geo- metric alignment of the image with respect to a known tem- plate. The relationship between appearance and geometric displacement is seldom linear, so the linearization process is typically repeated until convergence. Due to its importance, numerous extensions and variations upon the LK algorithm have subsequently been explored in literature [2]. At the heart of the LK algorithm is the notion that an ap- proximate linear relationship between pixel appearance and geometric displacement can be reliably established. Pixel intensities are not deterministically differentiable with re- spect to geometric displacement. Instead, the linear rela- tionship is established stochastically through spatial finite differences whose outputs we refer to as image gradients. The notion of estimating stochastic gradients on image in- tensities has a long and rich history dating back to some of the most seminal works of computer vision [29]. Further, it has been well documented that pixel intensities within natural images are strongly correlated over small spatial ar- eas further validating the assumed approximate linear rela- tionship between pixel intensities and geometric displace- ment [34]. Pixel intensities, however, have a problem when applied to most practical image alignment tasks. Specif- ically, they violate the brightness constancy assumption, which states that pixel intensities describing a scene shall remain constant under geometric distortion. Our proposed dense bit-planes descriptor offers a solution to this short- coming using a computationally efficient strategy. Contributions: In this work we explore the validity of a descriptor constancy assumption using photometrically invariant descriptors. In particular, we explore the effec- tiveness of one of the simplest and most efficient binary de- scriptors LBP [32] — or the Census Transform [42] — for robust and efficient dense correspondence estimation prob- lems. The concept of linearizing feature descriptors with re- spect to geometric displacement within the LK framework is a relatively new and emerging topic [8, 1]. Hitherto, most of the previously employed descriptors have a con- siderable computational footprint such as HOG [12], dense SIFT [6, 26], and even SURF [5] making them unsuitable for practical use in many vision applications requiring dense correspondences in real-time from high frame rate data. In 1 arXiv:1602.00307v1 [cs.CV] 31 Jan 2016

Upload: others

Post on 30-Jul-2020

13 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Bit-Planes: Dense Subpixel Alignment of Binary Descriptors · Bit-Planes: Dense Subpixel Alignment of Binary Descriptors Hatem Alismail Robotics Institute Carnegie Mellon University

Bit-Planes: Dense Subpixel Alignment of Binary Descriptors

Hatem AlismailRobotics Institute

Carnegie Mellon [email protected]

Brett BrowningRobotics Institute

Carnegie Mellon [email protected]

Simon LuceyRobotics Institute

Carnegie Mellon [email protected]

Abstract

Binary descriptors have been instrumental in the recentevolution of computationally efficient sparse image align-ment algorithms. Increasingly, however, the vision commu-nity is interested in dense image alignment methods, whichare more suitable for estimating correspondences from highframe rate cameras as they do not rely on exhaustive search.However, classic dense alignment approaches are sensitiveto illumination change. In this paper, we propose an easyto implement and low complexity dense binary descriptor,which we refer to as bit-planes, that can be seamlessly inte-grated within a multi-channel Lucas & Kanade framework.This novel approach combines the robustness of binarydescriptors with the speed and accuracy of dense align-ment methods. The approach is demonstrated on a tem-plate tracking problem achieving state-of-the-art robustnessand faster than real-time performance on consumer laptops(400+ fps on a single core Intel i7) and hand-held mobiledevices (100+ fps on an iPad Air 2).

1. IntroductionBinary descriptors such as BRIEF [10] & BRISK [24]

are powerful tools for solving sparse image alignment prob-lems due to their discriminative power, robustness to illumi-nation change, and low complexity [21, 16, 20, 39]. Match-ing binary descriptors is typically performed by exhaustivesearch [7, 23] using the Hamming distance. Exhaustivesearch, however, is inefficient when dense correspondencesare required in real-time [19, 30].

A classical way of speeding up the task of image align-ment is to linearize pixel intensities of an image with re-spect to geometric displacement. The most notable exampleof this strategy can be found in the seminal work of Lucas& Kanade [27]. The Lucas & Kanade (LK) algorithm at-tempts to establish an approximate linear relationship be-tween appearance and geometric displacement. Efficientlinear solvers can then be employed for finding the best geo-metric alignment of the image with respect to a known tem-

plate. The relationship between appearance and geometricdisplacement is seldom linear, so the linearization process istypically repeated until convergence. Due to its importance,numerous extensions and variations upon the LK algorithmhave subsequently been explored in literature [2].

At the heart of the LK algorithm is the notion that an ap-proximate linear relationship between pixel appearance andgeometric displacement can be reliably established. Pixelintensities are not deterministically differentiable with re-spect to geometric displacement. Instead, the linear rela-tionship is established stochastically through spatial finitedifferences whose outputs we refer to as image gradients.The notion of estimating stochastic gradients on image in-tensities has a long and rich history dating back to some ofthe most seminal works of computer vision [29]. Further,it has been well documented that pixel intensities withinnatural images are strongly correlated over small spatial ar-eas further validating the assumed approximate linear rela-tionship between pixel intensities and geometric displace-ment [34]. Pixel intensities, however, have a problem whenapplied to most practical image alignment tasks. Specif-ically, they violate the brightness constancy assumption,which states that pixel intensities describing a scene shallremain constant under geometric distortion. Our proposeddense bit-planes descriptor offers a solution to this short-coming using a computationally efficient strategy.

Contributions: In this work we explore the validity ofa descriptor constancy assumption using photometricallyinvariant descriptors. In particular, we explore the effec-tiveness of one of the simplest and most efficient binary de-scriptors LBP [32] — or the Census Transform [42] — forrobust and efficient dense correspondence estimation prob-lems. The concept of linearizing feature descriptors with re-spect to geometric displacement within the LK frameworkis a relatively new and emerging topic [8, 1]. Hitherto,most of the previously employed descriptors have a con-siderable computational footprint such as HOG [12], denseSIFT [6, 26], and even SURF [5] making them unsuitablefor practical use in many vision applications requiring densecorrespondences in real-time from high frame rate data. In

1

arX

iv:1

602.

0030

7v1

[cs

.CV

] 3

1 Ja

n 20

16

Page 2: Bit-Planes: Dense Subpixel Alignment of Binary Descriptors · Bit-Planes: Dense Subpixel Alignment of Binary Descriptors Hatem Alismail Robotics Institute Carnegie Mellon University

this paper we make the following contributions:• We propose the bit-planes descriptor, an adaptation of

the LBP descriptor, that can be used within the LKframework. Specifically, we propose a multi-channelLK adaptation that allows us to minimize the Ham-ming distance using standard least squares optimiza-tion.• The suitability of our bit-planes descriptor for lin-

earization is explored as a function of geometric dis-placement. We demonstrate that even though thedense bit-planes descriptor is inherently discontinuousit shares the same critical properties enjoyed by pixelintensities, which make them suitable for gradient-based optimization.• Unlike classical dense descriptors such as HOG and

dense SIFT, we demonstrate the efficiency of our bit-planes descriptor on planar target tracking achievingspeeds in excess of 400 fps on a laptop, and in excessof 120 fps on mobile devices. Furthermore, we demon-strate faster and more robust template tracking in com-parison to RANSAC-based algorithms on sparse fea-tures, especially with low- and ambiguously texturedobjects.

2. The Lucas & Kanade Algorithm

In this section we briefly review the LK algorithm in or-der to introduce notation. Let I0 : R2 → R be the tem-plate/reference image. After camera motion with parametervector θ ∈ Rp, we obtain an input/moving image I1. Wedesire to estimate the parameters of motion such that thefollowing quantity is minimized:

E(x;θ) =∑x∈Ω0

‖I0(x)− I1(x′(θ))‖22, (1)

where Ω0 is a subset of pixels in the template image, θ isan initial estimate of parameters and x′(θ) describes thetransformed pixel coordinates given the motion parameters,commonly known as the warping function. By performinga first-order Taylor expansion of Eq. (1) in the vicinity of θ,taking the derivative with respect to the parameter update,and equating it to zero, we arrive at the normal equations:

J(x;θ)>J(x;θ)∆θ = J(x;θ)>e(x;θ), (2)

where J(x;θ) is the matrix of partial derivatives of thewarped image intensities with respect to the motion param-eters evaluated at the current estimate of parameters θ, ande(x;θ) = I0(x) − I1(x′(θ)). Using the chain rule we ob-tain

J(x;θ) =∂I1(x)

∂θ=

∂I

∂x′∂x′

∂θ. (3)

where ∂I1/∂x′ is estimated stochastically through x- and y-finite difference filters, while ∂x′/∂θ is obtained determinis-tically using the closed-form of the warping function. Theoriginal formulation of LK is applicable to a wide varietyof problems. For special warps that satisfy a group re-quirement, however, a more efficient variation is Baker &Matthews’ Inverse Compositional algorithm (IC) [2] whichwe will use in the experimental portion of this paper.

Photometric variation: The classical formulation of LKrelies on the brightness constancy assumption [27], whichis seldom satisfied in real life applications. Techniques toaddress illumination change include: (i) estimating illumi-nation parameters alongside the motion parameters (eitherjointly [4] or in an alternating fashion [38]), (ii) using in-trinsically robust similarity measures, such as Mutual Infor-mation [14, 13], or the normalized correlation [22], and (iii)preprocessing the images to obtain a representation that ismore robust to intensity variations [28, 1, 40].

On the one hand, estimating illumination is sensitive tothe modeling assumptions and increases the dimensionalityof the state and vector, thereby increasing the complexityof the optimization. On the other hand, optimizing robustsimilarity metrics requires general purpose optimizers thatcannot exploit the special structure of least squares prob-lems.

Preprocessing the image does not typically require re-strictive assumptions, and does not affect the dimensionalityof the state vector. Traditionally, preprocessing an image isdone by convolving with filters, or other simple operationssuch as whitening the signal [17, 36]. Densely sampled fea-ture descriptors are another form of preprocessing, whichwe adopt in this work. In particular, we propose the use ofa dense bit-planes descriptor. During evaluation, we showthat our approach exceeds the robustness of algorithms thatexplicitly model illumination as well as methods that relyon robust cost metrics. Furthermore, our method is moreefficient, and simpler to implement. Central to our workis the multi-channel formulation of LK, which we reviewnext.

Multi-channel LK: In this section we present a general-ization of the LK algorithm to accommodate the applica-tion of multi-channel descriptors. Herein, we shall refer tothis generalization as the multi-channel LK algorithm. Letφ0 : R2 → Rd be the d-channel representation of the tem-plate/reference image. Employing a similar notation to theclassical LK algorithm, after camera motion with parametervector θ ∈ Rp, we obtain an input/moving d-channel rep-resentation φ1. To align descriptors using LK we seek tominimize

Eφ(x;θ) =∑x∈Ω0

‖φ0(x)− φ1(x′(θ))‖2. (4)

To linearize Eq. (4) we must obtain an estimate of the Ja-

Page 3: Bit-Planes: Dense Subpixel Alignment of Binary Descriptors · Bit-Planes: Dense Subpixel Alignment of Binary Descriptors Hatem Alismail Robotics Institute Carnegie Mellon University

cobian Jφ(x;θ) = ∂φ/∂θ ∈ Rd×p. Let the value of thej-th channel, as illustrated in Fig. 1, of the multi-channelrepresentation be described as φj(x), where φ(x) =[φ1(x), . . . ,φd(x)]>. The sought Jacobian for each chan-nel in Eq. (4) can be obtained using the chain rule

∂φj1(x)

∂θ=∂φj1∂x′

∂x′

∂θ(5)

for j = 1, . . . , d where ∂φj1/∂x′ is estimated stochastically

through x- and y- finite difference filters on φj1, and ∂x′/∂θis obtained deterministically from the warp function. Themulti-channel d× p Jacobian matrix can then be formed as

Jφ(x;θ) =∂φ1(x)

∂θ=

∂φ11(x)/∂θ

...∂φd

1(x)/∂θ

. (6)

Using this multi-channel linearization all extensions andvariations of the LK algorithm can be extended to differ-ent types of multi-channel descriptors. Recent work hasdemonstrated the utility of multi-channel LK [1, 8] usingclassical dense descriptors such as dense SIFT and HOG.A novel component of this paper is the derivation of a low-complexity dense binary descriptor that can be seamlesslyapplied within the multi-channel LK framework.

3. Dense Binary DescriptorsLocal Binary Patterns (LBP) [32] were among the first

binary descriptors proposed in vision. An almost iden-tical binary representation was independently developedby Zabih & Woodfill under the name: Census Transform(CT) [42], which is still commonly applied in stereo andoptical flow research [39, 20, 31, 35]. LBP is based onthe predicate of pixel comparisons in a small neighborhoodas illustrated in Fig. 2. By definition, the LBP descrip-tor is invariant to monotonic illumination change, which isdesirable in practical image alignment applications. Re-cently, binary descriptor research has progressed signifi-cantly with the development of several high performancedescriptors such as ORB [10] and BRISK [24] among oth-ers [10, 37, 3, 25].

LBP descriptor: When extracting a LBP descriptor abouta pixel position x one obtains,

φ(x) =

8∑i=1

2i−1[I(x) ./ I(x + ∆xi)

], (7)

where ∆xi8i=1 is the set of the eight relative coordinatedisplacements possible within a 3×3 neighborhood aroundthe center pixel location x. Other neighborhood sizes andsampling locations can be used, but we found a 3×3 regionto perform best. The operator ./∈ >,≥, <,≤ is a pixel

Figure 1. An example of the LBP descriptor evaluated on a 3 × 3neighborhood, which results in a 8−channel bit-plane descriptor.Unlike the classic LBP descriptor, the bit-plane descriptor can beemployed within a multi-channel LK framework using a sum ofsquared differences (SSD) cost measure.

8 12 200

56 42 55

128 16 11

(a)

8<42 12<42 200<42

56<42 55<42

128<42 16<42 11<42

(b)

1 1 0

0 0

0 1 1

(c)

Figure 2. The canonical LBP descriptor is obtained by performingpixel comparisons in a fixed order and converting the binary stringto a decimal value. In Fig. 2a the center pixel is in a 3 × 3 neigh-borhood is highlighted, and compared to its neighbors as shown inFig. 2b. Finally, the descriptor is obtained by combining the resultsof each comparison in Fig. 2c into a single scalar descriptor.

comparison/binary test, and the bracket notation representsthe indicator function. We refer to the LBP descriptor de-scribed in Equation Eq. (7) as single-channel since its out-put is a scalar at every pixel position x within the image. Avisual depiction of the single-channel LBP descriptor esti-mation process can be found in Fig. 2.

Bit-planes descriptor: When matching LBP descriptorsit is common practice to employ the Hamming distance.Hamming distance is useful, because it matches LBP de-scriptors in a fashion that is invariant to the ordering of pixelcomparisons within the 3×3 neighborhood. Other distancemetrics such as sum or squared distances (SSD) lack thisdesirable property and are dependent on the ordering spec-ified by ∆xi8i=1. This becomes problematic when em-ploying dense binary descriptors within the multi-channelLK framework due to its inherent dependence on the SSD.

To make dense binary descriptors compatible with LKwe propose the bit-planes descriptor given by

φ(x) =

I(x) ./ I(x + ∆x1)...

I(x) ./ I(x + ∆x8)

. (8)

For each pixel coordinate x in the image, this descriptorproduces an 8-channel vector of binary values 0, 1. No-tably, using the SSD with the multi-channel representationin Eq. (8) between two bit-planes descriptor is equivalent

Page 4: Bit-Planes: Dense Subpixel Alignment of Binary Descriptors · Bit-Planes: Dense Subpixel Alignment of Binary Descriptors Hatem Alismail Robotics Institute Carnegie Mellon University

to the Hamming distance between single-channel LBP de-scriptors. Specifically, the ordering of the pixel compar-isons within the 3 × 3 neighborhood of the bit-planes de-scriptor has no effect on the SSD. An analysis on the choiceof operator ./∈ >,≥, <,≤ is explored in the experi-ments section of this paper.

4. Linearizing Bit-PlanesIn order for our proposed bit-planes descriptor to be ef-

fective within a multi-channel LK framework we first needto establish that there exists an approximate linear relation-ship between the multi-channel bit-planes descriptor andgeometric displacements. Inspecting a visualization of thebit-planes descriptor in Fig. 1, one could be doubtful aboutthe existence of such relationship. Specifically, each chan-nel of the bit-planes descriptor is highly discontinuous (dueto its binary nature). In addition, estimating stochastic gra-dients per channel of the bit-planes descriptor seems strangeas they can take on only three possibilities: −1, 0,+1.SSD cost surface: However, the news is not all gloomy.In Fig. 3 we see the SSD cost surface between a patch withina natural image and shifted versions of itself in the x- andy- directions. This was repeated over a subset of naturalimages with the aggregate result being depicted in Fig. 3.Sub-pixel shifts are entertained here using bi-linear interpo-lation. In Fig. 3b one sees the cost surface for raw pixelintensities, and as expected, we see a quasi-convex cost sur-face surrounding the origin. This quasi-convex surface isimportant with respect to the effectiveness of the LK algo-rithm — as the LK objective relies on a graceful reductionof the SSD cost as a function of geometric displacementfrom ground-truth. In fact the LK algorithm can be inter-preted as attempting to hallucinate a convex quadratic rep-resentation of this SSD cost surface. Interestingly, wheninspecting Fig. 3a we see a similar quasi-convex cost sur-face, which indicates that bit-planes have similar propertiesto raw pixel intensities when it comes to the use of SSD asa measure of dissimilarity.

Linear predictions of bit-planes: Consider a translationaldisplacement warp ∆θ ∈ R2 where we attempt to lin-early predict an image representation R (raw pixels, or bit-planes) in the x- and y- directions,

R(x(0)) +∂R(0)

∂θ∆θ ≈ R(x(∆θ)) . (9)

We employ the neutral notation R to represent either rawpixels I or bit-planes φ. We can define the error of thislinear approximation to be:

ε(∆θ) =∑x∈Ω

‖R(x(0)) +∂R(0)

∂θ∆θ −R(x(∆θ))‖22

(10)

(a) bit-planes(b) raw intensity

Figure 3. Cost surface of our bit-planes descriptor Fig. 3a com-puted over a subset of natural images [41] in comparison to SSDover raw intensity Fig. 3b. Both cost surfaces are suitable for LK.

(a) Bit-planes. (b) Raw Intensity.Figure 4. Assessment of the linearization properties of our pro-posed bit-planes descriptor in terms of the signal-to-noise-ratio(SNR) as a function of translational displacement in x- and y- di-rections. One notes that even though raw pixels are superior, bit-planes offer a sufficient approximation to be used within a multi-channel LK framework. This thesis is the central focus of ourpaper.

The signal-to-noise-ratio (SNR) of the approximation canthen be defined as

SNR(∆θ) = 10 · [log∑x∈Ω

‖R(x(0))‖22− log ε(∆θ)] (11)

In Fig. 4 we depict the SNR of the linearized objective asa function of increasing translational shifts from the trueminima for both raw intensities, and the binary channels inbit-planes. The experiments were carried out in a similarmanner through the use of a subset of natural images andaggregated to form the results in Fig. 4. As expected, theSNR when using binary features is lower then using raw in-tensities. However, it seems that — at least qualitatively —bit-planes gradient estimates provide a good local linear ap-proximation of the objective. Hence, further justifying theuse of the bit-planes descriptor within the LK framework.

5. ExperimentsIn this section we shall attempt to answer a number of

important questions regarding the validity of our dense bit-planes descriptor for robust and efficient image alignment.

Page 5: Bit-Planes: Dense Subpixel Alignment of Binary Descriptors · Bit-Planes: Dense Subpixel Alignment of Binary Descriptors Hatem Alismail Robotics Institute Carnegie Mellon University

Pre-computing descriptors: An obvious question to askwhen considering the application of multi-channel descrip-tors, such as bit-planes, within the LK framework is:whether we can pre-compute the descriptors before warp-ing? Specifically, due to the iterative nature of the LK algo-rithm it becomes computationally expensive to re-computedense descriptors after each image warp. If one can pre-compute the descriptor before warping substantial efficien-cies can be integrated into any LK-based image alignment.

We attempted to answer this question in Fig. 5 wherewe evaluated a number of well-known LK variants [2]: for-ward additive (FA), forward compositional (FC), and in-verse compositional (IC) for the task of image alignmenton natural images. Random warp initializations and appear-ance variation of the form

I1(x) = 255

(αI0(θa(x)) + β

255

)1+γ

(12)

were included, where θa (·) are the 6DOF parameters of anaffine warp, α and β are multiplicative and additive lightingchange terms, and |γ| < 1 is used for gamma correction.We can see that warping feature channels is less accuratethan re-computing the descriptor on the warped image asshown in Fig. 5.

The degree to which warping the feature channels vs. re-computing them affects accuracy depends on the applica-tion and the type of warp. For simple warps such as 2Dtranslation, the relationship between intensity deformationas a function of warp parameters is linear. Hence, approx-imating multi-channel LK by warping the feature channelsis equivalent to re-computing the features on warped imagesas shown in Fig. 6. However, for more complicated warpswhere deformation of image intensities is nonlinearly de-pendent on the warp parameters we expect a pronounceddifference in alignment accuracy. This is because the valueof each descriptor channel might significantly differ aftera nonlinear warp. Overall, it is possible to approximatethe multi-channel objective in Eq. (4) with warping featurechannels depending on the type and accuracy requirementsof the application at hand. In our experiments, we chose torecompute descriptors after every iteration of image warp-ing.

LBP within LK: Employing bit-planes requires the align-ment of eight separate channels as opposed to a single chan-nel when working with raw intensities. In Fig. 8 we dis-cussed the problems of using a LBP descriptor within theLK framework. In particular, the representation is inher-ently sensitive to the ordering of pixel comparisons whenusing a SSD measure of dissimilarity. Using LBP descrip-tors within a LK framework as been reported to performwell [39, 20] given small displacements. However, undermoderate displacements the use of the LBP descriptor inLK introduces biases due to choices of the binary test and

FA FA-1 FC FC-1 IC IC-10.3

1.3

2.4

3.4·10−2

Poin

tRM

S

Figure 5. Recomputing descriptors after image warping showsconsistently better performance than warping feature images whentested with several LK variants. FA: Forward Addition, FC: For-ward Compositional, IC: Inverse Compositional. The suffix ‘-1‘indicates recomputed features on warped images.

FA FA-1 FC FC-1 IC IC-10

0.5

1

Poin

tRM

S

Figure 6. Recomputing descriptors after image warping is equiv-alent to warping feature channels given a linear relationship be-tween the warp parameters and intensities (translation warps). Thereason for reduced IC performance is due to higher noise at thetemplate.

−140 −100 −60 −20 20 60 100 1400

200

400

600

800

1000

1200

Error

Cou

nts

BitPlanes (18.44)

> (31.31)

≥ (35.47)

< (33.23)

≤ (34.19)

Figure 7. Histogram of final intensity errors when using our pro-posed multi-channel bit-planes vs. classical single-channel LBPdescriptors with different comparison operators ./∈ >,≥, <,≤, the RMS is shown in parenthesis.

neighborhood ordering. In Fig. 8 we show the effect of dif-fering binary comparison operators ./∈ >,≥, <,≤ com-pared to our proposed bit-plane descriptor. Our bit-planesdescriptor is unaffected by the ordering. In our experimentswe noticed indistinguishable differences in performance be-tween binary comparison operators when employing the bit-planes descriptor. As a result, we chose to use the > opera-tor for the rest of our experiments.

Real-time template tracking: We evaluate the perfor-mance of bit-planes for a template tracking problems us-ing the benchmark dataset collected by Gauglitz et al. [18].An example of the dataset is shown in Fig. 9. Our planetracker estimates an 8DOF homography using the IC algo-

Page 6: Bit-Planes: Dense Subpixel Alignment of Binary Descriptors · Bit-Planes: Dense Subpixel Alignment of Binary Descriptors Hatem Alismail Robotics Institute Carnegie Mellon University

(a) Template at t = 0 (b) Bit-planes result at t = 50

(c) ./ := > (d) ./ := ≥ (e) ./ := < (f) ./ := ≤Figure 8. Tracking drift when using LBP vs. bit-planes. The bot-tom row shows the result of template tracking using LPB. Theimage magnified for better visualization (compare with Fig. 8b).A quantitative analysis is shown in Fig. 7. Best viewed in color.

Algorithm # parameters # channels

BP (ours) 8 8ECC [15] 8 1DIC-1 [4] 10 1DIC-2 [4] 20 3DF [11] 8 5GC [9] 8 3GM 8 2LK 8 1

Table 1. Algorithms compared in this work. The number of pa-rameters indicates the DOF of the state vector, which is 8 for ahomography in addition to any photometric parameters. We usethe authors’ code for ECC and DIC.

rithm [2]. The template is extracted from the first framein each sequence and is kept fixed throughout the sequenceas we are interested in tracking robustness overtime. Toimprove convergence we use a 3-level pyramid and initial-ize the tracker for subsequent frames using the most recentestimate. We use Gauss-Newton as the optimization algo-rithm, without robust weighting, with a maximum of 100iterations. Tracking terminates early if the relative changein the estimated parameters between iterations drops below1× 10−6, or the relative change in the cost function reduc-tion drops below 1 × 10−5. For small motion, the trackertypically converges in less than 10 iterations. Our imple-mentation runs faster than real time as shown in Table 2.The efficiency is achieved by utilizing SIMD instructionson the CPU that allows us to process 16 pixels at once (or32 pixels with modern AVX instructions). Additionally, theoperations required to compute the LBP descriptor are lim-ited to bit shifts, ORs and ANDs, all of which can be per-formed with low latency.

We compare the performance of our algorithm against

Template area75× 57 150× 115 300× 230 640× 460

Intensity 650 360 140 45Bit-planes 460 170 90 35

Table 2. Plane template tracking runtime in frames per second(FPS) on a single core Intel i7-2460M CPU @ 2.8 Ghz.

a variety of template tracking methods summarized in Ta-ble 1. The algorithms are: the enhance correlation coeffi-cient ECC [15], which serves as an example of an intrinsi-cally robust cost function that is invariant up to an affine il-lumination change. The Dual Inverse Compositional (DIC)algorithm [4], which severs as an example of algorithms thatattempt to estimate illumination parameters. We use twovariations of the DIC: (i) the gain+bias model on grayscaleimages denoted by DIC-1, and (ii) using a full affine light-ing model the makes use of RGB image data denoted byDIC-2. We also compare the performance against a recentlypublished descriptor-based method [11] called DescriptorFields DF. Finally, we include baseline results from rawintensity BF, improved LK with the Gradient ConstraintGC [9], and alignment with the Gradient Magnitude GM.

We report two quantities in the evaluation. First, is thepercentage of successfully tracked frames. A frame is suc-cessfully tracked if the overlap between the estimate and theground truth is greater than 90%. The overlap is computedas o = (A ∩ B)/(A ∪ B), where A is the warped imagegiven each algorithm’s estimate, and B is the warped imagegiven the ground truth. Second, since we are also interestedin subpixel accuracy we show the mean percentage of over-lap across all frames given by m = 1/n

∑ni=1 oi, where n is

the number of frames in each sequence.

Real-time results: Results are compared for three typesof geometric and photometric variations. First is an out ofplane rotation, which induces perspective change as shownin Fig. 9b. Second, is dynamic lighting change where theimage is stationary but a illuminated with nonlinearly vary-ing light source. Finally a static lighting change, where thetransition phase of change in lighting is omitted.

Our evaluation results are shown in Table 3 and inFig. 10. Based on our experimentation, the top performingmethods are the ones that employ a descriptor constancy as-sumption, namely: bit-planes and DF. However, bit-planesis more efficient and it performed significantly better for the‘out of plane rotation’, which induces perspective change inthe image. In fact, all tested algorithms, with the excep-tion of bit-planes, performed poorly with this data. Algo-rithms that use a robust function (ECC) and the ones thatattempt to estimate illumination (DIC) performed well, butfell behind in comparison to descriptor constancy and gra-dient constraint.

Results on mobile devices: We further evaluate the work

Page 7: Bit-Planes: Dense Subpixel Alignment of Binary Descriptors · Bit-Planes: Dense Subpixel Alignment of Binary Descriptors Hatem Alismail Robotics Institute Carnegie Mellon University

br bu mi pa su wd

Out of Plane Rotation

BP 100.0 (99.38) 100.0 (99.51) 87.50 (99.38) 97.92 (99.26) 79.17 (99.57) 93.75 (99.30)ECC 25.00 (96.16) 33.33 (95.85) 25.00 (95.99) 33.33 (96.65) 20.83 (95.52) 18.75 (95.14)DIC-1 25.00 (96.20) 33.33 (95.83) 25.00 (95.98) 33.33 (96.73) 20.83 (95.95) 18.75 (95.46)DIC-2 25.00 (96.22) 35.42 (95.56) 25.00 (95.51) 35.42 (96.42) 25.00 (96.22) 18.75 (95.06)DF 91.67 (99.51) 93.75 (99.44) 79.17 (99.70) 85.42 (99.75) 70.83 (99.60) 83.33 (99.51)GC 100.0 (99.24) 95.83 (99.66) 87.50 (99.52) 93.75 (99.51) 62.50 (98.88) 91.67 (99.34)GM 62.50 (99.86) 83.33 (99.62) 77.08 (99.72) 77.08 (99.81) 58.33 (99.71) 62.50 (99.66)LK 93.75 (99.68) 91.67 (99.70) 83.33 (99.32) 91.67 (99.63) 37.50 (97.64) 66.67 (99.63)

Dynamic Lighting Change

BP 100.0 (98.97) 100.0 (99.08) 100.0 (99.13) 100.0 (98.91) 100.0 (98.98) 100.0 (99.02)ECC 16.33 (98.03) 19.39 (99.00) 100.0 (98.64) 100.0 (98.69) 100.0 (97.30) 67.35 (98.55)DIC-1 100.0 (98.40) 100.0 (99.04) 100.0 (98.77) 100.0 (98.60) 86.87 (96.02) 20.41 (95.36)DIC-2 100.0 (98.39) 100.0 (98.85) 100.0 (98.61) 100.0 (98.58) 85.86 (96.42) 26.53 (97.73)DF 100.0 (99.30) 100.0 (99.08) 100.0 (98.35) 100.0 (98.87) 20.41 (99.36) 68.37 (99.02)GC 17.35 (99.87) 100.0 (99.50) 22.45 (99.84) 18.37 (99.88) 12.24 (99.72) 17.35 (99.84)GM 17.35 (98.99) 19.39 (99.23) 23.47 (99.10) 19.39 (99.08) 0.00 (0.00) 0.00 (0.00)LK 13.27 (99.34) 31.63 (98.26) 18.37 (98.82) 18.37 (99.32) 12.24 (99.16) 16.33 (98.96)

Static lighting change

BP 100.0 (99.76) 100.0 (99.85) 100.0 (99.61) 100.0 (99.85) 100.0 (99.63) 100.0 (99.76)ECC 100.0 (97.33) 100.0 (97.67) 100.0 (97.75) 100.0 (97.41) 100.0 (96.79) 100.0 (97.55)DIC-1 100.0 (97.70) 100.0 (97.77) 100.0 (97.80) 100.0 (97.20) 98.72 (96.58) 89.74 (96.19)DIC-2 100.0 (97.58) 79.49 (97.59) 100.0 (97.07) 100.0 (97.13) 89.74 (95.75) 79.49 (96.38)DF 100.0 (99.68) 100.0 (99.51) 76.92 (99.71) 100.0 (99.77) 74.36 (99.70) 100.0 (99.83)GC 74.36 (99.73) 74.36 (99.84) 48.72 (99.97) 74.36 (99.76) 48.72 (99.74) 51.28 (99.88)GM 48.72 (99.88) 74.36 (99.75) 74.36 (99.66) 74.36 (99.81) 48.72 (99.76) 48.72 (99.83)LK 48.72 (99.80) 74.36 (99.67) 48.72 (99.95) 48.72 (99.93) 48.72 (99.40) 48.72 (99.94)

Table 3. Template tracking evaluation [18]. We show the percentage of successfully tracked frames. In parenthesis we show the averagepercentage of overlap for all successfully tracked frames. The available textures are: br (bricks), bu (building), mi (mission), pa (paris),su (sunset), and wd (wood).

(a) Sudden lighting change and ambiguous texture.

(b) Sudden lighting change and perspective distortion with medium texture.

(c) Sudden lighting change and motion blur with high texture.Figure 11. High frame rate data at 120 Hz captured using an iPhone 5s. Dataset contains different textures under sudden lighting change,low lighting, and motion blur. All data and code will be publicly available for the research community.

Page 8: Bit-Planes: Dense Subpixel Alignment of Binary Descriptors · Bit-Planes: Dense Subpixel Alignment of Binary Descriptors Hatem Alismail Robotics Institute Carnegie Mellon University

(a) Lighting change.

(b) Out-of-plane rotation.Figure 9. Tracking results using the Bricks dataset [18]. The toprow of each figure shows the performance of bit-planes, while thebottom row shows classical intensity-based LK.

0.5 0.6 0.7 0.8 0.9 10

0.2

0.4

0.6

0.8

1

Overlap fraction

Succ

ess

frac

tion

BP ECC DIC-1

DIC-2 DF GC

LK

Figure 10. Fraction of successfully tracked frames as function ofthe overlap area given the ground truth. Bit-planes and DF performbetter than other methods. However, in Table 3 we see that bit-planes’ performance is better with challenging sequences.

template iPad Air 2 iPhone 5ssize BP ORB BRISK BP ORB BRISK

70× 55 123 N/A N/A 50 N/A N/A150× 115 48 15 15 22 13 13311× 230 17 12 14 10 8 11

Table 4. Template tracking running time on ARM architecture us-ing a single CPU core in frames per second (FPS). The bottleneckfor bit-planes is image resizing and warping, which could be alle-viated using the GPU. Results are averaged over three videos ofchallenging data totalling 6446 frames.

on high frame rate data (Slo-mo) using two smart mobiledevices: the iPad Air 2 and the iPhone 5s. In additionto compression artifacts, we made the data more challeng-ing by turning off the lights multiples times to cause sud-den lighting change and low illumination. The videos arerecorded with unsteady hands causing further motion blur.An example of the videos is shown in Fig. 11 featuring anambiguously textured object in Fig. 11a, normal levels of

texture in Fig. 11b as well as higher amount of texture inFig. 11c. The first image in Fig. 11 shows the selected tem-plate, which we hold fixed throughout tracking. The totalnumber of frames from the videos combined is 6447.

We compare the performance of dense tracking using bit-planes with the RANSAC-based tracking by detection usingtwo types of binary descriptors, ORB [33] and BRISK [24].In terms of efficiency, even though our mobile device imple-mentation does not make use of NEON instructions or theGPU, we outperform opencv3’s optimized implementationsof ORB and BRISK by a substantial margin.

More importantly, our approach is more robust. Feature-based tracking failed approximately on 15% of the framesdue to either: (i) inability to detect features under low light,and (ii) RANSAC failure due to imprecise correspondencesunder motion blur.

Perhaps more interestingly, bit-planes is able to performwell and improve efficiency by reducing image resolution.In fact, tracking speed more than doubles when reducingthe template size by half. However, this is not the case withsparse features as memory overhead depends on the numberof extracted keypoints, which we kept fixed at 512. It ispossible to improve the tracking speed of ORB and BRISKby reducing the number of extracted keypoints. However,lowering the number of keypoints must be done carefully asnot to compromise the robustness of the system. We notethat the ability to work with lower resolution is importanton mobile devices, not only for efficiency considerations,but also for power consumption.

Finally, we note that while dense bit-planes tracking pro-duces faster and more accurate results, its main limitation isthe inability to recover if the template is lost due to occlu-sions or significant drift. In such cases, track by detectioncan be of immense value to re-initialize LK-based methodsif needed.

6. ConclusionsIn this work, we proposed a multi-channel representa-

tion that enables nonlinear gradient-based optimization al-gorithms to work with binary features. We arrive at the sur-prising result that binary data is suitable for gradient-basedoptimization, as the local approximation of gradients perchannel is well-approximated with a quadratic form.

We used the multi-channel representation in a Lucas &Kanade (LK) image alignment framework with our pro-posed bit-planes descriptors, which greatly improves the ro-bustness to arbitrary illumination variations without signif-icantly increasing computational demands. In addition, weobtained a precise subpixel localization of binary descrip-tors at speeds faster than real time.

In the context of binary features, least squares minimiza-tion over the multi-channel representation is equivalent tominimizing the Hamming distance. Hence, we are able to

Page 9: Bit-Planes: Dense Subpixel Alignment of Binary Descriptors · Bit-Planes: Dense Subpixel Alignment of Binary Descriptors Hatem Alismail Robotics Institute Carnegie Mellon University

minimize the Hamming distance in a standard least squaresoptimization framework.

References[1] E. Antonakos, J. Alabort-i Medina, G. Tzimiropoulos, and

S. Zafeiriou. Feature-Based Lucas-Kanade and Active Ap-pearance Models. Image Processing, IEEE Transactions on,24(9):2617–2632, Sept 2015. 1, 2, 3

[2] S. Baker and I. Matthews. Lucas-kanade 20 years on: A uni-fying framework. International Journal of Computer Vision,56(3):221–255, 2004. 1, 2, 5

[3] V. Balntas, L. Tang, and K. Mikolajczyk. BOLD - Bi-nary Online Learned Descriptor For Efficient Image Match-ing. The IEEE Conference on Computer Vision and PatternRecognition (CVPR), June 2015. 3

[4] A. Bartoli. Groupwise geometric and photometric direct im-age registration. IEEE Trans. on Pattern Analysis and Ma-chine Intelligence, 30(12):2098–2108, 2008. 2, 6

[5] H. Bay, A. Ess, T. Tuytelaars, and L. V. Gool. Speeded-UpRobust Features (SURF). Computer Vision and Image Un-derstanding, 110(3):346 – 359, 2008. Similarity Matchingin Computer Vision and Multimedia. 1

[6] A. Bosch, A. Zisserman, and X. Muoz. Image Classificationusing Random Forests and Ferns. In Computer Vision, 2007.ICCV 2007. IEEE 11th International Conference on, 2007.1

[7] E. Bostanci. Is Hamming distance only way for match-ing binary image feature descriptors? Electronics Letters,50(11):806–808, May 2014. 1

[8] H. Bristow and S. Lucey. In Defense of Gradient-BasedAlignment on Densely Sampled Sparse Features. In Densecorrespondences in computer vision. Springer, 2014. 1, 3

[9] T. Brox, A. Bruhn, N. Papenberg, and J. Weickert. HighAccuracy Optical Flow Estimation Based on a Theory forWarping. In ECCV, volume 3024 of Lecture Notes in Com-puter Science, pages 25–36. 2004. 6

[10] M. Calonder, V. Lepetit, C. Strecha, and P. Fua. BRIEF: Bi-nary Robust Independent Elementary Features. In K. Dani-ilidis, P. Maragos, and N. Paragios, editors, Computer Vision– ECCV 2010, volume 6314 of Lecture Notes in ComputerScience, pages 778–792. 2010. 1, 3

[11] A. Crivellaro and V. Lepetit. Robust 3D Tracking with De-scriptor Fields. In Conference on Computer Vision and Pat-tern Recognition (CVPR), 2014. 6

[12] N. Dalal and B. Triggs. Histograms of oriented gradients forhuman detection. In Computer Vision and Pattern Recogni-tion, 2005. CVPR 2005. IEEE Computer Society Conferenceon, volume 1, pages 886–893 vol. 1, June 2005. 1

[13] A. Dame and E. Marchand. Accurate real-time tracking us-ing mutual information. In Mixed and Augmented Reality(ISMAR), 2010 9th IEEE International Symposium on, pages47–56, Oct 2010. 2

[14] N. Dowson and R. Bowden. Mutual Information for Lucas-Kanade Tracking (MILK): An Inverse Compositional For-mulation. PAMI, 30(1):180–185, Jan 2008. 2

[15] G. D. Evangelidis and E. Z. Psarakis. Parametric imagealignment using enhanced correlation coefficient maximiza-tion. PAMI, 30(10), 2008. 6

[16] J. Figat, T. Kornuta, and W. Kasprzak. Performance Evalu-ation of Binary Descriptors of Local Features. In ComputerVision and Graphics, pages 187–194. Springer, 2014. 1

[17] D. Fleet and Y. Weiss. Optical Flow Estimation. InN. Paragios, Y. Chen, and O. Faugeras, editors, Handbook ofMathematical Models in Computer Vision, pages 237–257.Springer US, 2006. 2

[18] S. Gauglitz, T. Höllerer, and M. Turk. Evaluation of InterestPoint Detectors and Feature Descriptors for Visual Track-ing. International Journal of Computer Vision, 94(3):335–360, 2011. 5, 7, 8

[19] K. Grauman and R. Fergus. Learning binary hash codes forlarge-scale image search. In Machine learning for computervision, pages 49–87. Springer, 2013. 1

[20] D. Hafner, O. Demetz, and J. Weickert. Why Is the CensusTransform Good for Robust Optic Flow Computation? InScale Space and Variational Methods in Computer Vision,volume 7893. 2013. 1, 3, 5

[21] J. Heinly, E. Dunn, and J.-M. Frahm. Comparative evaluationof binary features. In Computer Vision–ECCV 2012, pages759–773. Springer, 2012. 1

[22] M. Irani and P. Anandan. Robust multi-sensor image align-ment. In Computer Vision, 1998. Sixth International Confer-ence on, pages 959–966, Jan 1998. 2

[23] S. Kaneko, I. Murase, and S. Igarashi. Robust image regis-tration by increment sign correlation. Pattern Recognition,35(10):2223 – 2234, 2002. 1

[24] S. Leutenegger, M. Chli, and R. Siegwart. BRISK: Bi-nary Robust invariant scalable keypoints. In Computer Vi-sion (ICCV), 2011 IEEE International Conference on, pages2548–2555, Nov 2011. 1, 3, 8

[25] G. Levi and T. Hassner. LATCH: Learned Arrangements ofThree Patch Codes. CoRR, abs/1501.03719, 2015. 3

[26] C. Liu, J. Yuen, and A. Torralba. SIFT Flow: Dense Corre-spondence across Scenes and Its Applications. IEEE Trans.Pattern Anal. Mach. Intell., 33(5):978–994, 2011. 1

[27] B. D. Lucas and T. Kanade. An Iterative Image RegistrationTechnique with an Application to Stereo Vision (DARPA). InProc. of the 1981 DARPA Image Understanding Workshop,pages 121–130, April 1981. 1, 2

[28] S. Lucey, R. Navarathna, A. B. Ashraf, and S. Sridharan.Fourier Lucas-Kanade Algorithm. Pattern Analysis and Ma-chine Intelligence, IEEE Transactions on, 35(6), 2013. 2

[29] D. Marr. Vision: A Computational Investigation into the Hu-man Representation and Processing of Visual Information.Henry Holt and Co., Inc., New York, NY, USA, 1982. 1

[30] M. Muja and D. G. Lowe. Fast matching of binary features.In Computer and Robot Vision (CRV), 2012 Ninth Confer-ence on, pages 404–410. IEEE, 2012. 1

[31] T. Müller, C. Rabe, J. Rannacher, U. Franke, and R. Mester.Illumination-Robust Dense Optical Flow Using Census Sig-natures. In Pattern Recognition, volume 6835 of LectureNotes in Computer Science, pages 236–245. 2011. 3

Page 10: Bit-Planes: Dense Subpixel Alignment of Binary Descriptors · Bit-Planes: Dense Subpixel Alignment of Binary Descriptors Hatem Alismail Robotics Institute Carnegie Mellon University

[32] T. Ojala, M. Pietikäinen, and D. Harwood. A comparativestudy of texture measures with classification based on fea-tured distributions. Pattern Recognition, 29:51–59, 1996. 1,3

[33] E. Rublee, V. Rabaud, K. Konolige, and G. Bradski. ORB:An efficient alternative to SIFT or SURF. In Computer Vi-sion (ICCV), 2011 IEEE International Conference on, pages2564–2571, Nov 2011. 8

[34] E. P. Simoncelli and B. A. Olshausen. Natural Image Statis-tics and Neural Representation. Annual Review of Neuro-science, 24:1193–1216, 2001. 1

[35] F. Stein. Efficient Computation of Optical Flow Using theCensus Transform. In Pattern Recognition, volume 3175 ofLecture Notes in Computer Science, pages 79–86. 2004. 3

[36] D. Sun, S. Roth, and M. Black. Secrets of optical flow es-timation and their principles. In Computer Vision and Pat-tern Recognition (CVPR), 2010 IEEE Conference on, pages2432–2439, June 2010. 2

[37] V. L. T. Trzcinski, M. Christoudias and P. Fua. Boosting Bi-nary Keypoint Descriptors. In Computer Vision and PatternRecognition, 2013. 3

[38] G. Tzimiropoulos, J. Alabort-i Medina, S. Zafeiriou, andM. Pantic. Generic Active Appearance Models Revisited. InK. Lee, Y. Matsushita, J. Rehg, and Z. Hu, editors, ComputerVision – ACCV 2012, pages 650–663. 2013. 2

[39] C. Vogel, S. Roth, and K. Schindler. An Evaluation ofData Costs for Optical Flow. In J. Weickert, M. Hein, andB. Schiele, editors, Pattern Recognition, Lecture Notes inComputer Science. 2013. 1, 3, 5

[40] A. Wedel, T. Pock, C. Zach, H. Bischof, and D. Cremers.An Improved Algorithm for TV-L1 Optical Flow. In Statisti-cal and Geometrical Approaches to Visual Motion Analysis,volume 5604 of Lecture Notes in Computer Science. 2009. 2

[41] J. Xiao, J. Hays, K. Ehinger, A. Oliva, A. Torralba, et al. Sundatabase: Large-scale scene recognition from abbey to zoo.In Computer vision and pattern recognition (CVPR), 2010IEEE conference on, pages 3485–3492. IEEE, 2010. 4

[42] R. Zabih and J. Woodfill. Non-parametric local transformsfor computing visual correspondence. In Computer Vision -ECCV’94, pages 151–158. Springer, 1994. 1, 3