retargeting

69
Retargeting Algorithms for Performance-Driven Animation J.P. Lewis Fred Pighin

Upload: kiora

Post on 12-Jan-2016

35 views

Category:

Documents


0 download

DESCRIPTION

Retargeting. Algorithms for Performance-Driven Animation. J.P. Lewis Fred Pighin. "Don't cross the streams.” (Ghostbusters). Why cross-mapping? Different character Imperfect source model Also known as: Performance-driven animation Motion retargeting. Performance Cloning History. - PowerPoint PPT Presentation

TRANSCRIPT

Page 1: Retargeting

RetargetingRetargeting

Algorithms for Performance-Driven Animation

J.P. Lewis Fred Pighin

Page 2: Retargeting

"Don't cross the streams.”(Ghostbusters)"Don't cross the streams.”(Ghostbusters)

• Why cross-mapping?

– Different character

– Imperfect source model

• Also known as:

– Performance-driven animation

– Motion retargeting

Page 3: Retargeting

Performance Cloning HistoryPerformance Cloning History

• L. Williams, Performance-driven Facial Animation, SIGGRAPH 1990

• SimGraphics systems, 1992-present

• LifeFX “Young at Heart” in Siggraph 2000 theater

• J.-Y. Noh and U. Neumann, Expression Cloning, SIGGRAPH 2001

• B. Choe and H. Ko, “Muscle Actuation Basis”, Computer Animation 2001 (used in Korean TV series)

• Wang et. al., EUROGRAPHICS 2003

• Polar Express movie, 2004

Page 4: Retargeting

Face space mappingFace space mapping

Source space Target space

skf

tkf

Page 5: Retargeting

Building a map from correspondencesBuilding a map from correspondences

skf

tkf

tif

sif

sjf

tjf

tlf

slf

Page 6: Retargeting

Building a map from correspondencesBuilding a map from correspondences

skf

tkf

tif

sif

sjf

tjf

tlf

slf

sf tf

( )}..1{),,(| NifffMf ti

si

st ∈=

Page 7: Retargeting

Main issuesMain issues

• How are the corresponding faces created?

• How to build mapping from correspondences?

Page 8: Retargeting

Linear vs. non-linear mappingLinear vs. non-linear mapping

• Linear– Global blend-shape mapping

• B. Choe and H. Ko, Analysis and Synthesis of Facial Expressions with Hand-Generated Muscle Actuation Basis, Computer Animation 2001.

• E. Chuang and C. Bregler, Performance Driven Facial Animation using Blendshape Interpolation, CS-TR-2002-02, Department of Computer Science, Stanford University

• Non-linear– Piece-wise blend-shape mapping

• I. Buck, A. Finkelstein, C. Jacobs, A. Klein, D. H. Salesin, J. Seims, R. Szeliski, and K. Toyama, Performance-driven hand-drawn animation, NPAR 2000.

• J.-Y. Noh and U. Neumann, Expression Cloning, SIGGRAPH 2001.

– Manifold learning

• Y. Wang, X. Huang, C.-S. Lee, S. Zhang, D. Samaras, D. Metaxas, A. Elgammal, and P. Huang, High Resolution Acquisition, Learning, and Transfer of Dynamic 3-D Facial Expressions, Eurographics 2004.

– Single-correspondence mapping

• J.-Y. Noh and U. Neumann, Expression Cloning, SIGGRAPH 2001.

Page 9: Retargeting

Linear MappingLinear Mapping

Page 10: Retargeting

Blendshape mappingBlendshape mappingsib

sjb

tib

tjb

si

n

i

sik

sk bwf ∑

=

=1

,ti

m

i

tik

tk bwf ∑

=

=1

,

(Global) Change of coordinate system

Page 11: Retargeting

Blendshape mappingBlendshape mappingsib

sjb

tib

tjb

si

n

i

sik

sk bwf ∑

=

=1

,ti

m

i

tik

tk bwf ∑

=

=1

,

(Global) Change of coordinate system

Page 12: Retargeting

Blendshapes: definitionBlendshapes: definition

• per-vertex view:

Industry term Math usage

wk Slider values weights

bk Blendshape target

Basis vector

Page 13: Retargeting

Blendshapes: definitionBlendshapes: definition

• Linear algebra of blendshapes:

• per-vertex view:

• global view:

• f : 3n x 1 vector containing all n vertices of the face, in some packing order e.g. xyzxyzxyz....

• B: 3n x m matrix; each column is one of the m blendshapes, using the same packing order.

• w: vector of m weights, animated over time

Page 14: Retargeting

Parallel blendshapesParallel blendshapes

skb

tkb

tibs

ib

sjb

tjb

tlb

slb

sik

tik ww ,, =Use same blending weights:

Page 15: Retargeting

Parallel Model ConstructionParallel Model Construction

• Have similar blendshape controls in source, target models

• Advantage: conceptually simple

• Disadvantage: twice the work (or more!) -- unnecessary!

• Disadvantage: cannot use PCA

Page 16: Retargeting

ImprovementsImprovements

• Adapt generic model to source (Choe et. al.)

• Derive source basis from data (Chuang and Bregler)

• Allow different source, target basis

Page 17: Retargeting

Source model adaptationSource model adaptation

• B. Choe and H. Ko, Analysis and Synthesis of Facial Expressions with Hand-Generated Muscle Actuation Basis, Computer Animation 2001

• Cross-mapping obtained simply by constructing two models with identical controls.

• Localized (delta) blendshape basis inspired by human muscles

• Face performance obtained from motion captured markers

Page 18: Retargeting

Choe and Ko Muscle actuation basisChoe and Ko Muscle actuation basis

• Model points corresponding to markers are identified

• Blendshape weights determined by least-squares fit of model points to markers

• Fit of model face to captured motion is improved with an alternating least squares procedure

Page 19: Retargeting

Choe and Ko Muscle actuation basisChoe and Ko Muscle actuation basis

• Fitting the model to the markers:

• alternate 1), 2)

– 1) solve for weights given markers and corresponding target points

– 2) solve for target points location

• warp the model geometry to fit the final model points using radial basis interpolation.

Page 20: Retargeting

Choe and Ko Muscle actuation basisChoe and Ko Muscle actuation basis

• Fitting the model to the markers:

• Alternate: solve for B, solve for W

• warp the model geometry to fit the final model points using radial basis interpolation.

Page 21: Retargeting

Choe & KoChoe & Ko

QuickTime™ and aSorenson Video 3 decompressorare needed to see this picture.

Page 22: Retargeting

Derive source blendshapes from dataDerive source blendshapes from data

• Principal Component Analysis

• [Chuang and Bregler, 2002]

Page 23: Retargeting

Recorded source motionsRecorded source motions

Source space Target space

Page 24: Retargeting

Source basis estimationSource basis estimation

Source space Target space

sib

sjb

skb

slb

Page 25: Retargeting

Target basis constructionTarget basis construction

Source space Target space

sib

sjb

skb

slb

tib

tjb

tkb

tlb

Page 26: Retargeting

Blendshapes by Principal Component Analysis (PCA)Blendshapes by Principal Component Analysis (PCA)

• Automatic construction of blendshape model (given movement data)

• Advantage: automatic; the most accurate model for a given number of sliders (L2 sense), easy

• Disadvantage: the resulting model is not intuitive

Page 27: Retargeting

Derive source basis from dataDerive source basis from data

• E. Chuang and C. Bregler, Performance Driven Facial Animation using Blendshape Interpolation, CS-TR-2002-02, Department of Computer Science, Stanford University

Page 28: Retargeting

Chuang and Bregler,

Derive source basis from data

Chuang and Bregler,

Derive source basis from data

• Parallel Model Construction approach:

– Source model automatically derived,

– Target manually sculpted

• Using PCA would be unpleasant

Page 29: Retargeting

Chuang and Bregler,

Derive source basis from data

Chuang and Bregler,

Derive source basis from data

• Using PCA would be unpleasant

• Solution: use a subset of the motion capture frames as the blendshape model.

– Subsets of the original motion capture start to “span the space” of that motion capture.

• Two new problems:

1) Which motion capture frames to use?

2) Source blendshape basis is not exact.

Page 30: Retargeting

Chuang and Bregler,

Derive source blendshapes from data

Chuang and Bregler,

Derive source blendshapes from data

Which motion capture frames to use?

Heuristic: for each of the leading PCA vectors,

Pick the mocap frame that have the largest (min,max) projections on that eigenvector.

Page 31: Retargeting

Chuang and Bregler,

Derive source blendshapes from data

Chuang and Bregler,

Derive source blendshapes from data

Two new problems:

1) Which motion capture frames to use?

2) Source blendshape basis is only approximate

Observation:

• Directly reusing weights works poorly when the source model is not exact

– Errors in representing the source can result in large cancelling basis combinations (nearly cancelling positive, negative weights)

– Transferring these cancelling weights to target results in poor shapes

Page 32: Retargeting

Chuang and Bregler,

Derive source blendshapes from data

Chuang and Bregler,

Derive source blendshapes from data

Errors in representing the source can result in large cancelling basis combinations (nearly cancelling positive, negative weights)

+

-

Page 33: Retargeting

Chuang and Bregler,

Derive source blendshapes from data

Chuang and Bregler,

Derive source blendshapes from data

QuickTime™ and aTIFF (LZW) decompressor

are needed to see this picture.

Page 34: Retargeting

• Solution

– Solve for the representation of the source with non-negative least squares. Prohibiting negative weights prevents the cancelling combinations.

– Robust cross mapping.

Chuang and Bregler,

Derive source basis from data

Chuang and Bregler,

Derive source basis from data

Page 35: Retargeting

QuickTime™ and aSorenson Video decompressorare needed to see this picture.

Page 36: Retargeting

Non-corresponding blendshapesNon-corresponding blendshapes

skb

tkb

tibs

ib

sjb

tjb

tlb

?

?

?

Page 37: Retargeting

Is there a “best” blendshape basis?Is there a “best” blendshape basis?

• There are an infinite number of different blendshape models that have exactly the same range of movement. Proof:

• And it’s easy to interconvert between different blendshapes

analogy: what is the best view of a 3D model? Why restrict yourself to only one view??

Page 38: Retargeting

Global blendshape mappingGlobal blendshape mapping

Motivating scenarios:

1) Use PCA for source!

2) Source or target model is pre-existing (e.g. from a library)

Page 39: Retargeting

Global blendshape mappingGlobal blendshape mapping

B,C n vertices, m blendshape targets

Manually create p ≥ m corresponding poses of each model, with weights vk, wk

Page 40: Retargeting

Global blendshape mappingGlobal blendshape mapping

• Gather pose weight vectors vk, wk in columns

of V,W

• Solve for the “cloning matrix” E:

this matrix converts weights for one model to produce the equivalent expression in the other model.

Page 41: Retargeting

Global blendshape mappingGlobal blendshape mapping

• Intentional source-target mismatch: style transfer (person on the right has asymmetric smile)

Page 42: Retargeting

Non-linear MappingNon-linear Mapping

• Piecewise linear

• Manifold learning

• Single-correspondence

Page 43: Retargeting

Piecewise linear mappingPiecewise linear mapping

sik

iik

sk bwf ,

3

1,∑

=

= tik

iik

tk bwf ,

3

1,∑

=

=(Local) Change of coordinate system

sb1

sb2

sb3

tb1tb2

tb3

Page 44: Retargeting

Piecewise linear mappingPiecewise linear mapping

• I. Buck, A. Finkelstein, C. Jacobs, A. Klein, D. H. Salesin, J. Seims, R. Szeliski, and K. Toyama, Performance-driven hand-drawn animation, NPAR 2000

Page 45: Retargeting

Buck et. al.

Piecewise linear mappingBuck et. al.

Piecewise linear mapping

• Project motion onto a 2D space (PCA)

• Construct Delaunay triangulation based on source blendshapes

• Within a triangle use barycentric coordinates as blending weights

Page 46: Retargeting

Buck et. al.

Piecewise linear mappingBuck et. al.

Piecewise linear mapping

• Split face into 3 regions

(eyes, mouth,

rest of the face)

• If frame outside of triangulated area, project onto convex hull

Page 47: Retargeting

Buck et. al.

TriangulationBuck et. al.

Triangulation

Page 48: Retargeting

VideoVideo

• hand_drawn.avi

Page 49: Retargeting

Manifold learningManifold learning

Source space Target space

skf

tkf

Page 50: Retargeting

Manifold learningManifold learning

• Y. Wang, X. Huang, C.-S. Lee, S. Zhang, D. Samaras, D. Metaxas, A. Elgammal, and P. Huang, High Resolution Acquisition, Learning, and Transfer of Dynamic 3-D Facial Expressions, Eurographics 2004.

Page 51: Retargeting

Source and target embeddingsSource and target embeddings

Page 52: Retargeting

Unified embeddingUnified embedding

Page 53: Retargeting

Final mappingFinal mapping

Page 54: Retargeting

Final mappingFinal mapping

Page 55: Retargeting

Wang et. al.

Manifold learningWang et. al.

Manifold learning

Manifold (curve) of smile motion obtained byLocal Linear Embedding (from Wang et. al.)

Page 56: Retargeting

VideoVideo

• final-video2-edit.mov

Page 57: Retargeting

Mapping from a single correspondenceMapping from a single correspondence

sf0 tf0

( )),(| 00tsst fffMf =

Page 58: Retargeting

Mapping from a single correspondenceMapping from a single correspondence

• J.-Y. Noh and U. Neumann, Expression Cloning, SIGGRAPH 2001.

Page 59: Retargeting

Noh et. al.

Two issuesNoh et. al.

Two issues

• Find dense geometric correspondences between the two face models

• Map motion using local geometric deformations from source to target face

Page 60: Retargeting

Noh et. al.

Estimating geometric correspondences

Noh et. al.

Estimating geometric correspondences

• Sparse correspondences through feature detection

• Dense correspondences by interpolating matching features (RBF)

Page 61: Retargeting

Noh et. al.

Local geometric motion transformation

Noh et. al.

Local geometric motion transformation

Source Target

Page 62: Retargeting

Noh et. al.

Animation as displacement from the neutral/rest face

Noh et. al.

Animation as displacement from the neutral/rest face

Source Target

Page 63: Retargeting

Noh et. al.

Local geometric motion transformation

Noh et. al.

Local geometric motion transformation

Source Target

?

Page 64: Retargeting

Noh et. al.

Local change of coordinates system

Noh et. al.

Local change of coordinates system

Source Target

Page 65: Retargeting

Noh et. al.

Local change of coordinates system

Noh et. al.

Local change of coordinates system

Source Target

Page 66: Retargeting

Noh et. al.

Local coordinates systemNoh et. al.

Local coordinates system

• Defined by

– Tangent plane and surface normal

– Scale factor: ratio of bounding boxes containing

all triangles sharing vertex

Page 67: Retargeting

VideoVideo

• monkeyExp.mov

Page 68: Retargeting

ReferencesReferences

• B. Choe and H. Ko, Analysis and Synthesis of Facial Expressions with Hand-Generated Muscle Actuation Basis, Computer Animation 2001.

• I. Buck, A. Finkelstein, C. Jacobs, A. Klein, D. H. Salesin, J. Seims, R. Szeliski, and K. Toyama, Performance-driven hand-drawn animation, NPAR 2000.

• E. Chuang and C. Bregler, Performance Driven Facial Animation using Blendshape Interpolation, CS-TR-2002-02, Department of Computer Science, Stanford University

• J.-Y. Noh and U. Neumann, Expression Cloning, SIGGRAPH 2001.

• Y. Wang, X. Huang, C.-S. Lee, S. Zhang, D. Samaras, D. Metaxas, A. Elgammal, and P. Huang, High Resolution Acquisition, Learning, and Transfer of Dynamic 3-D Facial Expressions, Eurographics 2004.

Page 69: Retargeting

Future workFuture work