on the concentration properties of interacting particle

42
On the Concentration Properties of Interacting Particle Processes Full text available at: http://dx.doi.org/10.1561/2200000026

Upload: others

Post on 30-May-2022

3 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: On the Concentration Properties of Interacting Particle

On the Concentration

Properties of Interacting

Particle Processes

Full text available at: http://dx.doi.org/10.1561/2200000026

Page 2: On the Concentration Properties of Interacting Particle

On the ConcentrationProperties of Interacting

Particle Processes

Pierre Del Moral

INRIA, Universite Bordeaux 1France

Peng Hu

INRIA, Universite Bordeaux 1France

Liming Wu

Institute of Applied MathematicsAMSS, Chinese Academy of Sciences

China

Universite Blaise PascalFrance

Boston – Delft

Full text available at: http://dx.doi.org/10.1561/2200000026

Page 3: On the Concentration Properties of Interacting Particle

Foundations and Trends R© inMachine Learning

Published, sold and distributed by:now Publishers Inc.PO Box 1024Hanover, MA 02339USATel. [email protected]

Outside North America:now Publishers Inc.PO Box 1792600 AD DelftThe NetherlandsTel. +31-6-51115274

The preferred citation for this publication is P. Del Moral, P. Hu, and L. Wu,On the Concentration Properties of Interacting Particle Processes, Foundation and

Trends R© in Machine Learning, vol 3, nos 3–4, pp 225–389, 2010.

ISBN: 978-1-60198-512-5c© 2012 P. Del Moral, P. Hu, and L. Wu

All rights reserved. No part of this publication may be reproduced, stored in a retrievalsystem, or transmitted in any form or by any means, mechanical, photocopying, recordingor otherwise, without prior written permission of the publishers.

Photocopying. In the USA: This journal is registered at the Copyright Clearance Cen-ter, Inc., 222 Rosewood Drive, Danvers, MA 01923. Authorization to photocopy items forinternal or personal use, or the internal or personal use of specific clients, is granted bynow Publishers Inc. for users registered with the Copyright Clearance Center (CCC). The‘services’ for users can be found on the internet at: www.copyright.com

For those organizations that have been granted a photocopy license, a separate systemof payment has been arranged. Authorization does not extend to other kinds of copy-ing, such as that for general distribution, for advertising or promotional purposes, forcreating new collective works, or for resale. In the rest of the world: Permission to pho-tocopy must be obtained from the copyright owner. Please apply to now Publishers Inc.,PO Box 1024, Hanover, MA 02339, USA; Tel. +1-781-871-0245; www.nowpublishers.com;[email protected]

now Publishers Inc. has an exclusive license to publish this material worldwide. Permissionto use this content must be obtained from the copyright license holder. Please apply to nowPublishers, PO Box 179, 2600 AD Delft, The Netherlands, www.nowpublishers.com; e-mail:[email protected]

Full text available at: http://dx.doi.org/10.1561/2200000026

Page 4: On the Concentration Properties of Interacting Particle

Foundations and Trends R© inMachine Learning

Volume 3 Issues 3–4, 2010

Editorial Board

Editor-in-Chief:Michael JordanDepartment of Electrical Engineering and Computer ScienceDepartment of StatisticsUniversity of California, BerkeleyBerkeley, CA 94720-1776

Editors

Peter Bartlett (UC Berkeley)

Yoshua Bengio (Universite de Montreal)

Avrim Blum (Carnegie Mellon University)

Craig Boutilier (University of Toronto)

Stephen Boyd (Stanford University)

Carla Brodley (Tufts University)

Inderjit Dhillon (University of Texas at

Austin)

Jerome Friedman (Stanford University)

Kenji Fukumizu (Institute of Statistical

Mathematics)

Zoubin Ghahramani (Cambridge

University)

David Heckerman (Microsoft Research)

Tom Heskes (Radboud University Nijmegen)

Geoffrey Hinton (University of Toronto)

Aapo Hyvarinen (Helsinki Institute for

Information Technology)

Leslie Pack Kaelbling (MIT)

Michael Kearns (University of

Pennsylvania)

Daphne Koller (Stanford University)

John Lafferty (Carnegie Mellon University)

Michael Littman (Rutgers University)

Gabor Lugosi (Pompeu Fabra University)

David Madigan (Columbia University)

Pascal Massart (Universite de Paris-Sud)

Andrew McCallum (University of

Massachusetts Amherst)

Marina Meila (University of Washington)

Andrew Moore (Carnegie Mellon

University)

John Platt (Microsoft Research)

Luc de Raedt (Albert-Ludwigs Universitaet

Freiburg)

Christian Robert (Universite

Paris-Dauphine)

Sunita Sarawagi (IIT Bombay)

Robert Schapire (Princeton University)

Bernhard Schoelkopf (Max Planck Institute)

Richard Sutton (University of Alberta)

Larry Wasserman (Carnegie Mellon

University)

Bin Yu (UC Berkeley)

Full text available at: http://dx.doi.org/10.1561/2200000026

Page 5: On the Concentration Properties of Interacting Particle

Editorial Scope

Foundations and Trends R© in Machine Learning will publish sur-vey and tutorial articles in the following topics:

• Adaptive control and signalprocessing

• Applications and case studies

• Behavioral, cognitive and neurallearning

• Bayesian learning

• Classification and prediction

• Clustering

• Data mining

• Dimensionality reduction

• Evaluation

• Game theoretic learning

• Graphical models

• Independent component analysis

• Inductive logic programming

• Kernel methods

• Markov chain Monte Carlo

• Model choice

• Nonparametric methods

• Online learning

• Optimization

• Reinforcement learning

• Relational learning

• Robustness

• Spectral methods

• Statistical learning theory

• Variational inference

• Visualization

Information for LibrariansFoundations and Trends R© in Machine Learning, 2010, Volume 3, 4 issues.ISSN paper version 1935-8237. ISSN online version 1935-8245. Also availableas a combined paper and online subscription.

Full text available at: http://dx.doi.org/10.1561/2200000026

Page 6: On the Concentration Properties of Interacting Particle

Foundations and Trends R© inMachine Learning

Vol. 3, Nos. 3–4 (2010) 225–389c© 2012 P. Del Moral, P. Hu, and L. Wu

DOI: 10.1561/2200000026

On the Concentration Propertiesof Interacting Particle Processes

Pierre Del Moral1, Peng Hu2 and Liming Wu3

1 INRIA, Bordeaux-Sud Ouest Center, INRIA, Bordeaux-Sud Ouest Center& Bordeaux Mathematical Institute, Universite Bordeaux 1, 351, Cours dela Liberation, Talence, 33405, France, [email protected]

2 INRIA, Bordeaux-Sud Ouest Center, INRIA, Bordeaux-Sud Ouest Center& Bordeaux Mathematical Institute, Universite Bordeaux 1, 351, Cours dela Liberation, Talence, 33405, France, [email protected]

3 Academy of Mathematics and Systems Science & Laboratoire deMathematiques, Universite Blaise Pascal, 63177 AUBIERE, France,Institute of Applied Mathematics, AMSS, CAS, Siyuan Building, Beijing,China, & Universite Blaise Pascal, Aubiere 63177, France,[email protected], [email protected]

Abstract

This monograph presents some new concentration inequalities forFeynman-Kac particle processes. We analyze different types of stochas-tic particle models, including particle profile occupation measures,genealogical tree based evolution models, particle free energies, as wellas backward Markov chain particle models. We illustrate these resultswith a series of topics related to computational physics and biology,stochastic optimization, signal processing and Bayesian statistics, andmany other probabilistic machine learning algorithms. Special empha-sis is given to the stochastic modeling, and to the quantitative perfor-mance analysis of a series of advanced Monte Carlo methods, includingparticle filters, genetic type island models, Markov bridge models, andinteracting particle Markov chain Monte Carlo methodologies.

Full text available at: http://dx.doi.org/10.1561/2200000026

Page 7: On the Concentration Properties of Interacting Particle

Full text available at: http://dx.doi.org/10.1561/2200000026

Page 8: On the Concentration Properties of Interacting Particle

Contents

1 Stochastic Particle Methods 1

1.1 Introduction 11.2 A Brief Review on Particle Algorithms 61.3 Feynman-Kac Path Integrals 91.4 Interacting Particle Systems 131.5 Basic Notation 22

2 Some Application Domains 27

2.1 Introduction 272.2 Boltzmann-Gibbs Measures 292.3 Rare Event Analysis 382.4 Sensitivity Measures 402.5 Partial Observation Models 442.6 Markov Chain Restriction Models 522.7 Particle Absorption Models 53

3 Feynman-Kac Semigroup Analysis 57

3.1 Introduction 573.2 Historical and Backward Models 583.3 Semigroup Models 593.4 Stability Properties 613.5 Mean Field Particle Models 703.6 Local Sampling Errors 72

ix

Full text available at: http://dx.doi.org/10.1561/2200000026

Page 9: On the Concentration Properties of Interacting Particle

4 Empirical Processes 75

4.1 Introduction 754.2 Statement of the Main Results 784.3 A Reminder on Orlicz’ Norms 834.4 Marginal Inequalities 924.5 Maximal Inequalities 954.6 Cramer-Chernov Inequalities 974.7 Perturbation Analysis 105

5 Interacting Empirical Processes 111

5.1 Introduction 1115.2 Finite Marginal Models 1135.3 Empirical Processes 1155.4 Covering Numbers and Entropy Methods 119

6 Feynman-Kac Particle Processes 123

6.1 Introduction 1236.2 First-Order Expansions 1246.3 A Stochastic Perturbation Analysis 1256.4 Concentration Inequalities 1296.5 Particle Free Energy Models 1346.6 Backward Particle Markov Models 143

References 159

Full text available at: http://dx.doi.org/10.1561/2200000026

Page 10: On the Concentration Properties of Interacting Particle

1

Stochastic Particle Methods

1.1 Introduction

Stochastic particle methods have come to play a significant role inapplied probability, numerical physics, Bayesian statistics, probabilisticmachine learning, and engineering sciences.

They are increasingly used to solve a variety of problems, includ-ing nonlinear filtering equations, data assimilation problems, rare eventsampling, hidden Markov chain parameter estimation, stochastic con-trol problems and financial mathematics. To name a few, They arealso used in computational physics for free energy computations, andSchrodinger operator’s ground states estimation problems, as well asin computational chemistry for sampling the conformation of polymersin a given solvent.

To illustrate these methods, we start with a classical filtering exam-ple. We consider a Markov chain Xk taking values in Rd, with priortransitions given by

P(Xk ∈ dxk | Xk−1 = xk−1) = pk(xk|xk−1) dxk, (1.1)

1

Full text available at: http://dx.doi.org/10.1561/2200000026

Page 11: On the Concentration Properties of Interacting Particle

2 Stochastic Particle Methods

Using some slight abuse of Bayesian notation, the observations Yk areRd′-valued random variables defined in terms of the likelihood functions

P(Yk ∈ dyk | Xk = xk) = pk(yk|xk) dyk, (1.2)

In the above display, dxk and dyk stand for the Lebesgue measuresin Rd and Rd′ . To compute the conditional distribution of the signalpath sequence (X0, . . . ,Xn), given the observations (Y0, . . . ,Yn), we canuse the genealogical tree model associated with a genetic type inter-acting particle model. This genetic algorithm is defined with mutationtransitions according to 1.1, and proportional selections with regard to(w.r.t.) the fitness functions 1.2. The occupation measures of the cor-responding genealogical tree provides an approximation of the desiredconditional distributions of the signal. More generally, for any functionf on the path space we have

limN↑∞

1N

N∑1

f(linen(i)) = E(f(X0, . . . ,Xn)|Y0 = y0, . . . ,Yn = yn) (1.3)

where linen(i) stands for the i−th ancestral line of the genealogicaltree, at time n.

More refined particle filters can be designed, including fixed param-eter estimates in hidden Markov chain models, unbiased particle esti-mates of the density of the observation sequence, and backward smooth-ing models based on complete ancestral trees. Section 2 presents a morerigorous and detailed discussion on these topics.

Rigorous understanding of these new particle Monte Carlo method-ologies leads to fascinating mathematics related to Feynman-Kac pathintegral theory and their interacting particle interpretations [17, 20, 38].In the last two decades, this line of research has been developed byusing methods from stochastic analysis of interacting particle systemsand nonlinear semigroup models in distribution spaces, but it has alsogenerated difficult questions that cannot be addressed without devel-oping new mathematical tools.

Let us survey some of the important challenges that arise.For numerical applications, it is essential to obtain nonasymptotic

quantitative information on the convergence of the algorithms. Forinstance, in the filtering problem presented at beginning of this section,

Full text available at: http://dx.doi.org/10.1561/2200000026

Page 12: On the Concentration Properties of Interacting Particle

1.1 Introduction 3

it is important to quantify the performance of the empirical particleestimate in 1.3. Asymptotic theory, including central limit theorems,moderate deviations, and large deviations principles have clearly lim-ited practical values. An overview of these asymptotic results in thecontext of mean field and Feynman-Kac particle models can be foundin the series of articles [13, 28, 29, 33, 41, 43].

Furthermore, when solving a given concrete problem, it is impor-tant to obtain explicit nonasymptotic error bounds estimates to ensurethat the stochastic algorithm is provably correct. While non asymptoticpropagation of chaos results provides some insights on the bias prop-erties of these models, it rarely provides useful effective convergencerates.

Last but not least, it is essential to analyze the robustness prop-erties, and more particularly the uniform performance of particlealgorithms w.r.t. the time horizon. By construction, these importantquestions are intimately related to the stability properties of com-plex nonlinear Markov chain semigroups associated with the limit-ing measure valued process. In the filtering example illustrated in thissection, the limiting measure valued process is given by the so-callednonlinear filtering equation. In this context, the stability property ofthese equations ensures that the optimal filter will correct any erroneousinitial conditions. This line of thought has been further developed inthe articles [13, 31, 38, 40], and in the books [17, 20].

Without any doubt, one of the most powerful mathematical toolsto analyze the deviations of Monte Carlo based approximations is thetheory of empirical processes and measure concentration theory. In thelast two decades, these new tools have become one of the most impor-tant steps forward in infinite dimensional stochastic analysis, advancedmachine learning techniques, as well as in the development of a statis-tical non asymptotic theory.

In recent years, much effort has been devoted to describing thebehavior of the supremum norm of empirical functionals around themean value of the norm. For an overview of these subjects, we referthe reader to the seminal books of Pollard [81], Van der Vaart andWellner [93], Ledoux and Talagrand [72], the remarkable articles byGine [56], Ledoux [70, 71], and Talagrand [90, 91, 92], and the more

Full text available at: http://dx.doi.org/10.1561/2200000026

Page 13: On the Concentration Properties of Interacting Particle

4 Stochastic Particle Methods

recent article by Adamczak [1]. The best constants in Talagrand’s con-centration inequalities were obtained by Klein and Rio [67]. In thisarticle, the authors proved the functional version of Bennett’s andBernstein’s inequalities for sums of independent random variables.

Two main difficulties we encountered in applying these concentra-tion inequalities to interacting particle models are of different order:

First, all of the concentration inequalities developed in the literatureon empirical processes still involve the mean value of the supremumnorm empirical functionals. In practical situations, these tail styleinequalities can only be used if we have some precise information on themagnitude of the mean value of the supremum norm of the functionals.

On the other hand, the range of application of the theory ofempirical processes and measure concentration theory is restrictedto independent random samples, or equivalently product measures,and more recently to mixing Markov chain models. In the reverseangle, stochastic particle techniques are not based on fully independentsequences, nor on Markov chain Monte Carlo principles, but on inter-acting particle samples combined with complex nonlinear Markov chainsemigroups. More precisely, in addition to the fact that particle modelsare built sequentially using conditionally independent random samples,their respective conditional distributions are still random. Also, in anonlinear way, they strongly depend on the occupation measure of thecurrent population.

In summary, the concentration analysis of interacting particle pro-cesses requires the development of new stochastic perturbation styletechniques to control the interaction propagation and the degree ofindependence between the samples.

Del Moral and Ledoux [36] extend empirical processes theory toparticle models. In this work, the authors proved Glivenko-Cantelli andDonsker theorems under entropy conditions, as well as nonasymptoticexponential bounds for Vapnik-Cervonenkis classes of sets or functions.Nevertheless, in practical situations these non asymptotic results tendto be a little disappointing, with very poor constants that degeneratew.r.t. the time horizon.

The second most important result on the concentration propertiesof the mean field particle model is found in [40]. This article is only

Full text available at: http://dx.doi.org/10.1561/2200000026

Page 14: On the Concentration Properties of Interacting Particle

1.1 Introduction 5

concerned with the finite marginal model. The authors generalize theclassical Hoeffding, Bernstein and Bennett inequalities for independentrandom sequences to interacting particle systems.

In this monograph, we survey some of these results, and we pro-vide new concentration inequalities for interacting empirical processes.We emphasize that this review does not give a comprehensive treat-ment of the theory of interacting empirical processes. To name a fewmissing topics, we do not discuss large deviation principles w.r.t. thestrong τ -topology, Donsker type fluctuation theorems, moderate devi-ation principles, and continuous time models. The first two topics aredeveloped [17], the third one is developed in [32], the last one is stillan open research subject.

Here, we emphasize a single stochastic perturbation method, withsecond-order expansion entering the stability properties of the limitingFeynman-Kac semigroups. The concentration results attained are prob-ably not the best possible of their kind. We have chosen to strive for justenough generality to derive useful and uniform concentration inequal-ities w.r.t. the time horizon, without having to impose complex andoften unnatural regularity conditions to squeeze them into the generaltheory of empirical processes.

Some of the results are borrowed from [40], and many others arenew. This monograph should be complemented with the books andarticles [17, 20, 31, 44]. A very basic knowledge in statistics and machinelearning theory will be useful, but not necessary. Good backgrounds inMarkov chain theory and in stochastic semigroup analysis are necessary.

We have done our best to give a self-contained presentation, withdetailed proofs. However, we assume some familiarity with Feynman-Kac models, and basic facts on the theory of Markov chains on abstractstate spaces. Only in subsection 4.6.1, have we skipped the proof ofsome tools from convex analysis. We hope that the essential ideas arestill accessible to the readers.

It is clearly not the scope of this monograph to give an exhaus-tive list of references to articles in computational physics, engineeringsciences, and machine learning, presenting heuristic-like particle algo-rithms to solve a specific estimation problem. With a few exceptions, wehave only provided references to articles with rigorous and well founded

Full text available at: http://dx.doi.org/10.1561/2200000026

Page 15: On the Concentration Properties of Interacting Particle

6 Stochastic Particle Methods

mathematical treatments on particle models. We apologize in advancefor possible errors, or for references that have been omitted due to thelack of accurate information.

This monograph grew from series of lectures the first author gavein the Computer Science and Communications Research Unit, of theUniversity of Luxembourg in February and March 2011. They werereworked, with the addition of new material on the concentration ofempirical processes for a course given at the Sino-French SummerInstitute in Stochastic Modeling and Applications (CNRS-NSFC JointInstitute of Mathematics), held at the Academy of Mathematics andSystem Science, Beijing, in June 2011. The Summer Institute wasably organized by Fuzhou Gong, Ying Jiao, Gilles Pages, and MingyuXu, and the members of the scientific committee, including Nicole ElKaroui, Zhiming Ma, Shige Peng, Liming Wu, Jia-An Yan, and NizarTouzi. The first author is grateful to them for giving to him the oppor-tunity to experiment on a receptive audience with material not entirelypolished.

In reworking the lectures, we have tried to resist the urge to pushthe analysis to general classes of mean field particle models, in the spiritof the recent joint article with E. Rio [40]. Our principal objective hasbeen to develop just enough analysis to handle four types of Feynman-Kac interacting particle processes, namely, genetic dynamic populationmodels, genealogical tree based algorithms, particle free energies, aswell as backward Markov chain particle models. These applicationmodels do not exhaust the possible uses of the theory developed inthese lectures.

1.2 A Brief Review on Particle Algorithms

Stochastic particle methods belong to the class of Monte Carlomethods. They can be thought of as a universal particle methodologyfor sampling complex distributions in highly dimensional state spaces.

We can distinguish two different classes of models, namely, diffu-sion type interacting processes, and interacting jump particle models.Feynman-Kac particle methods belongs to the second class of models,with rejection-recycling jump type interaction mechanisms. In contrast

Full text available at: http://dx.doi.org/10.1561/2200000026

Page 16: On the Concentration Properties of Interacting Particle

1.2 A Brief Review on Particle Algorithms 7

to conventional acceptance-rejection type techniques, Feynman-Kacparticle methods are equipped with an adaptive and interacting recy-cling strategy.

The common central feature of all the Monte Carlo particle method-ologies developed so far is to solve discrete generation, or continu-ous time integro-differential equations in distribution spaces. The firstheuristic-like description of these probabilistic techniques in mathemat-ical physics goes back to the Los Alamos report [49], and the article byEverett and Ulam in 1948 [48], and the short article by Metropolis andUlam [79], published in 1949.

In some instances, the flow of measures is dictated by the problem athand. In advanced signal processing, the conditional distributions of thesignal, given partial and noisy observations, are given by the so-callednonlinear filtering equation in distribution space (see for instance [15,16, 17, 20, 38], and references therein).

Free energies and Schrodinger operator’s ground states are given bythe quasi-invariant distribution of a Feynman-Kac conditional distri-bution flow of non absorbed particles in absorbing media. We refer thereader to the articles by Cances, Jourdain and Lelievre [5], El Makrini,Jourdain and Lelievre [46], Rousset [85], the pair of articles of DelMoral with Miclo [38, 39], with Doucet [19], and the book [17], and thereferences therein.

In mathematical biology, branching processes and infinite popula-tion models are also expressed by nonlinear parabolic type integro-differential equations. Further details on this subject can be found inthe articles by Dawson and his co-authors [11, 12, 14], the works ofDynkin [45], and Le Gall [69], and more particularly the seminal bookof Ethier and Kurtz [47], and the pioneering article by Feller [50].

In other instances, we formulate a given estimation problemin terms of a sequence of distributions with increasing complexityon state space models with increasing dimension. These stochasticevolutions can be related to decreasing temperature schedules inBoltzmann-Gibbs measures, multilevel decompositions for rare eventexcursion models on critical level sets, decreasing subsets strategiesfor sampling tail style distributions, and many other sequential

Full text available at: http://dx.doi.org/10.1561/2200000026

Page 17: On the Concentration Properties of Interacting Particle

8 Stochastic Particle Methods

importance sampling plans. For a more thorough discussion on thesemodels we refer the reader to [21].

From a purely probabilistic point of view, any flow of probabilitymeasures can be interpreted as the evolution of the laws of the randomstates of a Markov process. In contrast to conventional Markov chainmodels, the Markov transitions of these chains may depend on the dis-tribution of the current random state. The mathematical foundationsof these discrete generation models began in 1996 in [15] within thecontext of nonlinear filtering problems. Further analysis was developedin [38]. For a more thorough discussion on the origin and the perfor-mance analysis of these discrete generation models, we also refer thereader to the book [17], and the joint articles Del Moral with Guionnet[28, 29, 30, 31], and with Kouritzin [35].

The continuous time version of these nonlinear type Markov chainmodels take their origins from the 1960s, with the development of fluidmechanisms and statistical physics. We refer the reader to the pio-neering works of McKean [61, 63], as well as the more recent treat-ments by Bellomo and Pulvirenti [3, 4], the series of articles by Grahamand Meleard on interacting jump models [58, 59, 82], the articles byMeleard on Boltzmann equations [75, 76, 77, 78], and the lecture notesof Sznitman [89], and references therein.

In contrast to conventional Markov chain Monte Carlo techniques,these McKean type nonlinear Markov chain models can be thought ofas perfect importance sampling strategies, in the sense that the desiredtarget measures coincide at any time step with the law of the randomstates of a Markov chain. Unfortunately, as we mentioned above, thetransitions of these chains depend on the distributions of their randomstates. Thus, they cannot be sampled without an additional level ofapproximation. One natural solution is to use a mean field particleinterpretation model. These stochastic techniques belong to the class ofstochastic population models, with free evolutions mechanisms, coupledwith branching and/or adaptive interacting jumps. At any time step,the occupation measure of the population of individuals approximatesthe solution of the nonlinear equation, when the size of the systemtends to ∞.

Full text available at: http://dx.doi.org/10.1561/2200000026

Page 18: On the Concentration Properties of Interacting Particle

1.3 Feynman-Kac Path Integrals 9

In genetic algorithms and sequential Monte Carlo literature, thereference free evolution model is interpreted as a reference sequenceof twisted Markov chain samplers. These chains are used to performthe mutation/proposal transitions. As in conventional Markov chainMonte Carlo methods, the interacting jumps are interpreted as anacceptance-rejection transition, equipped with sophisticated interact-ing and adaptive recycling mechanism. In Bayesian statistics and engi-neering sciences, the resulting adaptive particle sampling model is oftencoined as a sequential Monte Carlo algorithm, genetic procedure, orsimply a Sampling Importance Resampling method, mainly because itis based on importance sampling plans and online approximations of aflow of probability measures.

Since the 1960s, the adaptive particle recycling strategy has alsobeen associated, in biology and engineering science, with severalheuristic-like paradigms, with a proliferation of botanical names,depending on the application area in which they are considered: boot-strapping, switching, replenishing, pruning, enrichment, cloning, recon-figurations, resampling, rejuvenation, acceptance/rejection, spawning.

Of course, the idea of duplicating online better-fitted individualsand moving them one step forward to explore state-space regions is thebasis of various stochastic search algorithms, such as:

Particle and bootstrap filters, Rao-Blackwell particle filters, sequen-tial Monte Carlo methods, sequentially interacting Markov chain MonteCarlo methods, genetic type search algorithms, Gibbs cloning searchtechniques, interacting simulated annealing algorithms, sampling-importance resampling methods, quantum Monte Carlo walkers,adaptive population Monte Carlo sampling models, and many othersevolutionary type Monte Carlo methods.

For a more detailed discussion on these models, with precise refer-ences, we refer the reader to the three books [17, 20, 44].

1.3 Feynman-Kac Path Integrals

Feynman-Kac measures represent the distribution of the paths ofa Markov process, weighted by a collection of potential functions.These functional models encapsulate traditional changes of probability

Full text available at: http://dx.doi.org/10.1561/2200000026

Page 19: On the Concentration Properties of Interacting Particle

10 Stochastic Particle Methods

measures, commonly used in importance sampling, posterior distribu-tions in Bayesian statistics, and the optimal filter in nonlinear filteringproblems.

These stochastic models are defined in terms of only two ingredients:A Markov chain Xn, with Markov transition Mn on some measur-

able state spaces (En,En) with initial distribution η0, and a sequenceof (0,1]-valued potential functions Gn on the set En.

The Feynman-Kac path measure associated with the pairs (Mn,Gn)is the probability measure Qn on the product state space

En := (E0 × . . . × En)

defined by the following formula

dQn :=1Zn

∏0≤p<n

Gp(Xp)

dPn (1.4)

where Zn is a normalizing constant and Pn is the distribution of therandom paths

Xn = (X0, . . . ,Xn) ∈ En

of the Markov process Xp from the origin p = 0 with initial distributionη0, up to the current time p = n. We also denote by

Γn = ZnQn (1.5)

its unnormalized version.The prototype model we have in mind is the traditional particle

absorbed Markov chain model

Xcn ∈ Ecn := En ∪ {c}

absorption ∼(1−Gn)−−−−−−−−−−−−−→ Xcnexploration ∼Mn+1−−−−−−−−−−−−→ Xc

n+1.

(1.6)

The chain Xcn starts at some initial state Xc

0 randomly chosen withdistribution η0. During the absorption stage, we set Xc

n = Xcn with

probability Gn(Xn), otherwise we put the particle in an auxiliary ceme-tery state Xc

n = c. When the particle Xcn is still alive (that is, if we have

Xcn ∈ En), it performs an elementary move Xc

n ; Xcn+1 according to the

Full text available at: http://dx.doi.org/10.1561/2200000026

Page 20: On the Concentration Properties of Interacting Particle

1.3 Feynman-Kac Path Integrals 11

Markov transition Mn+1. Otherwise, the particle is absorbed and weset Xc

p = Xcp = c, for any time p > n.

If we let T be the first time Xcn = c, then we have the Feynman-Kac

representation formulae

Qn = Law((Xc0, . . . ,X

cn) | T ≥ n) and Zn = Proba(T ≥ n).

For a more thorough discussion on the variety of application domainsof Feynman-Kac models, we refer the reader to Section 2.

We also denote by ηn and γn, the n-th time marginal of Qn and Γn.It is a simple exercise to check that

γn = γn−1Qn and ηn+1 = Φn+1(ηn) := ΨGn(ηn)Mn+1 (1.7)

with the positive integral operator

Qn(x,dy) = Gn−1(x) Mn(x,dy)

and the Boltzmann-Gibbs transformation

ΨGn(ηn)(dx) =1

ηn(Gn)Gn(x)ηn(dx). (1.8)

In addition, the normalizing constants Zn can be expressed in termsof the flow of marginal measures ηp, from the origin p = 0 up to thecurrent time n, with the following multiplicative formulae:

Zn := γn( ) = E

∏0≤p<n

Gp(Xp)

=∏

0≤p<nηp(Gp). (1.9)

This multiplicative formula is easily checked using the induction

γn+1(1) = γn(Gn) = ηn(Gn)γn(1).

The abstract formulae discussed above are more general than theymay appear. For instance, they can be used to analyze, without furtherwork, path spaces models, including historical processes or transitionspace models, as well as finite excursion models. These functionalmodels also encapsulate quenched Feynman-Kac models, Browniantype bridges and linear Gaussian Markov chains conditioned on startingand end points.

Full text available at: http://dx.doi.org/10.1561/2200000026

Page 21: On the Concentration Properties of Interacting Particle

12 Stochastic Particle Methods

For a more thorough discussion on these path space models, we referthe reader to subsections 2.4 and 2.6, Chapters 11–12 in the book [17],as well as to the Section 2, in this monograph.

When the Markov transitions Mn are absolutely continuous withrespect to some measures λn on En, and for any (x,y) ∈ (En−1 × En)we have

Hn(x,y) :=dMn(x, .)dλn

(y) > 0. (1.10)

We also have the following backward formula

Qn(d(x0, . . . ,xn)) = ηn(dxn)n∏q=1

Mq,ηq−1(xq,dxq−1) (1.11)

with the collection of Markov transitions defined by

Mn+1,ηn(x,dy) ∝ Gn(y)Hn+1(y,x)ηn(dy). (1.12)

The proof of this formula is postponed to subsection 3.2.Before launching into the description of the particle approximation

of these models, we end this subsection with some connexions betweendiscrete generation Feynman-Kac models and more conventional con-tinuous time models arising in physics and scientific computing.

The Feynman-Kac models presented above play a central role inthe numerical analysis of certain partial differential equations, offeringa natural way to solve these functional integral models by simulat-ing random paths of stochastic processes. These Feynman-Kac modelswere originally presented by Mark Kac in 1949 [66] for continuous timeprocesses.

These continuous time models are used in molecular chemistryand computational physics to calculate the ground state energy ofsome Hamiltonian operators associated with some potential functionV describing the energy of a molecular configuration (see, for instance,[5, 17, 39, 46, 85], and references therein). To better connect thesepartial differential equation models with (1.4), let us assume thatMn(xn−1,dxn) is the Markov probability transitionXn = xn ; Xn+1 =xn+1 coming from a discretization in time Xn = X ′tn of a continuoustime E-valued Markov process X ′t on a given time mesh (tn)n≥0 with

Full text available at: http://dx.doi.org/10.1561/2200000026

Page 22: On the Concentration Properties of Interacting Particle

1.4 Interacting Particle Systems 13

a given time step (tn − tn−1) = ∆t. For potential functions of the formGn = e−V∆t, the measures Qn '∆t→0 Qtn represent the time discretiza-tion of the following distribution:

dQt =1Zt

exp(−∫ t

0V (X ′s)ds

)dPX

′t

where PX′t stands for the distribution of the random paths (X ′s)0≤s≤twith a given infinitesimal generator L. The marginal distributions γt attime t of the unnormalized measures Zt dQt are the solution of the so-called imaginary time Schrodinger equation, given in weak formulationon sufficiently regular function f by the following integro–differentialequation

d

dtγt(f) := γt(LV (f)) with LV = L − V.

The errors introduced by the discretization of the time are well under-stood for regular models; we refer the interested reader to [34, 42, 68, 80]in the context of nonlinear filtering.

1.4 Interacting Particle Systems

Our aim here is to design an interacting particle approximation of theFeynman-Kac measures introduced in the previous subsection. Theseparticle methods can be interpreted in different ways, depending on theapplication domain in which they are considered.

In the filtering example presented at the beginning of this mono-graph, these particle algorithms can be seen as a stochastic adaptivefixed approximation of the filtering equations. From a purely statisti-cal point of view, these algorithms can also be seen as a sophisticatedacceptance-rejection technique with an interacting recycling transition.

The particle model is defined as follows:We start with a population of N candidate possible solutions

(ξ10 , . . . , ξ

N0 ) randomly chosen w.r.t. some distribution η0.

The coordinates ξi0 are also called individuals or phenotypes, with1 ≤ N . The random evolution of the particles is decomposed into twomain steps : the free exploration and the adaptive selection transition.

Full text available at: http://dx.doi.org/10.1561/2200000026

Page 23: On the Concentration Properties of Interacting Particle

14 Stochastic Particle Methods

During the updating-selection stage, multiple individuals in the cur-rent population (ξ1

n, . . . , ξNn ) at time n ∈ N are stochastically selected

based on the fitness function Gn. In practice, we choose a random pro-portion Bi

n of an existing solution ξin in the current population with amean value ∝ Gn(ξin) to breed a brand new generation of “improved”solutions (ξ1

n, . . . , ξNn ). For instance, for every index i, with a proba-

bility εnGn(ξin), we set ξin = ξin, otherwise we replace ξin with a newindividual ξin = ξjn randomly chosen from the whole population witha probability proportional to Gn(ξjn). The parameter εn ≥ 0 is a tun-ing parameter that must satisfy the constraint εnGn(ξin) ≤ 1, for every1 ≤ i ≤ N . During the prediction-mutation stage, every selected indi-vidual ξin moves to a new solution ξin+1 = x randomly chosen in En+1,with a distribution Mn+1(ξin,dx).

If we interpret the updating-selection transition as a birth and deathprocess, then the important notion of the ancestral line of a currentindividual arises. More precisely, when a particle ξin−1 −→ ξin evolvesto a new location ξin, we can interpret ξin−1 as the parent of ξin. Lookingbackwards in time and recalling that the particle ξin−1 has selected asite ξjn−1 in the configuration at time (n − 1), we can interpret this siteξjn−1 as the parent of ξin−1 and therefore as the ancestor denoted ξin−1,n

at level (n − 1) of ξin. Running backwards in time we may trace thewhole ancestral line as

ξi0,n←− ξi1,n←− ·· · ←− ξin−1,n←− ξin,n = ξin. (1.13)

Most of the terminology we have used is drawn from filtering andgenetic evolution theories.

In the filtering example presented in the subsection 1.1, the formerparticle model is dictated by the two steps prediction-updating learningequations of the conditional distributions of the signal processXk, givensome noisy and partial observations Yk. In this setting, the potentialfunctions represent the likelihood function of the current observation,while the free exploration transitions are related to the Markov transi-tions of the signal process. More formally, using the notation we usedin example (1.2), we have:

dpk(xk|xk−1) = Mk(xk−1|xk)

Full text available at: http://dx.doi.org/10.1561/2200000026

Page 24: On the Concentration Properties of Interacting Particle

1.4 Interacting Particle Systems 15

and

pk(yk|xk) = Gk(xk).

In biology, the mutation-selection particle model presented aboveis used to mimic genetic evolutions of biological organisms and, moregenerally, natural evolution processes. For instance, in gene analysis,each population of individuals represents a chromosome and each indi-vidual particle is called a gene. In this setting the fitness potentialfunction is usually time-homogeneous and it represents the quality andthe adaptation potential value of the set of genes in a chromosome [62].These particle algorithms are also used in population analysis to modelchanges in the structure of population in time and in space.

The different types of particle approximation measures associatedwith the genetic type particle model described above are summarizedin the following synthetic picture corresponding to the case N = 3.

In the next four subsections we give an overview of the four parti-cle approximation measures that can be extracted from the interact-ing population evolution model described above. We also provide somebasic formulation of the concentration inequalities that will be treatedin greater detail later. As a service to the reader we also provide precisepointers to their location within each section of the monograph.

We have already mentioned that the proofs of these results are quitesubtle. In the further development of the next subsections, c1 standsfor a finite constant related to the bias of the particle model, while c2

is related to the variance of the scheme. The value of these constantsmay vary from one line to another, but in all the situations they do notdepend on the time parameter.

The precise form of the constants in these exponential inequali-ties depends on the contraction properties of Feynman-Kac flows. Our

Full text available at: http://dx.doi.org/10.1561/2200000026

Page 25: On the Concentration Properties of Interacting Particle

16 Stochastic Particle Methods

stochastic analysis requires us to combine the stability properties of thenonlinear semigroup of the Feynman-Kac distribution flow ηn, with thedeep convergence results of empirical processes theory associated withinteracting random samples.

1.4.1 Last Population Models

The occupation measures of the current population, represented by thered dots in the above figure

ηNn :=1N

N∑i=1

δξin

converge to the n-th time marginals ηn of the Feynman-Kac mea-sures Qn. We shall measure the performance of these particle estimatesthrough several concentration inequalities, with a special emphasis onuniform inequalities w.r.t. the time parameter. Our results will basi-cally be stated as follows.

1) For any time horizon n ≥ 0, any bounded function f , any N ≥ 1,and for any x ≥ 0, the probability of the event

[ηNn − ηn](f) ≤ c1

N(1 + x +

√x) +

c2√N

√x

is greater than 1 − e−x.We have already mentioned one important consequence of these

uniform concentration inequalities for time homogeneous Feynman-Kacmodels. Under some regularity conditions, the flow of measures ηn tendsto some fixed point distribution η∞, in the sense that

‖ηn − η∞‖tv ≤ c3 e−δn (1.14)

for some finite positive constants c3 and δ. In the above display‖ν − µ‖tv stands for the total variation distance. The connexionsbetween these limiting measures and the top of the spectrum ofSchrodinger operators is discussed in subsection 2.7.1. We also referthe reader to subsection 2.7.2 for a discussion on these quasi-invariantmeasures and Yaglom limits. Quantitative contraction theorems forFeynman-Kac semigroups are developed in subsection 3.4.2. As a directconsequence of the above inequalities, we find that for any x ≥ 0, the

Full text available at: http://dx.doi.org/10.1561/2200000026

Page 26: On the Concentration Properties of Interacting Particle

1.4 Interacting Particle Systems 17

probability of the following events is greater than 1 − e−x:

[ηNn − η∞](f) ≤ c1

N(1 + x +

√x) +

c2√N

√x + c3e

−δn.

2) For any x = (xi)1≤i≤d ∈ En = Rd, we set (−∞,x] =∏di=1(−∞,xi]

and we consider the repartition functions

Fn(x) = ηn(1(−∞,x]) and FNn (x) = ηNn (1(−∞,x]).

The probability of the following event√N‖FNn − Fn‖ ≤ c

√d(x + 1)

is greater than 1 − e−x, for any x ≥ 0, for some universal constantc <∞ that does not depend on the dimension, nor on the time param-eter. In the above display ‖F‖ = supx |F (x)| stands for the uniformnorm. Furthermore, under the stability properties (1.14), if we set

F∞(x) = η∞(1(−∞,x])

then, the probability of the following event

‖FNn − F∞‖ ≤c√N

√d(x + 1) + c3e

−δn

is greater than 1 − e−x, for any x ≥ 0, for some universal constantc <∞ that does not depend on the dimension.

For more precise statements, we refer the reader to Corollary 6.4,and Corollary 6.9, respectively.

The concentration properties of the particle measures ηNn aroundtheir limiting values are developed in Section 6. In subsection 6.3,we design a stochastic perturbation analysis that allows us to enterthe stability properties of the limiting Feynman-Kac semigroup. Finitemarginal models are discussed in subsection 6.4.1. Subsection 6.4.2 isconcerned with the concentration inequalities of interacting particleprocesses w.r.t. some collection of functions.

1.4.2 Particle Free Energy Models

Mimicking the multiplicative formula (1.9), we set

ZNn =∏

0≤p<nηNp (Gp) and γNn (dx) = ZNn × ηNn (dx). (1.15)

Full text available at: http://dx.doi.org/10.1561/2200000026

Page 27: On the Concentration Properties of Interacting Particle

18 Stochastic Particle Methods

We have already mentioned that these rather complex particle mod-els provide an unbiased estimate of the unnormalized measures. Thatis, we have that

E

ηNn (fn)∏

0≤p<nηNp (Gp)

= E

fn(Xn)∏

0≤p<nGp(Xp)

. (1.16)

The concentration properties of the unbiased particle free energiesZNn around their limiting values Zn are developed in subsection 6.5.Our results will basically be stated as follows.

For any N ≥ 1, and any ε ∈ {+1,−1}, the probability of each of thefollowing events

ε

nlogZNnZn≤ c1

N(1 + x +

√x) +

c2√N

√x

is greater than 1 − e−x. A more precise statement is provided inCorollary 6.14.

1.4.3 Genealogical Tree Model

The occupation measure of the N -genealogical tree model representedby the lines linking the blue dots converges as N →∞ to the distribu-tion Qn

limN→∞

1N

N∑i=1

δ(ξi0,n,ξi1,n,...,ξ

in,n) = Qn. (1.17)

Our concentration inequalities will basically be stated as follows.A more precise statement is provided in Corollary 6.5.

For any n ≥ 0, any bounded function fn on the path space En,such that (s.t.) ‖fn‖ ≤ 1, and any N ≥ 1, the probability of each of thefollowing events[

1N

N∑i=1

fn(ξi0,n, ξi1,n, . . . , ξ

in,n) − Qn(fn)

]

≤ c1n + 1N

(1 + x +√x) + c2

√(n + 1)N

√x

is greater than 1 − e−x.

Full text available at: http://dx.doi.org/10.1561/2200000026

Page 28: On the Concentration Properties of Interacting Particle

1.4 Interacting Particle Systems 19

The concentration properties of genealogical tree occupation mea-sures can be derived more or less directly from those of the currentpopulation models. This rather surprising assertion comes from thefact that the n-th time marginal ηn of a Feynman-Kac measure asso-ciated with a reference historical Markov process has the same formas in the measure (1.4). This equivalence principle between Qn andthe marginal measures are developed in subsection 3.2, dedicated tohistorical Feynman-Kac models.

Using these properties, we prove concentration properties for inter-acting empirical processes associated with genealogical tree models.Our concentration inequalities will basically be stated as follows.A more precise statement is provided in subsection 6.4.2. We let Fnbe the set of product functions of cell indicators in the path spaceEn = (Rd0 × . . . ,×Rdn), for some dp ≥ 1, p ≥ 0. We also denote by ηNnthe occupation measure of the genealogical tree model. In this notation,the probability of the following event

supfn∈Fn

|ηNn (fn) − Qn(fn)| ≤ c(n + 1)

√∑0≤p≤n dp

N(x + 1)

is greater than 1 − e−x, for any x ≥ 0, for some universal constantc <∞ that does not depend on the dimension.

1.4.4 Complete Genealogical Tree Models

Mimicking the backward model (1.11) and the above formulae, we set

ΓNn = ZNn × QNn (1.18)

with

QNn (d(x0, . . . ,xn)) = ηNn (dxn)

n∏q=1

Mq,ηNq−1(xq,dxq−1).

Notice that the computation of sums w.r.t. these particle measuresare reduced to summations over the particle locations ξin. It is thereforenatural to identify a population of individuals (ξ1

n, . . . , ξNn ) at time n

to the ordered set of indexes {1, . . . ,N}. In this case, the occupation

Full text available at: http://dx.doi.org/10.1561/2200000026

Page 29: On the Concentration Properties of Interacting Particle

20 Stochastic Particle Methods

measures and the functions are identified with the following line andcolumn vectors

ηNn :=[

1N,. . . ,

1N

]and fn :=

fn(ξ1n)

...fn(ξNn )

and the matrices Mn,ηNn−1

by the (N × N) matrices

Mn,ηNn−1:=

Mn,ηNn−1

(ξ1n, ξ

1n−1) · · · Mn,ηNn−1

(ξ1n, ξ

Nn−1)

......

...Mn,ηNn−1

(ξNn , ξ1n−1) · · · Mn,ηNn−1

(ξNn , ξNn−1)

(1.19)

with the (i, j)-entries

Mn,ηNn−1(ξin, ξ

jn−1) =

Gn−1(ξjn−1)Hn(ξjn−1, ξin)∑N

k=1Gn−1(ξkn−1)Hn(ξkn−1, ξin).

For instance, the Qn-integration of normalized additive linear function-als of the form

fn(x0, . . . ,xn) =1

n + 1

∑0≤p≤n

fp(xp) (1.20)

is given the particle matrix approximation model

QNn (fn) =

1n + 1

∑0≤p≤n

ηNn Mn,ηNn−1Mn−1,ηNn−2

. . .Mp+1,ηNp(fp).

These type of additive functionals arise in the calculation of thesensitivity measures discussed in subsection 2.4.1.

The concentration properties of the particle measures QNn around

the Feynman-Kac measures Qn are developed in subsection 6.6. Spe-cial emphasis is given to the additive functional models (1.20). Insubsection 6.6.3, we extend the stochastic perturbation methodologydeveloped in subsection 6.3 for time marginal models to the particlebackward Markov chain associated with the random stochastic matrices(1.19). This technique allows us to enter not only the stability prop-erties of the limiting Feynman-Kac semigroup, but also those of theparticle backward Markov chain model.

Full text available at: http://dx.doi.org/10.1561/2200000026

Page 30: On the Concentration Properties of Interacting Particle

1.4 Interacting Particle Systems 21

Our concentration inequalities will basically be stated as follows.A more precise statement is provided in Corollary 6.20 and in Corol-lary 6.24.

For any n ≥ 0, any normalized additive functional of the form (1.20),with max0≤p≤n ‖fp‖ ≤ 1, and any N ≥ 1, the probability of each of thefollowing events

[QNn − Qn](fn) ≤ c1

1N

(1 + x +√x) + c2

√x

N(n + 1)

is greater than 1 − e−x.For any a = (ai)1≤i≤d ∈ En = Rd, we denote by Ca the cell

Ca := (−∞,a] =d∏i=1

(−∞,ai]

and fa,n the additive functional

fa,n(x0, . . . ,xn) =1

n + 1

∑0≤p≤n

1(−∞,a](xp).

The probability of the following event

supa∈Rd|QN

n (fa,n) − Qn(fa,n)| ≤ c√

d

N(x + 1)

is greater than 1 − e−x, for any x ≥ 0, for some constant c <∞ thatdoes not depend on the dimension, nor on the time horizon.

Remark 1.1. One way to turn all of these inequalities into of Bern-stein style concentration inequalities is as follows. For any exponentialinequality of the form

∀x ≥ 0 P(X ≤ ax +√

2bx + c) ≤ 1 − e−x

for some non negative constants (a,b,c), we also have

∀y ≥ 0 P(X ≤ y + c) ≤ 1 − exp(− y2

2(b + ay)

).

A proof of this result is provided in Lemma 4.13.

Full text available at: http://dx.doi.org/10.1561/2200000026

Page 31: On the Concentration Properties of Interacting Particle

22 Stochastic Particle Methods

1.5 Basic Notation

Here we provide some background from stochastic analysis and integraloperator theory, which we require for our proofs. Most of the resultswith detailed proofs can be located in the book [17] on Feynman-Kacformulae and interacting particle methods. Our proofs also containcross-references to this rather well known material, so the reader maywish to skip this subsection and proceed directly to Section 2, which isdedicated to some application domains of Feynman-Kac models.

1.5.1 Integral Operators

We denote respectively byM(E),M0(E), P(E), and B(E), the set ofall finite signed measures on some measurable space (E,E), the con-vex subset of measures with null mass, the set of all probability mea-sures, and the Banach space of all bounded and measurable functionsf equipped with the uniform norm ‖f‖. We also denote by Osc1(E),and by B1(E) the set of E-measurable functions f with oscillationsosc(f) ≤ 1, and respectively with ‖f‖ ≤ 1. We let

µ(f) =∫µ(dx)f(x)

be the Lebesgue integral of a function f ∈ B(E), with respect to ameasure µ ∈M(E).

We recall that the total variation distance on M(E) is defined forany µ ∈M(E) by

‖µ‖tv =12

sup(A,B)∈E2

(µ(A) − µ(B)).

We recall that a bounded integral operatorM turned from a measurablespace (E,E) into an auxiliary measurable space (F,F) is an operatorf 7→M(f) from B(F ) into B(E) such that the functions

M(f)(x) :=∫FM(x,dy)f(y)

are E-measurable and bounded, for any f ∈ B(F ). A Markov kernel is apositive and bounded integral operator M with M(1) = 1. Given a pair

Full text available at: http://dx.doi.org/10.1561/2200000026

Page 32: On the Concentration Properties of Interacting Particle

1.5 Basic Notation 23

of bounded integral operators (M1,M2), we let (M1M2) represent thecomposition operator defined by (M1M2)(f) = M1(M2(f)). For timehomogeneous state spaces, we denote by Mm = Mm−1M = MMm−1

the m-th composition of a given bounded integral operator M , withm ≥ 1. A bounded integral operator M turned from a measurable space(E,E) into an auxiliary measurable space (F,F) also generates a dualoperator

µ(dx) 7→ (µM)(dx) =∫µ(dy)M(y,dx)

fromM(E) intoM(F ) defined by (µM)(f) := µ(M(f)). We also usedthe notation

K([f − K(f)]2)(x) := K([f − K(f)(x)]2)(x)

for some bounded integral operator K and some bounded function f .We prefer to avoid unnecessary abstraction and technical assump-

tions, so we frame the standing assumption that all the test functionsare in the unit sphere, and the integral operators, and all the randomvariables are sufficiently regular so that we are justified in computingintegral transport equations, regular versions of conditional expecta-tions, and so forth.

1.5.2 Contraction Coefficients

When the bounded integral operator M has a constant mass, that is,when M(1)(x) = M(1)(y) for any (x,y) ∈ E2, the operator µ 7→ µM

maps M0(E) into M0(F ). In this situation, we let β(M) be theDobrushin coefficient of a bounded integral operator M defined by theformula

β(M) := sup {osc(M(f)); f ∈ Osc(F )}.

Notice that β(M) is the operator norm of M on M0(E), and we havethe equivalent formulations

β(M) = sup{‖M(x, .) −M(y, .)‖tv ; (x,y) ∈ E2}

= supµ∈M0(E)

‖µM‖tv/‖µ‖tv.

Full text available at: http://dx.doi.org/10.1561/2200000026

Page 33: On the Concentration Properties of Interacting Particle

24 Stochastic Particle Methods

A detailed proof of these well known formulae can be found in [17].Given a positive and bounded potential function G on E, we also

denote by ΨG the Boltzmann-Gibbs mapping from P(E) into itself,defined for any µ ∈ P(E) by

ΨG(µ)(dx) =1

µ(G)G(x)µ(dx).

For [0,1]-valued potential functions, we also mention that ΨG(µ) canbe expressed as a non linear Markov transport equation

ΨG(µ) = µSµ,G (1.21)

with the Markov transitions

Sµ,G(x,dy) = G(x)δx(dy) + (1 − G(x))ΨG(µ)(dy).

We notice that

ΨG(µ) − ΨG(ν) = (µ − ν)Sµ + ν(Sµ − Sν)

and

ν(Sµ − Sν) = (1 − ν(G))[ΨG(µ) − ΨG(ν)]

from which we find the formula

ΨG(µ) − ΨG(ν) =1

ν(G)(µ − ν)Sµ.

In addition, using the fact that

∀(x,A) ∈ (E,E) Sµ(x,A) ≥ (1 − ‖G‖)ΨG(µ)(A)

we prove that β(Sµ) ≤ ‖G‖ and

‖ΨG(µ) − ΨG(ν)‖tv ≤‖G‖

µ(G) ∨ ν(G)‖µ − ν‖tv.

If we set Φ(µ) = ΨG(µ)M , for some Markov transition M , then wehave the decomposition

Φ(µ) − Φ(ν) =1

ν(G)(µ − ν)SµM (1.22)

Full text available at: http://dx.doi.org/10.1561/2200000026

Page 34: On the Concentration Properties of Interacting Particle

1.5 Basic Notation 25

for any couple of measures ν,µ on E. From the previous discussion, wealso find the following Lipschitz estimates

‖Φ(µ) − Φ(ν)‖tv ≤‖G‖

µ(G) ∨ ν(G)β(M)‖µ − ν‖tv. (1.23)

We end this subsection with an interesting contraction property of aMarkov transition

MG(x,dy) =M(x,dy)G(y)M(G)(x)

= ΨG(δxM)(dy) (1.24)

associated with a ]0,1]-valued potential function G, with

g = supx,y

G(x)/G(y) <∞. (1.25)

It is easily checked that

|MG(f)(x) −MG(f)(y)| = |ΨG(δxM)(f) − ΨG(δyM)(f)|

≤ g‖δxM − δyM‖tv

from which we conclude that

β(MG) ≤ gβ(M). (1.26)

1.5.3 Orlicz Norms and Gaussian Moments

We let πψ[Y ] be the Orlicz norm of an R-valued random variable Yassociated with the convex function ψ(u) = eu

2 − 1, and defined by

πψ(Y ) = inf {a ∈ (0,∞) :E(ψ(|Y |/a)) ≤ 1}

with the convention inf∅ =∞. Notice that

πψ(Y ) ≤ c⇐⇒ E(ψ(Y/c)) ≤ 1.

For instance, the Orlicz norm of a Gaussian and centered randomvariable U , s.t. E(U2) = 1, is given by πψ(U) =

√8/3. We also recall

that

E(U2m) = b(2m)2m := (2m)m 2−m

E(|U |2m+1) ≤ b(2m + 1)2m+1 :=(2m + 1)(m+1)√

m + 1/22−(m+1/2) (1.27)

Full text available at: http://dx.doi.org/10.1561/2200000026

Page 35: On the Concentration Properties of Interacting Particle

26 Stochastic Particle Methods

with (q + p)p := (q + p)!/q!. The second assertion comes from the factthat

E(U2m+1)2 ≤ E(U2m)E(U2(m+1))

and therefore

b(2m + 1)2(2m+1) = E(U2m)E(U2(m+1))

= 2−(2m+1) (2m)m (2(m + 1))(m+1).

This formula is a direct consequence of the following decompositions

(2(m + 1))(m+1) =(2(m + 1))!

(m + 1)!= 2

(2m + 1)!m!

= 2(2m + 1)(m+1)

and

(2m)m =1

2m + 1(2m + 1)!

m!=

12m + 1

(2m + 1)(m+1).

We also mention that

b(m) ≤ b(2m). (1.28)

Indeed, for even numbers m = 2p we have

b(m)2m = b(2p)4p = E(U2p)2 ≤ E(U4p) = b(4p)4p = b(2m)2m

and for odd numbers m = (2p + 1), we have

b(m)2m = b(2p + 1)2(2p+1) = E(U2p)E(U2(p+1))

≤ E(

(U2p)(2p+1)p

) p2p+1

E(

(U2(p+1))(2p+1)p+1

) p+12p+1

= E(U2(2p+1)) = b(2(2p + 1))2(2p+1) = b(2m)2m.

Full text available at: http://dx.doi.org/10.1561/2200000026

Page 36: On the Concentration Properties of Interacting Particle

References

[1] R. Adamczak, “A tail inequality for suprema of unbounded empirical processeswith applications to Markov chains,” Electronic Journal of Probability, vol. 13,no. 34, pp. 1000–1034, 2008.

[2] C. Andrieu, A. Doucet, and R. Holenstein, “Particle Markov chain Monte Carlomethods,” Journal Royal Statistical Society B, vol. 72, no. 3, pp. 269–342, 2010.

[3] N. Bellomo and M. Pulvirenti, “Generalized kinetic models in applied sci-ences,” in Modeling in Applied Sciences, Modeling and Simulation in Science,Engineering, and Technology, pp. 1–19, Birkhauser, Boston, 2000.

[4] N. Bellomo and M. Pulvirenti, eds., Modeling in Applied Sciences, Modeling andSimulation in Science, Engineering, and Technology. Birkhauser, Boston, 2000.

[5] E. Cances, B. Jourdain, and T. Lelievre, “Quantum Monte Carlo simulations offermions. A mathematical analysis of the fixed-node approximation,” ESAIM:M2AN, vol. 16, no. 9, pp. 1403–1449, 2006.

[6] R. Carmona, J.-P. Fouque, and D. Vestal, “Interacting particle systems for thecomputation of rare credit portfolio Losses,” Finance and Stochastics, vol. 13,no. 4, pp. 613–633, 2009.

[7] F. Caron, P. Del Moral, A. Tantar, and E. Tantar, “Simulation partic-ulaire, Dimensionnement en conditions extremes,” Research contract no.2010-IFREMER-01 with IFREMER, (92p) September, 2010.

[8] F. Cerou, P. Del Moral, Fr. Le Gland, and P. Lezaud, “Genealogical modelsin entrance times rare event analysis,” Alea, Latin American Journal ofProbability And Mathematical Statistics, vol. 1, pp. 181–203, 2006.

[9] F. Cerou, P. Del Moral, T. Furon, and A. Guyader, “Sequential Monte Carlofor rare event estimation,” Statistics and Computing, to appear, 2011.

159

Full text available at: http://dx.doi.org/10.1561/2200000026

Page 37: On the Concentration Properties of Interacting Particle

160 References

[10] F. Cerou, P. Del Moral, A. Guyader, F. LeGland, P. Lezaud, and H. Topart,“Some recent improvements to importance splitting,” in (preliminary version),Proceedings of 6th International Workshop on Rare Event Simulation,Bamberg, Germany, 2006.

[11] D. A. Dawson, “Stochastic evolution equations and related measure processes,”J. Multivariate Anal., vol. 3, pp. 1–52, 1975.

[12] D. A. Dawson, “Measure-valued Markov processes,” in Ecole d’Ete de Prob-abilites de Saint-Flour 1991, Lecture Notes Math. 1541, pp. 1–260, Springer,Berlin, 1993.

[13] D. A. Dawson and P. Del Moral, “Large deviations for interacting processes inthe strong topology,” in Statistical Modeling and Analysis for Complex DataProblem, (P. Duchesne and B. Rmillard, eds.), pp. 179–209, Springer, 2005.

[14] D. A. Dawson, K. J. Hochberg, and V. Vinogradov, “On weak convergence ofbranching particle systems undergoing spatial motion,” in Stochastic Analysis:Random Fields and Measure-valued Processes, pp. 65–79, Israel Math. Conf.Proc. 10. Bar-Ilan Univ., Ramat Gan., 1996.

[15] P. Del Moral, “Non linear filtering: Interacting particle solution,” MarkovProcesses and Related Fields, vol. 2, no. 4, pp. 555–580, 1996.

[16] P. Del Moral, “Measure valued processes and interacting particle systems.Application to nonlinear filtering problems,” Annals of Applied Probability,vol. 8, no. 2, pp. 438–495, 1998.

[17] P. Del Moral, Feynman-Kac Formulae: Genealogical and Interacting ParticleSystems with Applications. New York: Springer-Verlag, 2004.

[18] P. Del Moral and A. Doucet, “On a class of genealogical and interactingMetropolis Models,” in Seminaire de Probabilites XXXVII, (J. Azema,M. Emery, M. Ledoux, and M. Yor, eds.), pp. 415–446, Lecture Notes inMathematics 1832, Springer-Verlag, Berlin, 2003.

[19] P. Del Moral and A. Doucet, “Particle motions in absorbing medium withhard and soft obstacles,” Stochastic Analysis and Applications, vol. 22, no. 5,pp. 1175–1207, 2004.

[20] P. Del Moral and A. Doucet, Sequential Monte Carlo & Genetic ParticleModels. Theory and Practice. Chapman & Hall, Green series, Statistics andApplied Probability, to appear, 2012.

[21] P. Del Moral, A. Doucet, and A. Jasra, “Sequential Monte Carlo samplers,”J. Royal Statist. Soc. B, vol. 68, pp. 411–436, 2006.

[22] P. Del Moral, A. Doucet, and A. Jasra, “An adaptive sequential Monte Carlomethod for approximate Bayesian computation,” Research report, ImperialCollege, 2008.

[23] P. Del Moral, A. Doucet, and S. S. Singh, “A backward particle interpretationof Feynman-Kac formulae,” ESAIM M2AN, vol. 44, no. 5, pp. 947–976, 2010.

[24] P. Del Moral, A. Doucet, and S. S. Singh, “Forward smoothing using sequen-tial Monte Carlo,” Technical Report 638. Cambridge University EngineeringDepartment., 2010.

[25] P. Del Moral, A. Doucet, and S. S. Singh, “Uniform stability of a particleapproximation of the optimal filter derivative,” Technical Report CUED/F-INFENG/TR 668. Cambridge University Engineering Department., 2011.

Full text available at: http://dx.doi.org/10.1561/2200000026

Page 38: On the Concentration Properties of Interacting Particle

References 161

[26] P. Del Moral and J. Garnier, “Genealogical particle analysis of rare events,”Annals of Applied Probability, vol. 15, no. 4, pp. 2496–2534, 2005.

[27] P. Del Moral and J. Garnier, “Simulations of rare events in fiber optics byinteracting particle systems,” (preliminary version), Optics Communications,Elsevier, 2006.

[28] P. Del Moral and A. Guionnet, “Large deviations for interacting particlesystems. Applications to nonlinear filtering problems,” Stochastic ProcessesAppl., vol. 78, pp. 69–95, 1998.

[29] P. Del Moral and A. Guionnet, “A central limit theorem for nonlinearfiltering using interacting particle systems,” Ann. Appl. Probab., vol. 9, no. 2,pp. 275–297, 1999.

[30] P. Del Moral and A. Guionnet, “On the stability of measure valued processeswith applications to filtering,” C.R. Acad. Sci. Paris, vol. t. 329, no. Serie I,pp. 429–434, 1999.

[31] P. Del Moral and A. Guionnet, “On the stability of interacting processes withapplications to filtering and genetic algorithms,” Ann. Inst. Henri Poincare,vol. 37, no. 2, pp. 155–194, 2001.

[32] P. Del Moral, S. Hu, and L. M. Wu, “Moderate deviations for mean fieldparticle models,” in preparation, 2011.

[33] P. Del Moral and J. Jacod, “The Monte-Carlo method for filtering withdiscrete-time observations: Central limit theorems,” in Numerical Methods andStochastics (Toronto, ON, 1999), Vol. 34 of Fields Inst. Commun., pp. 29–53,American Mathematical Society, Providence, RI, 2002.

[34] P. Del Moral, J. Jacod, and P. Protter, “The Monte-Carlo method for filteringwith discrete-time observations,” Probability Theory and Related Fields,vol. 120, pp. 346–368, 2001.

[35] P. Del Moral, M. A. Kouritzin, and L. Miclo, “On a class of discrete generationinteracting particle systems,” Electronic Journal of Probability, vol. 16,pp. 1–26, 2001.

[36] P. Del Moral and M. Ledoux, “On the convergence and the applications ofempirical processes for interacting particle systems and nonlinear filtering,”Journal of Theoretical Probability, vol. 13, no. 1, pp. 225–257, 2000.

[37] P. Del Moral and P. Lezaud, “Branching and interacting particle interpreta-tion of rare event probabilities,” in Stochastic Hybrid Systems: Theory andSafety Critical Applications, (H. Blom and J. Lygeros, eds.), Springer-Verlag,Heidelberg, 2006.

[38] P. Del Moral and L. Miclo, “Branching and interacting particle systemsapproximations of Feynman-Kac formulae with applications to nonlinear fil-tering,” in Seminaire de Probabilites XXXIV, Vol. 1729, (J. Azma, M. Emery,M. Ledoux, and M. Yor, eds.), pp. 1–145, Lecture Notes in Mathematics,Springer-Verlag Berlin, 2000.

[39] P. Del Moral and L. Miclo, “Particle approximations of Lyapunov exponentsconnected to Schrodinger operators and Feynman-Kac semigroups,” ESAIMProbability and Statistics, vol. 7, pp. 169–207, 2003.

[40] P. Del Moral and E. Rio, “Concentration inequalities for mean field particlemodels,” Annals of Applied Probability, vol. 21, no. 3, pp. 1017–1052, 2011.

Full text available at: http://dx.doi.org/10.1561/2200000026

Page 39: On the Concentration Properties of Interacting Particle

162 References

[41] P. Del Moral and T. Zajic, “A note on Laplace-Varadhan’s integral lemma,”Bernoulli, vol. 9, no. 1, pp. 49–65, 2003.

[42] G. B. Di Masi, M. Pratelli, and W. G. Runggaldier, “An approximation forthe nonlinear filtering problem with error bounds,” Stochastics, vol. 14, no. 4,pp. 247–271, 1985.

[43] R. Douc, A. Guillin, and J. Najim, “Moderate deviations for particle filtering,”The Annals of Applied Probability, vol. 15, no. 1B, pp. 587–614, 2005.

[44] A. Doucet, N. De Freitas, and N. Gordon, eds., Sequential Monte CarloMethods in Pratice, Statistics for Engineering and Information Science. NewYork: Springer, 2001.

[45] E. B. Dynkin, “An introduction to branching measure-valued processes,” inCRM Monograph Series, Vol. 6. Amer. Math. Soc., Providence, 1994.

[46] M. El Makrini, B. Jourdain, and T. Lelievre, “Diffusion Monte Carlomethod: Numerical analysis in a simple case,” ESAIM: M2AN, vol. 41, no. 2,pp. 189–213, 2007.

[47] S. N. Ethier and T. G. Kurtz, Markov Processes: Characterization andConvergence. Wiley, New York, 1995.

[48] C. J. Everett and S. Ulam, Proc. Nat. Acad. Sciences, vol. 403, no. 33, 1948.[49] C. J. Everett and S. Ulam, USAEC, Los Alamos report, LADC-533, LADC-534.

declassified, 1948.[50] W. Feller, “Diffusion processes in genetics,” in Proc. Second Berkeley Symp.

Math. Statist. Prob., pp. 227–246, Berkeley: University of California Press, 1951.[51] P. A. Ferrari, H. Kesten, S. Martınez, and P. Picco, “Existence of quasi-

stationary distributions. A renewal dynamical approach,” Ann. Probab.,vol. 23, pp. 501–521, 1995.

[52] Fr. Caron, P. Del Moral, A. Doucet, and M. Pace. On the ConditionalDistributions of Spatial Point Processes HAL-INRIA RR-7232. To appear inAdvances in Applied Probability, 2011.

[53] Fr. Caron, P. Del Moral, A. Doucet, and M. Pace, “Particle approximations ofa class of branching distribution flows arising in multi-target tracking,” HAL-INRIA RR-7233 (2010) [29p], SIAM Journal on Control and Optimization,vol. 49, no. 4, pp. 1766–1792, 2011.

[54] Fr. Caron, P. Del Moral, M. Pace, and B. N. Vo, “On the stability and theapproximation of branching distribution flows, with applications to nonlinearmultiple target filtering,” Stoch. Analysis and Applications, vol. 29, no. 6,pp. 951–997, 2011.

[55] Fr. Cerou, A. Guyader, R. Rubinstein, and R. Vaismana, “On the use ofsmoothing to improve the performance of the splitting method,” StochasticModels, vol. 27, no. 4, pp. 629–650, 2011.

[56] E. Gine, “Lectures on some aspects of the bootstrap,” in Lecture Notes inMaths, Vol. 1665, Springer, 1997.

[57] F. Gosselin, “Asymptotic behavior of absorbing Markov chains conditional onnonabsorption for applications in conservation biology,” Ann. Appl. Probab.,vol. 11, pp. 261–284, 2001.

[58] C. Graham and S. Meleard, “Stochastic particle approximations for generalizedBoltzmann models and convergence estimates,” Ann. Probab., vol. 25, no. 1,pp. 115–132, 1997.

Full text available at: http://dx.doi.org/10.1561/2200000026

Page 40: On the Concentration Properties of Interacting Particle

References 163

[59] C. Graham and S. Meleard, “Probabilistic tools and Monte-Carlo approxima-tions for some Boltzmann equations,” in CEMRACS 1999 (Orsay), Vol. 10of ESAIM Proceedings, pp. 77–126, (electronic). Societe de MathematiquesAppliquees et Industrielles, Paris, 1999.

[60] A. Guyader, N. Hengartner, and E. Matzner-Løber, “Simulation and estima-tion of extreme quantiles and extreme probabilities,” Applied Mathematics &Optimization, vol. 64, no. 2, pp. 171–196, 2011.

[61] H. P. McKean, Jr, “A class of Markov processes associated with nonlinearparabolic equations,” Proc. Natl. Acad. Sci. U.S.A., vol. 56, pp. 1907–1911,1966.

[62] J. H. Holland, Adaptation in Natural and Artificial Systems. University ofMichigan Press, Ann Arbor, 1975.

[63] H.P. McKean, Jr, “Propagation of chaos for a class of non-linear parabolicequations,” in Stochastic Differential Equations (Lecture Series in DifferentialEquations, Session 7, Catholic University, 1967), pp. 41–57, Air Force Officeof Scientific Research, Arlington, VA, 1967.

[64] A. M. Johansen, P. Del Moral, and A. Doucet, “Sequential Monte Carlo sam-plers for rare events,” in (preliminary version), Proceedings of 6th InternationalWorkshop on Rare Event Simulation, Bamberg, Germany, 2006.

[65] W. B. Johnson, G. Schechtman, and J. Zinn, “Best constants in momentinequalities for linear combinations of independent and exchangeable randomvariables,” Ann. Probab., vol. 13, no. 1, pp. 234–253, 1985.

[66] M. Kac, “On distributions of certain Wiener functionals,” Trans. AmericanMath. Soc., vol. 61, no. 1, pp. 1–13, 1949.

[67] T. Klein and E. Rio, “Concentration around the mean for maxima of empiricalprocesses,” Ann. Probab., vol. 33, no. 3, pp. 1060–1077, 2005.

[68] H. Korezlioglu and W. J. Runggaldier, “Filtering for nonlinear systems drivenby nonwhite noises: An approximating scheme,” Stochastics and StochasticsRep., vol. 44, no. 1–2, pp. 65–102, 1983.

[69] J. F. Le Gall, “Spatial branching processes, random snakes, and partial differentequations,” in Lectures Given in the Nachdiplomsvorlesung, ETH Zurich, Pub-lished in the series, Lectures in Mathematics, ETH Zurich, Birkhauser, 1999.

[70] M. Ledoux, “On Talagrand’s deviation inequalities for product measures,”ESAIM Probability and Statistics, vol. 1, pp. 63–87, 1996.

[71] M. Ledoux, “The concentration of measure phenomenon,” AMS Monographs,Providence, 2001.

[72] M. Ledoux and M. Talagrand, Probability in Banach Spaces. Springer-Verlag,1991.

[73] R. Mahler, “A theoretical foundation for the Stein-Winter Probability Hypoth-esis Density (PHD) multi-target tracking approach,” in Proc. MSS Nat’l Symp.on Sensor and Data Fusion, Vol. 1, 2000.

[74] R. Mahler, “Multi-target Bayes filtering via first-order multi-target moments,”IEEE Trans. Aerospace & Electronic Systems, vol. 39, no. 4, pp. 1152–1178,2003.

[75] S. Meleard, “Asymptotic behaviour of some interacting particle systems;McKean-Vlasov and Boltzmann models,” in Probabilistic Models for Nonlinear

Full text available at: http://dx.doi.org/10.1561/2200000026

Page 41: On the Concentration Properties of Interacting Particle

164 References

Partial Differential Equations, Montecatini Terme, 1995, (D. Talay andL. Tubaro, eds.), Lecture Notes in Mathematics 1627. Springer-Verlag, Berlin,1996.

[76] S. Meleard, “Convergence of the fluctuations for interacting diffusions withjumps associated with Boltzmann equations,” Stochastics Stochastics Rep.,vol. 63, no. 3–4, pp. 195–225, 1998.

[77] S. Meleard, “Probabilistic interpretation and approximations of some Boltz-mann equations,” in Stochastic models (Spanish) (Guanajuato, 1998), Vol. 14of Aportaciones Mat. Investig., pp. 1–64, Soc. Mat. Mexicana, Mexico, 1998.

[78] S. Meleard, “Stochastic approximations of the solution of a full Boltzmannequation with small initial data,” ESAIM Probab. Stat., vol. 2, pp. 23–40, 1998.

[79] N. Metropolis and S. Ulam, “The Monte Carlo method,” Journal of theAmerican Statistical Association, vol. 44, no. 247, pp. 335–341, 1949.

[80] J. Picard, “Approximation of the non linear filtering problems and order ofconvergence,” in Lecture Notes in Control and Inf. Sc, Vol. 61, 1984.

[81] D. Pollard, Empirical Processes: Theory and Applications, Vol. 2. NSF-CBMSRegional Conference Series in Probability and Statistics. Publication of theInstitute of Mathematical Statistics. Hayward, CA, 1990.

[82] M. Pulvirenti, “Kinetic limits for stochastic particle systems,” in ProbabilisticModels for Nonlinear Partial Differential Equations (Montecatini Terme, 1995),Vol. 1627 of Lecture Notes in Mathematics, pp. 96–126, Springer, Berlin, 1996.

[83] S. T. Rachev, Probability Metrics and the Stability of Stochastic Models. Wiley,New York, 1991.

[84] E. Rio, “Local invariance principles and their applications to densityestimation,” Probability and related Fields, vol. 98, pp. 21–45, 1994.

[85] M. Rousset, “On the control of an interacting particle approximation ofSchrodinger ground states,” SIAM J. Math. Anal., vol. 38, no. 3, pp. 824–844,2006.

[86] W. Runggaldier and L. Stettner, Approximations of Discrete Time PartiallyObserved Control Problems. Applied Mathematics Monographs CNR, GiardiniEditori, Pisa, 1994.

[87] Z. Skolicki and K. De Jong, “The influence of migration sizes and intervals onisland models,” in Proceedings of the Genetic and Evolutionary ComputationConference (GECCO-2005), ACM Press, 2005.

[88] D. Steinsaltz and S. N. Evans, “Markov mortality models: Implications ofquasistationarity and varying initial distributions,” Theor. Pop. Biol., vol. 65,pp. 319–337, 2004.

[89] A. S. Sznitman, “Topics in propagation of chaos,” in Ecole d’Ete de Prob-abilites de Saint-Flour XIX-1989, (P. L. Hennequin, ed.), Lecture Notes inMathematics 1464. Springer-Verlag, Berlin, 1991.

[90] M. Talagrand, “Concentration of measure and isoperimetric inequalities inproduct spaces,” Inst. Hautes Etudes Sci. Publ. Math., vol. 81, pp. 73–205, 1995.

[91] M. Talagrand, “New concentration inequalities in product spaces,” Invent.Math., vol. 125, pp. 505–563, 1996.

[92] M. Talagrand, “A new look at independence,” Ann. Probab., vol. 24, pp. 1–34,1996.

Full text available at: http://dx.doi.org/10.1561/2200000026

Page 42: On the Concentration Properties of Interacting Particle

References 165

[93] A. N. Van der Vaart and J. A. Wellner, Weak Convergence and EmpiricalProcesses with Applications to Statistics. New York: Springer Series inStatistics. Springer, 1996.

[94] J. Vanneste, “Estimating generalized Lyapunov exponents for products ofrandom matrices,” Phys. Rev. E, vol. 81, no. 3, p. 036701, 2010.

[95] J.-P. Vila and V. Rossi, “Nonlinear filtering in discrete time: A particleconvolution approach, Biostatistic group of Montpellier,” Technical Report04-03, (available at http://vrossi.free.fr/recherche.html), 2004.

[96] D. Whitley, S. Rana, and R. B. Heckendorn, “The island model geneticalgorithm: On separability, population size and convergence,” Journal ofComputing and Information Technology, vol. 7, no. 1, pp. 33–47, 1999.

Full text available at: http://dx.doi.org/10.1561/2200000026