data fusion modeling for groundwater systemsmmc2.geofisica.unam.mx/cursos/geoest/articulos... ·...

33
Ž . Journal of Contaminant Hydrology 42 2000 303–335 www.elsevier.comrlocaterjconhyd Data fusion modeling for groundwater systems David W. Porter a, ) , Bruce P. Gibbs a , Walter F. Jones b , Peter S. Huyakorn b , L. Larry Hamm c , Gregory P. Flach c a Fusion and Control Technology, Annapolis, MD 21401, USA b HydroGeoLogic, Herndon, VA 20170, USA c Westinghouse SaÕannah RiÕer Company, Aiken, SC 29808, USA Received 22 December 1997; accepted 26 August 1999 Abstract Engineering projects involving hydrogeology are faced with uncertainties because the earth is heterogeneous, and typical data sets are fragmented and disparate. In theory, predictions provided by computer simulations using calibrated models constrained by geological boundaries provide answers to support management decisions, and geostatistical methods quantify safety margins. In practice, current methods are limited by the data types and models that can be included, Ž . computational demands, or simplifying assumptions. Data Fusion Modeling DFM removes many of the limitations and is capable of providing data integration and model calibration with quantified uncertainty for a variety of hydrological, geological, and geophysical data types and models. The benefits of DFM for waste management, water supply, and geotechnical applications are savings in time and cost through the ability to produce visual models that fill in missing data and predictive numerical models to aid management optimization. DFM has the ability to update field-scale models in real time using PC or workstation systems and is ideally suited for parallel processing implementation. DFM is a spatial state estimation and system identification methodol- ogy that uses three sources of information: measured data, physical laws, and statistical models for uncertainty in spatial heterogeneities. What is new in DFM is the solution of the causality problem in the data assimilation Kalman filter methods to achieve computational practicality. The Kalman filter is generalized by introducing information filter methods due to Bierman coupled with a Markov random field representation for spatial variation. A Bayesian penalty function is imple- mented with Gauss–Newton methods. This leads to a computational problem similar to numerical Ž . simulation of the partial differential equations PDEs of groundwater. In fact, extensions of PDE solver ideas to break down computations over space form the computational heart of DFM. State estimates and uncertainties can be computed for heterogeneous hydraulic conductivity fields in multiple geological layers from the usually sparse hydraulic conductivity data and the often more ) Corresponding author. Present address: Applied Physics Laboratory, The Johns Hopkins University, Laurel, MD 20723, USA. Fax: q 1-410-451-4736; e-mail: [email protected] 0169-7722r00r$ - see front matter Published by Elsevier Science B.V. Ž . PII: S0169-7722 99 00081-9

Upload: others

Post on 09-Sep-2020

2 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Data fusion modeling for groundwater systemsmmc2.geofisica.unam.mx/cursos/geoest/Articulos... · 2015. 10. 23. · Data Fusion Modelling DFM has value added as an engineering decision

Ž .Journal of Contaminant Hydrology 42 2000 303–335www.elsevier.comrlocaterjconhyd

Data fusion modeling for groundwater systems

David W. Porter a,), Bruce P. Gibbs a, Walter F. Jones b,Peter S. Huyakorn b, L. Larry Hamm c, Gregory P. Flach c

a Fusion and Control Technology, Annapolis, MD 21401, USAb HydroGeoLogic, Herndon, VA 20170, USA

c Westinghouse SaÕannah RiÕer Company, Aiken, SC 29808, USA

Received 22 December 1997; accepted 26 August 1999

Abstract

Engineering projects involving hydrogeology are faced with uncertainties because the earth isheterogeneous, and typical data sets are fragmented and disparate. In theory, predictions providedby computer simulations using calibrated models constrained by geological boundaries provideanswers to support management decisions, and geostatistical methods quantify safety margins. Inpractice, current methods are limited by the data types and models that can be included,

Ž .computational demands, or simplifying assumptions. Data Fusion Modeling DFM removes manyof the limitations and is capable of providing data integration and model calibration withquantified uncertainty for a variety of hydrological, geological, and geophysical data types andmodels. The benefits of DFM for waste management, water supply, and geotechnical applicationsare savings in time and cost through the ability to produce visual models that fill in missing dataand predictive numerical models to aid management optimization. DFM has the ability to updatefield-scale models in real time using PC or workstation systems and is ideally suited for parallelprocessing implementation. DFM is a spatial state estimation and system identification methodol-ogy that uses three sources of information: measured data, physical laws, and statistical models foruncertainty in spatial heterogeneities. What is new in DFM is the solution of the causality problemin the data assimilation Kalman filter methods to achieve computational practicality. The Kalmanfilter is generalized by introducing information filter methods due to Bierman coupled with aMarkov random field representation for spatial variation. A Bayesian penalty function is imple-mented with Gauss–Newton methods. This leads to a computational problem similar to numerical

Ž .simulation of the partial differential equations PDEs of groundwater. In fact, extensions of PDEsolver ideas to break down computations over space form the computational heart of DFM. Stateestimates and uncertainties can be computed for heterogeneous hydraulic conductivity fields inmultiple geological layers from the usually sparse hydraulic conductivity data and the often more

) Corresponding author. Present address: Applied Physics Laboratory, The Johns Hopkins University,Laurel, MD 20723, USA. Fax: q1-410-451-4736; e-mail: [email protected]

0169-7722r00r$ - see front matter Published by Elsevier Science B.V.Ž .PII: S0169-7722 99 00081-9

Page 2: Data fusion modeling for groundwater systemsmmc2.geofisica.unam.mx/cursos/geoest/Articulos... · 2015. 10. 23. · Data Fusion Modelling DFM has value added as an engineering decision

( )D.W. Porter et al.rJournal of Contaminant Hydrology 42 2000 303–335304

plentiful head data. Further, a system identification theory has been derived based on statisticallikelihood principles. A maximum likelihood theory is provided to estimate statistical parameterssuch as Markov model parameters that determine the geostatistical variogram. Field-scale applica-tion of DFM at the DOE Savannah River Site is presented and compared with manual calibration.DFM calibration runs converge in less than 1 h on a Pentium Pro PC for a 3D model with morethan 15,000 nodes. Run time is approximately linear with the number of nodes. Furthermore,conditional simulation is used to quantify the statistical variability in model predictions such ascontaminant breakthrough curves. Published by Elsevier Science B.V.

Keywords: Groundwater modeling; Calibration; Geostatistics; Inverse modeling; Kalman filter; Data assimila-tion; Data fusion

1. Introduction

Engineering projects involving hydrogeology are often driven by uncertainties. Foractivities such as environmental remediation or siting of waste management facilities,cost-effective solutions frequently depend on the confidence with which groundwaterpredictions can be made. For example, to create a plume capture zone, answers areneeded to questions about the number, depth, location and pumping schedules for purgewells. Prediction of the groundwater system response using computer simulation pro-vides answers. In theory, simulation can provide real-time monitoring of remediation soa plume can be ‘‘seen’’ as it is being cleaned up. But simulation is only as good as itsgeological and hydrological inputs. The earth is very heterogeneous, and typical datasets are fragmented and disparate so there are substantial uncertainties.

Currently, engineers do not have adequate tools to calibrate models and quantifyuncertainty so they often rely solely on their judgment to build in sufficient safetymargins. This tends to lead to overly conservative decisions that waste money. Data

Ž .Fusion Modelling DFM has value added as an engineering decision tool for dataintegration, model calibration, and quantification of uncertainty. Fusion uses modelingmethods in two different ways. First, models provide prior knowledge of physical andstatistical relationships between fragmented and disparate data sets, and fusion usesthese relationships to extract hydrological and geological information from the measureddata. Second, models are used in computer simulation of remediation to optimizeremediation and to quantify safety margins.

With reduced cleanup budgets and movement to risk assessment with predictivemodels having quantified uncertainties, DFM provides an enabling technology. It has

Ž .been applied for the Department of Energy DOE at Savannah River, Fernald, Hanford,and Pantex; for DoD at the Massachusetts Military Reservation, Beale Air Force Base,and at the Letterkenney Army Depot; and at a former manufactured gas plant inMarshalltown, IA. The DFM system is available in PC or workstation software systems

Ž .with Graphical User Interface GUI and visualization software. DFM gives decisionmakers a quantitative basis for action so savings in cost and time can be realized throughthe following:Ø Real-time monitoring and optimization of remediation,Ø Remediation design with quantified safety margins,

Page 3: Data fusion modeling for groundwater systemsmmc2.geofisica.unam.mx/cursos/geoest/Articulos... · 2015. 10. 23. · Data Fusion Modelling DFM has value added as an engineering decision

( )D.W. Porter et al.rJournal of Contaminant Hydrology 42 2000 303–335 305

Ø Improved uncertainty for risk assessment,Ø Real-time determination of data worth for guidance of site investigation.

Section 2 provides motivation and a review of related areas. Then the theory ispresented for a baseline groundwater system and a description is provided of an

Ž .application at the DOE Savannah River Site SRS .

2. Background

Hydrogeological problems that motivated DFM are described and the key features ofthe DFM methodology are explained. Our DFM methodology builds on a foundation ofwork from hydrogeology, meteorology, oceanography, image processing, and the datafusion communities in military and space science areas. A literature review is presented,and it is explained how DFM relates to current methods and generalizes them to helpsolve some long-standing problems in hydrogeology.

DFM fits into the framework for hydrogeological decision analysis presented inŽ .Freeze et al. 1990 . A stochastic methodology is described for waste management,

water supply, and geotechnical applications. It is stated that uncertainty in heterogeneoushydraulic conductivity must be viewed as an autocorrelated spatial stochastic process inorder to account for uncertainty in subsurface investigation. Bayesian estimation isadvocated as the way to update prior information with new information from dataconsistent with sequential engineering decision making. DFM views hydraulic conduc-tivity as a stochastic process and produces Bayesian updates.

Ž .It is explained in Freeze et al. 1990 that the current Bayesian theory has limitationsin the application of inverse modeling to take advantage of important data worth. Forexample, it is clear that hydraulic head data used in conjunction with hydraulicconductivity data have considerable data worth, but Bayesian methods have difficultycombining head and conductivity data. This is because of limitations in current methodsconcerning data and model types, computational demands, and simplifying assumptions.DFM removes most of these limitations by building on the foundation of existingmethods.

The methods that are most limited by data and model types are traditional inverseŽ . Ž .modeling methods such as those in Cooley 1982 and Hill 1992 and classical kriging

Ž . Ž .methods such as those in Cressie 1991 and Journal 1991 . Inverse modeling isperformed using non-linear regression in the MODFLOWP software described in HillŽ .1992 . In theory, a least squares fit to head data can be achieved with the incorporationof point data on conductivity as prior information in MODFLOWP. However, in orderto avoid overparameterization, the estimation must be performed on a larger scale of

Ž .variability than appropriate for the conductivity data as explained in Hill 1992 . It isŽ .explained in Kitanidis and Vomvoris 1983 that the overparameterization problem can

be solved using a random field representation that can account for spatial variability atvarious scales. The lack of a random field representation in traditional inverse modelinglimits the data types that can be combined.

Classical kriging methods use nested variability structures as described in CressieŽ .1991 to account for variability at different scales. However, the classical methods do

Page 4: Data fusion modeling for groundwater systemsmmc2.geofisica.unam.mx/cursos/geoest/Articulos... · 2015. 10. 23. · Data Fusion Modelling DFM has value added as an engineering decision

( )D.W. Porter et al.rJournal of Contaminant Hydrology 42 2000 303–335306

not use groundwater numerical models so there is no way to combine conductivity datawith head data based on the physical relationship between them. The lack of agroundwater numerical model in classical kriging limits the data types that can becombined.

Numerous methods are able to work with a variety of data and model types inhydrology, meteorology, and oceanography such as those in Kitanidis and VomvorisŽ . Ž . Ž . Ž .1983 , Hoeksema and Kitinidis 1984 , Dagan 1985 , Gelhar 1986 , Carrera and

Ž . Ž . Ž . Ž .Neuman 1986a; b; c , Ghil 1989 , Carrera and Glorioso 1991 , Van Geer et al. 1991 ,Ž . Ž . Ž . Ž .Yangxiao et al. 1991 , Bennett 1992 , Sun and Yeh 1992 , Courtier et al. 1993 ,

Ž . Ž . Ž .Poeter and McKenna 1994 , LaVenue et al. 1995 , McLaughlin 1995 , RamaRao etŽ . Ž . Ž .al. 1995 , Zou and Parr 1995 , Eppstein and Dougherty 1996 , McLaughlin and

Ž . Ž . Ž .Townley 1996 , Daniel and Willsky 1997 and Zimmerman et al. 1998 . Many of thecrucial issues in moving from theory to practice are well understood. Ill-posednessproblems and problems of scale have been addressed. It is understood that modeling isinherently a conceptual iteration of hypothesizing model structure, estimating parameterswithin the structure, testing model validity, and repeating the process until an adequatemodel is obtained. Several unified theories exist for state estimation and systemidentification to support modeling conceptual iteration. The main remaining issues arelimitations concerning computational demands and simplifying assumptions.

In other fields, Kalman filtering solved apparently similar problems. Even thoughŽ .Kalman filtering has been introduced for groundwater systems in Van Geer et al. 1991 ,

Ž . Ž . Ž .Yangxiao et al. 1991 , McLaughlin 1995 , Zou and Parr 1995 , Eppstein andŽ . Ž . Ž .Dougherty 1996 , McLaughlin and Townley 1996 and Daniel and Willsky 1997 , the

limitations still remain. Part of our contribution is to recognize that the difficulty is theŽ .assumption of causality used by Kalman in his classical paper Kalman, 1960 . The

Kalman filter uses the idea of causality that there is a past causing a future in order tobreak down computations over time to achieve computational efficiency. But the spatialphysical and statistical relationships for groundwater are non-causal so the Kalman filtercan not break down the computations over space the way it does over time.

We have produced the first solution to the causality problem that generalizes theKalman filter to produce a spatial state estimation and system identification methodol-ogy that is computationally practical. The Kalman filter is generalized by introducing

Žinformation filter methods due to Bierman Bierman, 1977; Bierman and Porter, 1988;. ŽPorter, 1991 coupled with a Markov random field representation Whittle, 1954; Clark

.and Yuille, 1990; Chellappa and Jain, 1991 for spatial variation. A Bayesian penaltyfunction is implemented with Gauss–Newton methods to replace the non-linear estima-tion problem with a sequence of linear problems. This leads to a computational problem

Ž .similar to numerical simulation of the partial differential equations PDEs of groundwa-ter. In fact, extensions of PDE solver ideas to break down computations over space form

Ž .the computational heart of DFM. Consequently, realistic size three-dimensional 3Dgroundwater systems can be calibrated using DFM.

DFM includes a system identification theory based on statistical likelihood methods.In practice residual model fit errors provide diagnostic information for outlier detectionand for data and model validation. The statistical likelihood function needed to testvalidity of hypothesized model structures is computed from residual fit errors and state

Page 5: Data fusion modeling for groundwater systemsmmc2.geofisica.unam.mx/cursos/geoest/Articulos... · 2015. 10. 23. · Data Fusion Modelling DFM has value added as an engineering decision

( )D.W. Porter et al.rJournal of Contaminant Hydrology 42 2000 303–335 307

Ž .estimator outputs by generalizing the ideas of Bierman et al. 1990 . A maximumlikelihood theory is provided to estimate statistical parameters such as Markov modelparameters that determine the geostatistical variogram by generalizing ideas in Levy and

Ž .Porter 1992 .DFM draws on the mathematically related field of image restoration and reconstruc-

tion for an understanding of the roles of Bayesian penalty functions and MarkovŽ .representations Jain, 1981; Sezan and Tekalp, 1990 . Image restoration estimates an

ideal image from a blurred and noisy image. Spatial variability in the ideal image ismodeled as a Markov random field that is passed through a linear blur operator and anon-linear sensor response to be corrupted by measurement noise to produce the blurredand noisy image. A virtue of the Markov random field is that variability over any scalecan be modeled with only local equations in space. This provides analytical andcomputational advantages that make Markovian processes popular in statistical physicswhere they originated and in image processing where they have been more recently usedŽ .see Geman and Geman, 1984 .

Ž .As explained in the book about data fusion methods by Clark and Yuille 1990 ,Ž .image restoration can be accomplished using maximum a posteriori MAP methods. A

non-linear Bayesian penalty function is constructed with terms that penalize data fiterrors and excessive spatial variation according to the Markov random field model. Therestored image is the image that minimizes the penalty function, and simulated annealingmethods can be used to perform the minimization. The MAP estimator can alsoreconstruct image discontinuities using a binary line Markov process. The term thatpenalizes excessive spatial variability is relaxed at surfaces of discontinuity, and anadditional term is used to penalize inclusion of discontinuities.

DFM uses a Bayesian penalty function with a Markov term, but minimizes thepenalty function using information filter methods. A promising area for further researchis to directly estimate discontinuities such as faults or channels by including additionalpenalty terms. This would provide the basis for joint inversion of hydrological,geological, and geophysical data.

DFM draws on concepts and methods from the military and space science data fusionŽ .communities Bierman, 1977; Hall, 1992 . One of the key methods is information

filtering. The information filter reformulates the Kalman filter to work in terms ofŽ .information rather than covariance quantities Bierman, 1977 . The Kalman filter

problem is posed as a Bayesian least squares problem. The assumption of causality intime of physical and statistical models leads to a linear algebra problem that is sparseand banded. Then, linear algebra methods are used to breakdown the processing overtime. Filtered estimates of current states are obtained by a forward sweep through thedata, and smoothed estimates of past states using all the data are obtained by a backwardpass through the filter results.

Ž .A close examination of the information filter in Bierman 1977 reveals that the moregeneral assumption of Markov physical and statistical models also leads to a sparse andbanded linear algebra problem. The physical and statistical models for groundwaterproblems are not causal in space, but they do have a Markov representation. This meansthat sparse direct techniques using Householder reflections and iterative techniquesusing conjugate gradients can be used to provide computationally practical solutions for

Page 6: Data fusion modeling for groundwater systemsmmc2.geofisica.unam.mx/cursos/geoest/Articulos... · 2015. 10. 23. · Data Fusion Modelling DFM has value added as an engineering decision

( )D.W. Porter et al.rJournal of Contaminant Hydrology 42 2000 303–335308

DFM. The sparse methods avoid the excessive storage and computational demands thathave plagued the application of Kalman filter methods to groundwater problems.

The objective of DFM is to give modelers a computational engine with the flexibilityand computational efficiency to work with many data and model types for heteroge-neous field sites. It can be used to calibrate 3D 100,000 node systems with multipleaquifers of differing levels of spatial variability. The system can be heterogeneous with adifferent hydraulic conductivity estimated at each of the 100,000 nodes. In addition,trend and zonal parameters can be estimated, and measurement bias parameters can beestimated. Field experience has shown that Gauss–Newton methods with backtrackingor trust region techniques provide robust non-linear convergence. Modelers spend lesstime waiting for long computer runs, solving convergence problems, and wonderingabout the effects of simplifying assumptions. They spend more time with the models andthe data and more time working through the conceptual iteration process.

In other fields, one of the greatest benefits of Kalman filtering has been the ability toput all the physical and statistical modeling assumptions and the data in the computer toreveal inconsistencies in assumptions and data. Distinguishable patterns in residuals,residual outliers, or unrealistic estimates flag difficulties and often suggest to experi-enced modelers where to look for the difficulty. DFM provides this capability forgroundwater systems. This helps to determine data outliers and data problems such asbiases. Further, this helps to find inadequacies in physical and statistical models such asunmodeled faults and channels or physical effects that have been left out or are tooapproximate.

3. Mathematical development

3.1. Information filter methods

Information filter methods solve the same estimation problem that the Kalman filterŽsolves but work in terms of statistical information rather than covariance Bierman,

.1977 . Time domain techniques are explained here in preparation for extension to spatialestimation. The prior physical and statistical models for the state x and the model for themeasurements z over time ks0,1,2, . . . that are used by the information filter are asfollows:

x sF x qu qw ; z sH x qÕ 1Ž .kq1 k k k k k k k k

where x is the physical and statistical state vector with initial mean x and covarianceˆk 0

P , F is the state transition matrix, u is the known deterministic input, w is the white0 k k k

process noise with mean zero and covariance Q , H is the measurement sensitivityk k

matrix, and Õ is the white measurement noise with mean zero and covariance R .k kŽ .Eq. 1 is statistically normalized so all the equations have equal weight for

computational purposes. This is accomplished using the lower triangular Choleskyfactors C, C , and C for P , Q and R , respectively. Using the Cholesky factors1k 2 k 0 k k

Page 7: Data fusion modeling for groundwater systemsmmc2.geofisica.unam.mx/cursos/geoest/Articulos... · 2015. 10. 23. · Data Fusion Modelling DFM has value added as an engineering decision

( )D.W. Porter et al.rJournal of Contaminant Hydrology 42 2000 303–335 309

Ž . X X Xand rearranging Eq. 1 produces the following set of equations where w , w , and Õ allk k

have identity covariance.

C Tx sC T x qwX ; yC T F x qCT x sC T u qwX ;ˆ0 0 1k k k 1k kq1 1k k k

C T H x sC T z yÕX 2Ž .2 k k k 2 k k k

Ž .The physical and statistical models corresponding to the first two equations in Eq. 2Ž .are interpreted using the data equation idea of Duncan and Horn 1972 and Bierman

Ž .1977 . This means that the state is estimated by solving a regression problem where theterms dependent on u and x are taken to be measurements, the wX terms are taken toˆk o

be measurement noises, and the model dynamics on the left-hand-side of the equationare taken to be measurement sensitivity. Remarkably, the regression estimate andestimate error covariance are the same as the Kalman filter estimate and estimate error

Ž .covariance for the problem represented by Eq. 1 . But the mathematically equivalentregression problem is simpler conceptually and well established numerical methods existfor its efficient solution.

The regression estimate is formed by doing a least squares fit of the state to theŽ .measurements in Eq. 2 . This is done by solving the generally overdetermined system

Ž .that is the same as Eq. 2 except that there is no measurement noise on theright-hand-side. The system will be overdetermined if sufficient measurement equationsand prior physical and statistical model equations are available. In other words, theproblem will not be ill-posed if there are sufficient measurements and prior physical andstatistical information. Solution of the overdetermined system is achieved in thecomputer by manipulation of the information array given by

T T<C C x̂0

T T<C H C z20 0 20 0

T T T<yC F C C u10 0 10 10 0

T T<C H C z21 1 21 13Ž .T T T<yC F C C u11 1 11 11 1

T T<C H C z22 2 22 2

T T T<yC F C C u12 2 12 12 2. .<. .. .

A family of solution methods exist to take advantage of different problem structuresŽ .for computational efficiency Bierman, 1977; Porter, 1991 . Banded sparse solution

methods using Householder reflections and Givens rotations break down the computa-tions over time to achieve the Kalman filter solution. If the actual measurements areremoved, then the information filter solves the physical system. Consequently, theinformation filter can be viewed as a generalization of the numerical solver for thephysical system. When we generalize to spatial estimation, we take advantage of this touse extensions of PDE solvers to achieve computationally efficient estimation.

Ž .The system of equations represented by Eq. 1 is compatible with models used forŽ . Ž .hydrogeological applications in Van Geer et al. 1991 , Yangxiao et al. 1991 ,

Page 8: Data fusion modeling for groundwater systemsmmc2.geofisica.unam.mx/cursos/geoest/Articulos... · 2015. 10. 23. · Data Fusion Modelling DFM has value added as an engineering decision

( )D.W. Porter et al.rJournal of Contaminant Hydrology 42 2000 303–335310

Ž . Ž . Ž .McLaughlin 1995 , Zou and Parr 1995 , Eppstein and Dougherty 1996 andŽ .McLaughlin and Townley 1996 . Causal Markov processes that are correlated in time

and are parameters or inputs for the physical model can be adjoined to the state vectorwhere appropriate modifications are made to the state transition matrix, process noise,

Ž .and measurement sensitivity matrix Bierman, 1977 . Non-causal processes that arespatially correlated and are parameters or inputs for the physical model can also beadjoined to the state vector. However, the statistical normalization and solution using

Ž . Ž .Eqs. 2 and 3 can be prohibitive computationally for realistic field scale parameteriza-tion. In later sections, it will be shown that DFM generalizes previous results byadjoining Markov random field states to the state vector. The Markov random field leadsto a simple normalization and equations that are local in space for a computationallypractical solution.

3.2. Groundwater DFM theory

DFM is developed for a baseline groundwater system commonly encountered. This isdone in order to make the fundamental ideas concrete without being distracted byextraneous model details. The focus is on the combined study of point data for hydraulichead and conductivity that is so often emphasized in the hydrogeological literature.DFM theory has also been applied to data integration of geological data and geophysicaldata such as time-domain electromagnetic and seismic data to delineate geologicalboundaries.

DFM performs a non-linear iteration to minimize a Bayesian penalty function usingGauss–Newton methods. Within the Gauss–Newton method, we have implemented

Ž .options for back tracking Press et al., 1992 and the generalization of the Levinberg–ŽMarquardt approach known as the model trust region method Vandergraft, 1985; Press

.et al., 1992 . We also include inequality constraints to keep the state estimates physicallyreasonable. To make the ideas concrete, we present the simplest Gauss–Newton optionof replacing the non-linear problem with a sequence of locally linearized problems.

The locally linearized problem also has a number of solution options. We haveimplemented several methods that include triangularizing reflections and rotations using

ŽHouseholder reflections and Givens rotations Lawson and Hanson, 1974; Bierman,. Ž1977 and conjugate gradient related methods Paige and Saunders, 1982; Golub and

.Van Loan, 1989; Kightly and Forsyth, 1991 . The problem is also ideally suited toŽ .parallel processing using the ideas of Kung 1991 and using parallel versions of

conjugate gradient related methods. We will present the triangularizing reflection androtation methods. Sections 3.3, 3.5, 3.6, 3.7 and 3.8 describe the Markov random fielddata equations, Bayesian penalty function, state estimation, prediction uncertainty, andsystem identification.

The Gaussian assumption is often used in estimation methods. DFM Bayesian leastsquares state estimation does not require the Gaussian assumption. The wide-sense

Ž .assumption of Duncan and Horn 1972 is made that the first and second moments of thesources of randomness are well defined, but no distributional assumptions need be made.If the Gaussian assumption is made, then the DFM estimates are MAP estimates. DFMsystem identification does use the Gaussian assumption. However, our many years ofexperience with system identification methods suggest that the Gaussian approaches are

Page 9: Data fusion modeling for groundwater systemsmmc2.geofisica.unam.mx/cursos/geoest/Articulos... · 2015. 10. 23. · Data Fusion Modelling DFM has value added as an engineering decision

( )D.W. Porter et al.rJournal of Contaminant Hydrology 42 2000 303–335 311

robust, near-optimal, and can provide useful results even if the problem is not Gaussian.Furthermore, our experience is supported by theoretical results on approximate system

Ž .modeling Caines, 1978 .

3.3. MarkoÕ random field

The spatial autoregressive representation of a Markov random field for the spatialŽ . Žstochastic process f on a regular two-dimensional 2D grid is given by Whittle, 1954;.Clark and Yuille, 1990; Chellappa and Jain, 1991

f s a f qw 4Ž .Ýi j k , l iqk , jql i js

where a are Markov interpolation coefficients, w is Markov process noise that isk , l i, j

spatially white with variance q, and S is a local spatial region of support.Spatial correlation is modeled in terms of interpolation over a local neighborhood S

about the interpolation point. For a stationary random field with a rational powerspectrum, the interpolation error w is white over space for sufficiently high-orderinterpolation. The Markov random field extends to three dimensions. Furthermore, thecorrelation structure can be anisotropic with different correlation distances in differentdirections.

The Markov random field provides a general and simple way to model spatialvariability over any scale with only a local equation. This amazing property is behind its

Žuse in statistical physics and computer vision Geman and Geman, 1984; Clark and.Yuille, 1990 . We have extended the Markov model to a non-uniform slightly deformed

grid to be compatible with standard finite-difference and curvilinear finite-element gridsover which groundwater numerical models are often defined. Spatially varying Markovcoefficients are computed so the underlying correlation structure stays fixed as the gridvaries. This is accomplished by using a stochastic PDE representation for the Markovrandom field and computing the coefficients so the underlying structure stays fixed.

We use nested structures for spatial variability at different scales similarly toŽ . Ž .Kitanidis and Vomvoris 1983 and Carrera and Glorioso 1991 , using ideas fromŽ .Journal and Huijbregts 1991 . At the observational scale, measurement noise and

microstructure appear as white noise. At intermediate scales larger than observationalbut smaller than the aquifer, spatial variability is modeled as a stationary process usingthe Markov random field. At the aquifer scale, non-stationary effects are taken intoaccount with a deterministic trend having unknown coefficients that are estimated asadditional states in DFM. The distinctions between scales are problem and objectivedependent. If there are too few terms in the trend, then the Markov process will beforced to account for non-stationary behavior for which it is not well suited. On theother hand, too many terms in the trend leads to familiar problems with overparameteri-zation. Our experience has been that decisions about problems of scale are a natural andreasonably straightforward part of modeling conceptual iteration. To focus on incorpora-tion of the Markov random field and to keep the presentation simple, the trend terms arenot included in the ensuing description of the theory.

Page 10: Data fusion modeling for groundwater systemsmmc2.geofisica.unam.mx/cursos/geoest/Articulos... · 2015. 10. 23. · Data Fusion Modelling DFM has value added as an engineering decision

( )D.W. Porter et al.rJournal of Contaminant Hydrology 42 2000 303–335312

3.4. Data equations

The baseline groundwater system is steady-state flow with transient advective–dis-persive transport for which point measurements of hydraulic head, hydraulic conductiv-ity, and contaminant concentration are available and for which the log-conductivity isviewed as a Markov random field. Consider the finite-difference representation. Takingthe Gauss–Newton viewpoint and linearizing the equations about the current estimate,we obtain the data equations

A hqA fsq qw Steady Flow Model1 2 1 1

M hsz yÕ Head Measurements1 1 1

M fsz yÕ Conductivity Measurements2 2 2

A fsq qw Markov Random Field3 2 2

A C qA c qA hqA fsq qw Transient Transport Model4, tq1 tq1 5, t t 6, t 7, t 3, t 3, t

M c sz yÕ Contaminant Data3, t t 3, t 3, t

5Ž .where h is hydraulic head relative to the current estimate; f is log conductivity relativeto the current estimate; A , A , and A through A are banded sparse matrices for flow1 2 4 7

and transport models; A is the banded sparse matrix for the Markov model; q and q3 1 3

are sinkrsource and local linearization terms; q is the Markov linearization term; w ,2 1

w , and w are process noises with zero means and normalized to identity covariances;2 3

M , M , and M are measurement sensitivity matrices; z , z , and z are measured1 2 3 1 2 3

data and local linearization terms; Õ , Õ , and Õ are measurement noises with zero1 2 3

means and identity covariances; and c is contaminant concentration at time t relative tot

the current estimate.This problem will be solved and the original problem will be linearized about the

solution, and so forth, until convergence of the Gauss–Newton method to the solution ofthe non-linear problem. Boundary, initial conditions, and sinkrsource terms are assumedknown, although they can be estimated also. In addition, zonal parameters and measure-

Ž .ment bias parameters can be estimated. Since Eq. 5 is linearized, the states h, f , and care relative to the current state estimates about which the equations are linearized, i.e.,they are perturbations from the current estimates. Also, the q and z terms for sourcesand sinks and for measurements have lumped in with them linearization terms. The

Ž .Markov random field represented by Eq. 4 has been incorporated with a nomenclatureŽ .appropriate for Eq. 5 and a q term to account for referencing to the current estimate.

The measurement sensitivity matrices for the point data perform an interpolation fromthe grid where the states are defined to the points where the measurements are made.Data sets that are not point data such as seismic refraction for determining geologicalboundaries can be incorporated by expanding the measurement sensitivity matrices.

All process noises and measurement noises are normalized to identity covariance forlater computational reasons. Notice that identity covariance does not mean that there isno spatial correlation structure in the data equations. The Markov random field equationsproduce a spatially correlated structure from an identity covariance input. One of the keyideas behind Kalman filtering is that correlated processes can be produced by linear

Page 11: Data fusion modeling for groundwater systemsmmc2.geofisica.unam.mx/cursos/geoest/Articulos... · 2015. 10. 23. · Data Fusion Modelling DFM has value added as an engineering decision

( )D.W. Porter et al.rJournal of Contaminant Hydrology 42 2000 303–335 313

operations on uncorrelated processes. For example, if part of the noise input w to1

steady flow were spatially correlated, then it could be produced by an identity covari-ance input to Markov random field equations. The correlated part of w would be1

treated as part of the state vector. The resulting augmented data equations would haveŽ .the same structure as Eq. 5 and would still have identity covariance noises.

3.5. Bayesian penalty function

DFM estimation is accomplished by finding h, f , and c that minimize the Bayesianpenalty function, J, defined as follows for the linearized equations:

Ž .6

5 5where denotes the standard Euclidean norm. The terms in the penalty functionpenalize measurement fit errors to the models, physical model errors, and excessivespatial variability in log-conductivity. The penalty function can be viewed as mathemati-cally formalizing the criteria under which manual trial-and-error model calibration isperformed. There is no additional weighting of the terms in the penalty function sincethe measurement and process noises have already been normalized to identity covariance

Ž . Ž .in Eq. 5 . Eq. 6 represents the linearized least squares problem solved at eachGauss–Newton iteration.

The wide sense random assumption is made that the first and second moments of therandom terms are well defined, but no distributional assumptions are required. The widesense assumption is adequate to compute the estimate error covariance within locallinear perturbation as shown later in this section. If the Gaussian assumption is made,then within an additive constant the penalty function is the negative twice log of the aposteriori statistical distribution of the states given the measured data. Consequently,minimizing the Bayesian penalty function is equivalent to the non-linear MAP estimator.In other words, DFM computes the state estimates that have the highest probability ofbeing true given the measured data.

Ž .The minimization of the penalty function given by Eq. 6 is a least-squares problem.Ž .By inspection, the solution of Eq. 6 is equivalent to the solution of a regression

Ž .problem treating the model equations and the Markov random field equation in Eq. 5as if they were equations for measured data q in combination with the actual measured

Page 12: Data fusion modeling for groundwater systemsmmc2.geofisica.unam.mx/cursos/geoest/Articulos... · 2015. 10. 23. · Data Fusion Modelling DFM has value added as an engineering decision

( )D.W. Porter et al.rJournal of Contaminant Hydrology 42 2000 303–335314

Ž .data z. This equivalence is the data equation idea expressed in Eq. 2 that leads to theŽ .information array of Eq. 3 and a version of Kalman filtering over time. Consequently,

Ž .the physical and statistical model terms of Eq. 5 can be interpreted as data equationsthat are equivalent to measured data.

3.6. State estimation

In order to focus on the spatial aspects of estimation and on the combined study ofpoint measurements of head and conductivity, attention is now limited to steady flowwith head and conductivity measurements and with a Markov random field model forlog-conductivity. Using the kind of techniques for assembly that PDE solvers use, we

Ž .assemble the relevant parts of Eq. 5 in the banded sparse form:

hmy b

fmy b...

Eg hm m s z yÕm mfEh , f m...

hmq b 7Ž .fmq b

fmy b...

fL s q qwmm m m...

fmq b

where h and f are head and log-conductivity states in region m relative to the currentm m

estimate; g denotes non-linear physical model and measurement equations; z denotesm m

measurements, sources and sinks, and linearization terms that are all treated as measure-ments; L denotes Markov coefficients; q is the Markov linearization term; and Õm m m

and w are noises with zero mean and identity covariance.m

The states in a region m of space have been defined so that the data equations breakdown into blocks with a block bandwidth of b. For example, if a 2D grid is used withphysical and statistical models at any grid cell depending only on adjacent cells and withpoint data, then region m can be selected to be grid column m and the block bandwidthb is one.

Ž .The partial derivative is shown in Eq. 7 to emphasize the local linearization in thenon-linear Gauss–Newton method. The equation for the Markov random field is shownseparately to emphasize that Markov states have been adjoined to the state vector forspatial estimation in much the same way they are adjoined to states of the Kalman filterfor temporal estimation. For computational efficiency, the Markov equations are actuallyinterleaved with the model and measurement equations.

Page 13: Data fusion modeling for groundwater systemsmmc2.geofisica.unam.mx/cursos/geoest/Articulos... · 2015. 10. 23. · Data Fusion Modelling DFM has value added as an engineering decision

( )D.W. Porter et al.rJournal of Contaminant Hydrology 42 2000 303–335 315

Ž .For computer processing, Eq. 7 is expressed in an information array in which the xsdenote non-zero block matrices as shown below.

col.a myb m mqbØ

ØØzØ my 1

Øx Ø Ø x Ø x qmy 1x Ø Ø x Ø xØ z 8Ž .mx Ø Ø x Ø xØ

qmx Ø Ø x Ø xØzxx Ø Ø Ø Ø x mq 1

xx Ø Ø Ø Ø x qmq 1

Ø ØØ Ø

Ž .The regression solution of Eq. 7 is accomplished by manipulation of the informationŽ Ž ..array Eq. 8 .

The information array is processed by performing triangularizing rotations andŽ .reflections TR as illustrated below for a three-state case:

Ž .9a

Ž .9b

The problem is rotated and reflected until the solution, y , y , and y can be determinedˆ ˆ ˆ1 2 3

by inspection as follows:

zX zX yH y zX yHX y yHX yXˆ ˆ ˆ3 2 23 3 1 12 2 13 3 25 5y s , y s , y s ; Js e 9cŽ .ˆ ˆ ˆX X X3 2 1H H H33 22 11

TRs are applied to make the matrix on the left upper triangular. The correspondingpenalty function is shown on the right broken into a fixed part that involves an e-termand a variable part in terms of y. But the variable part can be set to zero, and no bettersolution is possible since the term can not be negative. Simple algebra reveals thesolution as backward substitution.

TR can be implemented using the numerically stable work horses of linear algebra,Householder and Givens transformations. Since TR transformations are non-singular,information is preserved. Because TR transformations rotate or reflect they do not

Page 14: Data fusion modeling for groundwater systemsmmc2.geofisica.unam.mx/cursos/geoest/Articulos... · 2015. 10. 23. · Data Fusion Modelling DFM has value added as an engineering decision

( )D.W. Porter et al.rJournal of Contaminant Hydrology 42 2000 303–335316

Ž .magnify numerical errors and are numerically stable Lawson and Hanson, 1974 .Furthermore, because they are orthogonal transformations they maintain the identity

Ž . Ž . Ž .covariance of the transformed noise terms shown in Eqs. 9a , 9b and 9c . The TRtransformations are implemented by direct manipulation of the information array for

Ž .computer implementation Bierman, 1977 .The covariance of the estimate error is determined by the sequence of relationships

H yX szX yÕX ; HX yszX ; HX yyy syÕ

X ;Ž .ˆ ˆT Xy1 XyTE yyy yyy sH H 10Ž . Ž . Ž .ˆ ˆ

The inverse of the HX matrix is performed in a numerically efficient and stable mannerŽ . Ž .using the same type of backward substitution scheme that was used in Eqs. 9a , 9b

Ž .and 9c for the estimate itself. Then squaring the result provides the estimate errorcovariance.

State estimation is broken down over space by proceeding recursively down theŽ .diagonal of the information array of Eq. 8 using a sequence of TRs to triangularize the

Ž . Ž .array as shown in Eqs. 11a and 11b , where xs and asterisks denote block matrices.Each recursion m starts with the triangularized array from the preceding recursion. Then

Ž .the data equation for region m from Eq. 8 is appended where new states are includedonly as they are encountered moving down the information array. The TR transforma-tion is applied to triangularize the array. Old states that are not going to be seen any

Ž . Ž .more moving down the information array and the starred arrays in Eqs. 11a and 11bthat are associated with them are left behind but saved.

Ž .11a

Page 15: Data fusion modeling for groundwater systemsmmc2.geofisica.unam.mx/cursos/geoest/Articulos... · 2015. 10. 23. · Data Fusion Modelling DFM has value added as an engineering decision

( )D.W. Porter et al.rJournal of Contaminant Hydrology 42 2000 303–335 317

Ž .11b

The estimates and covariances for the h and f states in the last region m at thebottom of the state vector are computed first using the bottom two rows of blocks in thetriangularized information array. Back substitution is used for the estimate as in Eqs.Ž . Ž . Ž .9a , 9b and 9c . The covariance is computed by inverting the block in the informa-

Ž .tion array and squaring it as in Eq. 10 .Then the h and f states just estimated are moved to the top of the state vector. The

estimates and covariances for the next h and f states back up the state vector arecomputed by retriangularizing the information array and repeating the same processingas above. The retriangularizing is efficiently broken down in space because of thebanded structure. This is repeated until all states and covariances are computed. Thenthe system is relinearized and the next Gauss–Newton iteration is performed. At thefinal Gauss–Newton iteration the estimate is obtained and the covariance of the estimateerror is obtained within linear perturbation.

3.7. Prediction uncertainty

Since DFM quantifies the uncertainty in groundwater flow, it can quantify predictionuncertainty for contaminant concentration. The direct approach of propagating thecovariance of estimate errors through transport is not computationally practical.

Monte Carlo methods are implemented in DFM that avoid propagating the covarianceŽ . Ž .of the estimate errors by directly using the information array of Eqs. 11a and 11b .

Ž .The sequence of relationships in Eq. 10 shows that the left-hand-side of the triangular-ized information array times the estimate error equals a spatially white noise withidentity covariance. Monte Carlo trials can be efficiently generated by sampling thenoise from a random number generator and back solving through the entire informationarray that includes all the starred arrays that were saved. This generates a realization of

Page 16: Data fusion modeling for groundwater systemsmmc2.geofisica.unam.mx/cursos/geoest/Articulos... · 2015. 10. 23. · Data Fusion Modelling DFM has value added as an engineering decision

( )D.W. Porter et al.rJournal of Contaminant Hydrology 42 2000 303–335318

state estimate errors that is added to the estimate to obtain a conditional simulation usingthe same idea that kriging uses to produce conditional simulations. To better capturenon-linear effects, conditional simulations are generated only for independent quantitiessuch as hydraulic conductivity. Then the dependent quantities in the state such as headare regenerated through Monte Carlo simulation with numerical models along with allthe other quantities of interest such as pathlines, travel times, and breakthrough curves.In the SRS application described in a later section, 1000 Monte Carlo trials weregenerated to produce confidence bounds for breakthrough curves for a complex highlyheterogeneous groundwater system.

3.8. System identification

DFM uses a combination of residual analysis and likelihood theory for systemidentification. The likelihood theory is similar to concepts in Kitanidis and VomvorisŽ . Ž .1983 and Carrera and Glorioso 1991 . Measurement outliers are rejected automati-cally based on an adaptive residual thresholding method for which the user sets inputparameters. Measurement residuals and physical and statistical data equation residualsare monitored for data and model validation. Under the Gaussian assumption and withinlinear perturbation, the minimized Bayesian penalty function should be statisticallydistributed as a x 2 random variable with degrees of freedom equal to the number ofactual scalar measurements. This does not include the physical and statistical dataequations which are treated as measurements only for forming state estimates andestimate error covariances.

Under the Gaussian assumption and within linear perturbation, the statistical likeli-hood function can be computed from information array quantities using relationships

Ž .from Bierman et al. 1990 . For the same measurements and models as Section 3.6, thenegative twice log-likelihood within an additive constant is given by

<Ž .L u sy2ln p z , z uŽ .Ž .1 2

Ž . Ž .L u s J of Eq. 6 evaluatedat state estimate

Žq2ln Product of scalar diagonals of the left-hand-side of triangularized information array in

Ž . .Eq. 11b including all starred arrays

Žy2ln Same as preceding term with information array computed with no measured data but

.linearized about same converged estimateŽ .y2ln Product of reciprocals of measurement noise standard deviations over all measured data

12Ž .

where z denotes head measurements; z denotes log conductivity measurements; u1 2

denotes a vector of unknown parameters in the Markov model, physical model, andŽ < .measurement noise model; and p z , z u is the probability density of measured data1 2

given u .Maximum likelihood estimates for parameters such as the Markov process noise

variance can be formed by minimizing L. This provides estimates for parameters thatmaximize the probability of receiving the measured data that is actually received.Maximum likelihood estimates have many desirable properties such as being asymptoti-cally unbiased, Gaussian, and minimum variance under mild regularity conditions.

Page 17: Data fusion modeling for groundwater systemsmmc2.geofisica.unam.mx/cursos/geoest/Articulos... · 2015. 10. 23. · Data Fusion Modelling DFM has value added as an engineering decision

( )D.W. Porter et al.rJournal of Contaminant Hydrology 42 2000 303–335 319

Modeler judgment and variogram methods from geostatistics provide model parame-ters. When more information is needed, maximum likelihood methods offer the virtue ofusing all the available information. For instance, the variogram for hydraulic conductiv-ity can be determined not only from measured conductivity data but from head data also.

Ž .Kitanidis and Vomvoris 1983 were the first to actually achieve this.Under the Gaussian assumption and within linear perturbation, DFM uses the

Ž . Ž .expectation maximization EM method of Levy and Porter 1992 to provide maximumlikelihood estimates. EM views the measured data as incomplete data for the purpose of

Ž .estimating the model parameters Dempster et al., 1977 . A complete data set isspecified from which the parameters could easily have been estimated if the completedata had been measured. If the incomplete data is a deterministic function of thecomplete data independent of the parameters being estimated and the relevant distribu-tions are not singular, then EM provides a simple iterative estimation procedure.

A sufficient statistic of the complete data is estimated in an expectation step based onthe previous estimate for the parameters. The parameters are updated based on thesufficient statistic in a likelihood maximization step and the process is iterated. EM hasthe desirable statistical convergence property that it will never make the likelihoodworse. Furthermore, it is usually easy to constrain the parameter estimates to bephysically reasonable. The EM method has been used in a host of agricultural, economic

Ž .and scientific applications Dempster et al., 1977 , and used for military, scientific andŽ .image processing applications Levy and Porter, 1992 .

The EM answers are usually intuitive and simple. For example, it may be desired toŽ .estimate the process noise variance q for a Markov process of Eq. 4 in order to model

one of the hydraulic conductivity fields in multiple geological layers. If the log-conduc-Ž .tivity f were directly measured noise free at all grid locations, then Eq. 4 could be

solved for the process noise w at all locations and q would be estimated as the samplevariance of the interpolation errors as follows:

21qs f y a f 13Ž .ˆ Ý Ýi j k , l iqk jq1½ 5N ij S

The complete data is defined as all values of f augmented by the actual noisy measureddata. The incomplete measured data is a deterministic function of the complete dataindependent of the parameter being estimated because the incomplete data is actuallypart of the complete data. Therefore, an EM iteration can be formulated.

It might seem reasonable to use the current estimate for q to estimate f , use theŽ .estimate for f in Eq. 4 to estimate w

ˆ ˆw s f y a f 14Ž .ˆ Ýi j i j k , l iqk jq1S

and then estimate q as the sample variance of the estimate of w. This is almost theanswer that EM provides, but it would be biased downward due to estimate error in w.EM removes the bias by adding in the estimate error variance in w computed by theestimator to give the estimate for q:

1 22qs w qE w yw 15Ž .ˆ ˆ ˆŽ .Ýž /i j i j i jN ij

Page 18: Data fusion modeling for groundwater systemsmmc2.geofisica.unam.mx/cursos/geoest/Articulos... · 2015. 10. 23. · Data Fusion Modelling DFM has value added as an engineering decision

( )D.W. Porter et al.rJournal of Contaminant Hydrology 42 2000 303–335320

The above procedure is iterated to convergence. Notice how simple the estimate is andhow it is naturally constrained to have non-negative values. This is typical for EMestimates for the model parameters that need to be estimated.

4. Application at DOE Savannah River Site

ŽIn conjunction with the VAM3DF finite-element formulation Huyakorn and Panday,.1995 , the DFM approach has been used to construct a groundwater flow and transport

Ž .model of the Old Burial Ground OBG at the U.S. Department of Energy’s SavannahŽ .River Site SRS . The resulting DFM model was compared to an existing model that

was calibrated via the typical trial-and-error manual calibration method. The OBG waschosen because a substantial amount of hydrogeologic information is available, and thecalibration and numerical issues were challenging with standard approaches. The DFM

Ž .flow model developed here is similar to the flow model by Flach et al. 1996 , which isŽ .based on the FACT finite-element code Hamm et al., 1995 . This allows comparison of

the two flow models and validates the utility of DFM.The SRS occupies approximately 310 mile2 along the Savannah River in southwest-

ern South Carolina. Since 1950, SRS has been a controlled area for the production ofnuclear material for national defense and civilian purposes. The OBG is located in the

Ž .central portion of SRS within the General Separations Area GSA . The OBG is theŽ .original solid radioactive and hazardous waste burial ground in the GSA Fig. 1 . Solid

waste was deposited in the OBG from about 1952 to 1972. Groundwater contaminantsŽ .of concern at the OBG include tritium, trichloroethylene TCE and tetrachloroethylene

Ž .PCE . Groundwater plumes emanating from the OBG are migrating to the south towardŽ .the Fourmile Branch Flach et al., 1996 . The contaminant of interest for this study is

tritium, because it is a geochemically conservative tracer that has been monitored alongthe seepline near the F-Area effluent and Fourmile Branch for several years. Fig. 2shows the conceptual model for the OBG from the horizontal perspective. Fig. 3 showsthe classical layer-cake hydrostratigraphy that has been used in the past in the vicinity ofthe OBG and the conceptual model vertical perspective.

The development of the DFM flow and transport model of the OBG relied upon theŽ . Ž .work of Westinghouse Savannah River Company WSRC . Flach et al. 1996 presented

the hydrogeologic information on the OBG and developed the FACT flow model.Detailed lithologic data was used to construct an extremely heterogeneous conductivityfield vs. the typical layer-cake approach that had been utilized in the past. Flach et al.Ž .1996 presented a model of the OBG tritium source, which was used in the transportsimulations for this study.

4.1. Numerical grid and boundary conditions

The DFM numerical grid is a 3D curvilinear grid with an areal extent of 6760 ft inŽ .the x-direction and 5070 ft in the y-direction Fig. 1 . The grid is 27=27=21 with

uniform spacing along the x- and y-axes. This is the same region considered by Flach etŽ .al. 1996 . Fig. 1 includes the local base map. Hydrostratigraphic picks obtained from

Page 19: Data fusion modeling for groundwater systemsmmc2.geofisica.unam.mx/cursos/geoest/Articulos... · 2015. 10. 23. · Data Fusion Modelling DFM has value added as an engineering decision

( )D.W. Porter et al.rJournal of Contaminant Hydrology 42 2000 303–335 321

Fig. 1. Areal view of model mesh overlaid on local base map.

WSRC were used to define the elevation of the 21 nodal layers, so that each nodal layerwas assigned to only one aquifer unit. The upper nodal layer conformed to the landsurface while the bottom nodal layer conformed to the top of the Ellenton Clays, whichis a regional aquitard at the SRS. The Green Clay was discretized by the fourth nodallayer from the bottom.

The boundary conditions applied to DFM flow model are very similar to boundaryŽ .conditions in Flach et al. 1996 . The rechargerdrain boundary condition was applied to

the top nodal layer and the no-flow boundary condition was applied to the interior nodesof the bottom nodal layer, which roughly approximates a no-flow condition across theEllenton. All other boundary nodes were assigned constant head boundary conditions.Heads from observation wells were used to construct an areal 2D head distribution forthe Water Table, BarnwellrMcBean and Congaree Aquifers. These three head distribu-tions were used to define the constant head boundary conditions.

4.2. Head and conductiÕity data

Heads from 237 observation wells were calibration targets for the DFM flow model.In order to facilitate a comparison between the DFM flow model and the flow model in

Ž .Flach et al. 1996 , conductivity field information used in their report was used asconductivity data. Mud fraction data from the foot-by-foot description of 84 cores were

Page 20: Data fusion modeling for groundwater systemsmmc2.geofisica.unam.mx/cursos/geoest/Articulos... · 2015. 10. 23. · Data Fusion Modelling DFM has value added as an engineering decision

( )D.W. Porter et al.rJournal of Contaminant Hydrology 42 2000 303–335322

Ž .Fig. 2. Overall flow and transport conceptual model horizontal perspective .

Ž .interpolated on to a fine-scale 3D grid 23=23=251 . The fine-scale mud fraction wasŽ . Ž .mapped to horizontal conductivity K and vertical conductivity K using anh v

empirical function. Mud fraction was used to assign one foot thick sediment intervals toŽ . Ž .four conductivity categories: Sand 0–10% mud , Clayey Sand 10–25% mud , Sandy

Ž . Ž .Clay 25–50% mud , and Clay 50–100% mud . Conductivity values for each classŽ .were estimated through calibration in the preceding modeling effort Flach et al., 1996 .

Ž .Fig. 3. Overall flow and transport conceptual model vertical perspective .

Page 21: Data fusion modeling for groundwater systemsmmc2.geofisica.unam.mx/cursos/geoest/Articulos... · 2015. 10. 23. · Data Fusion Modelling DFM has value added as an engineering decision

( )D.W. Porter et al.rJournal of Contaminant Hydrology 42 2000 303–335 323

For horizontal conductivity the values were: Sand, 40 ftrday; Clayey Sand, 5 ftrday;Sandy Clay, 0.15 ftrday, and Clay 0.0003 ftrday. Vertical conductivities were set to1r3 of the horizontal component, based on laboratory permeability data that showed anaverage anisotropy ratio of 3. The fine-scale conductivities on the 23=23=251 grid

Ž .were then converted to the vertically coarser 53=40=30 grid using arithmetic K hŽ .and harmonic K averaging. Of the 84 cores used in interpolation, 27 lay inside thev

modeled OBG area. The conductivities above the Green Clay at the 27 interior corelocations of the 53=40=30 grid were used as conductivity measurements for the DFMflow model. Both the vertical and horizontal conductivities were used as measurements.A total of 880 conductivity measurements were used with 415 being horizontalconductivities and 465 being vertical conductivities.

4.3. Parameters estimated in DFM

The Markov spatial variability model for hydraulic conductivity is applied to eachhydrostratigraphic layer separately in the DFM. Thus the fusion OBG model consistedof three hydrostratigraphic layers. Layer 1 contained all nodes above the Green Clay.Layer 2 contained the fourth nodal layer from the bottom, which represented the GreenClay. Layer 3 contained the bottom three nodal layers, which represented the CongareeAquifer and the top of the Ellenton Clays.

Ž .Head and log horizontal conductivity, ln K were estimated at each node in thehŽ .model where ln K is treated as a Markov random field. The standard deviation andh

Ž . Ž .correlation distances distance at which correlations1re for ln K for each hydros-h

tratigraphic layer are shown in Table 1. The coefficients of the polynomial modelŽ .defining the mean trend for ln K within each layer are also estimated. The priorh

standard deviation for each polynomial coefficient is specified in Table 1. In the OBGcase, there did not appear to be a significant change in the trend across the OBG, so the

Ž .trend model was limited to a constant mean value.The DFM methodology allows considerable flexibility in specifying and estimating

vertical hydraulic conductivity: log anisotropy can be defined by a spatial polynomial inŽa hydrostratigraphic unit, or vertical conductivity can be defined for specified zones sets

.of nodes . In either case, the coefficients can be fixed or estimated separately for eachzone. Within hydrostratigraphic layer 1, the vertical anisotropy was initially defined

Table 1Prior on horizontal and vertical conductivities, K and Kh v

Ž . Ž . Ž . Ž .Hydrostratigraphic ln K trend ln K heterogeneity ln K rK ln Kh h v h v

layerMean s s Horizontal Vertical Mean s Mean s

correlation correlationlengths lengths

1. Water Tabler 1.70 0.30 0.50 1000 20 4.0 0.3 – –Barnwell–Mcbean

2. Green Clay y10.41 0.20 0.01 1000 100 – – y12.21 0.303. Congaree 3.689 0.30 0.3 1000 50 1.1 – – –

Page 22: Data fusion modeling for groundwater systemsmmc2.geofisica.unam.mx/cursos/geoest/Articulos... · 2015. 10. 23. · Data Fusion Modelling DFM has value added as an engineering decision

( )D.W. Porter et al.rJournal of Contaminant Hydrology 42 2000 303–335324

Ž .using a linear 3D polynomial four coefficients , but the estimated linear coefficientswere small, so the anisotropy model was reduced to a constant. The vertical conductivityin the tan clay was defined separately and is discussed below. In hydrostratigraphic layerŽ .2 Green Clay , vertical conductivity was estimated directly, rather than using anisotro-

py. This was done because vertical flow dominates in aquitards and thus the relationbetween vertical and horizontal conductivity would not be observable from the data.

Ž .Finally, the anisotropy in layer 3 Congaree was somewhat arbitrarily fixed at 3, sincethere is little well information in this layer and model results are not sensitive to thisparameter.

In addition to conductivity parameters, the maximum recharge coefficient was alsoestimated in the DFM. This ‘‘maximum recharge coefficient’’ is the maximum allowedrecharge at an unsaturated surface node as used in a rechargerdrain boundary condition.This must be estimated because it has a strong influence on the flow model.

4.4. Parameter prior statistics

Ž .Table 1 lists the prior means, standard deviations s as well as correlation lengthsŽ .c.l. used for parameters estimated in the OBG data fusion modelling. The values used

Ž .for the means were based on information in Flach et al. 1996 . Standard deviations andcorrelation lengths were selected as reasonable estimates for the variation in theparameters, with some modifications if the data would not support estimation of a

Ž Ž .parameter e.g., the standard deviation of the Green clay ln K random component wash.set to 0.01 .

The prior information used for the maximum recharge coefficient was means17.0in.ryear and standard deviations0.88 in.ryear. Developed or capped areas used acoefficient of zero. Other input parameters included an effective porosity of 0.25 and aresidual saturation used in the pseudo-soil function of 0.01.

4.5. Incorporating the Tan Clay

Ž .Flach et al. 1996 discussed how mud fraction data from 84 cores in GSA were usedto construct a conductivity field on a 53=40=30 grid. FORTRAN code was acquiredfrom WSRC for construction of the conductivity field on the 27=27=21 grid. Nodesin the vicinity of the Tan Clay with K F0.0025 ftrday were identified as Tan Clayv

Ž . Ž .nodes. For these Tan Clay nodes, ln K was set equal to y8.0 ln ftrday in the DFMv

flow model.

4.6. DFM parameter estimates

Ž .The VAM3DF Huyakorn and Panday, 1995 flow model option of DFM was used tocompute the steady-state flow model. The resulting DFM estimated heads and horizontalconductivities are shown in Figs. 4 and 5, respectively. Figs. 4 and 5 show two cut-awayviews with corresponding data. Because the conductivity correlation distances input tothe DFM were relatively large, the DFM estimated conductivity field is smoother than

Page 23: Data fusion modeling for groundwater systemsmmc2.geofisica.unam.mx/cursos/geoest/Articulos... · 2015. 10. 23. · Data Fusion Modelling DFM has value added as an engineering decision

( )D.W. Porter et al.rJournal of Contaminant Hydrology 42 2000 303–335 325

Fig.

4.E

stim

ated

head

and

sele

cted

head

mea

sure

men

ts.

Page 24: Data fusion modeling for groundwater systemsmmc2.geofisica.unam.mx/cursos/geoest/Articulos... · 2015. 10. 23. · Data Fusion Modelling DFM has value added as an engineering decision

()

D.W

.Porter

etal.r

JournalofC

ontaminantH

ydrology42

2000303

–335

326

Fig. 5. Estimated log horizontal conductivity and selected measurements.

Page 25: Data fusion modeling for groundwater systemsmmc2.geofisica.unam.mx/cursos/geoest/Articulos... · 2015. 10. 23. · Data Fusion Modelling DFM has value added as an engineering decision

( )D.W. Porter et al.rJournal of Contaminant Hydrology 42 2000 303–335 327

Ž .Fig. 6. 3D Pathlines starting at zs240 ft and head contours at zs178 ft horizontal perspective .

Ž .the conductivity field in Flach et al. 1996 , but has similar trends. Figs. 6 and 7 show anareal view and side view, respectively, of 3D pathlines. The pathlines originating near

Ž .Fig. 7. 3D Pathlines starting at zs240 ft and head contours at xs2600 ft vertical perspective .

Page 26: Data fusion modeling for groundwater systemsmmc2.geofisica.unam.mx/cursos/geoest/Articulos... · 2015. 10. 23. · Data Fusion Modelling DFM has value added as an engineering decision

( )D.W. Porter et al.rJournal of Contaminant Hydrology 42 2000 303–335328

the OBG start in the vadose zone, travel through the saturated zone and discharge on thetop of the model near the Fourmile Branch. The DFM pathlines from the OBG to the

Ž .F-Area effluent are similar to the pathlines presented by Flach et al. 1996 . Also, theseeplines surrounding the Fourmile Branch and the F-Area effluent are similar to the

Ž .seeplines in Flach et al. 1996 . The main difference between the two flow models isŽ .that the water divide plane at which the flow separates below the OBG is further south

Ž . Žcloser to the Fourmile Branch in the fusion model, which is due to the lower on.average horizontal conductivities in the water table aquifer of the DFM. There are also

fine-scale differences in the velocity field, which are due to differences in the conductiv-ity fields.

Estimated values of the hydraulic conductivity parameters are shown in Table 2. Notethat most of the parameter estimates are close to the prior values, except for meanŽ . Ž .ln K in the upper layer, which is slightly higher, and ln K in the Green clay, whichh v

is slightly smaller.The DFM estimated maximum recharge was 15.31 in.ryear with an a posteriori

1ys uncertainty of 0.67 in.ryear. This flow model computed a discharge to the3 Ž 3 . ŽFourmile Branch of 119,773 ft rday 1.4 ft rs and the average recharge total

.volumetric inflowrarea of the top of the model was 12.3 in.ryear. In Flach et al.Ž .1996 , the maximum recharge was set equal to 17 in.ryear. The resulting discharge toFourmile Branch was 129,600 ft3rday and the average recharge was 13.9 in.ryear.

ŽThus the DFM estimate of discharge is closer to previous estimates based on measure-. 3ments of 115,000 ft rday.

The a posteriori estimate error standard deviations for most of the nodal heads in theŽ .central region in layer 1 were about 0.20–0.30 ft. The corresponding nodal ln K h

uncertainty was about 0.35–0.40, which is somewhat smaller than the prior 1ys ofŽ .0.58 including trend and variability , but is still significant. In addition to computing

estimate error standard deviations, the DFM also computes the correlations betweenstates. Surprisingly, very few correlation coefficients were greater than 0.3, except for

Žthe conductivity parameters within the Green clay or Congaree where little data was.available . The most significant correlations were between the layer 1 anisotropy and the

heads at the nodes near the Fourmile Branch. There was also significant correlationbetween recharge and the conductivities in the same region. This correlation is expectedbecause the discharge of water gained by the system through recharge is stronglyinfluenced by K near the stream.v

Table 2DFM estimated horizontal and vertical conductivities

Ž . Ž . Ž .Layer ln K trend ln K rK ln Kh v h v

Mean s Mean s Mean s

1. Water Tabler 1.87 0.09 3.74 0.10 – –Barnwell–Mcbean

2. Green Clay y10.42 0.20 – – y13.57 0.203. Congaree 3.69 0.30 – – – –

Page 27: Data fusion modeling for groundwater systemsmmc2.geofisica.unam.mx/cursos/geoest/Articulos... · 2015. 10. 23. · Data Fusion Modelling DFM has value added as an engineering decision

( )D.W. Porter et al.rJournal of Contaminant Hydrology 42 2000 303–335 329

4.7. Residuals and edited data

The DFM software edits any measurements for which the normalized residualŽ .residual divided by nominal accuracy is greater than a user specified threshold

Ž .multiplying the root-mean-squared RMS normalized residual from the previous itera-tion. The user must also input an initial residual RMS to use on the first iteration. Thus,the editing threshold is very high on early iterations and becomes gradually smaller inlater iterations as the fit to the data improves. The editing threshold used in the OBGmodeling was set equal to 4.0 and the initial residual RMS was set to 200. At the finaliteration, the normalized residual RMS was 1.7 with 1117 measurements processed. Thisindicates that the measurement errors are only slightly larger than the expected accuracy.

DFM rejected nine head measurements as outliers. DFM accepted 228 head measure-Ž .ments with an RMS not normalized of 1.3 ft, 415 horizontal conductivity estimatesŽ .with an RMS of 1.3 ln ftrday , and 465 vertical conductivity estimates with an RMS of

Ž . Ž .3.0 ln ftrday . For the flow model developed in Flach et al. 1996 , the RMS for the228 accepted head measurements was 3.0 ft.

4.8. Transport model

The numerical grid used for the transport model was identical to the DFM flowmodel grid. The tritium migration was simulated for approximately 80 years, from 1956to 2034. To minimize run time, the model was tested to determine the smallest numberof time steps that could be used without affecting the model results. The transport model

Ž .was executed using 780 time steps and data was output every 36.5 days 0.1 year .Ž .Values of longitudinal horizontal dispersivity a , longitudinal vertical dispersivityLH

Ž . Ž . Ž .a , transverse horizontal dispersivity a , and transverse vertical dispersivity aLV TH TV

were 20, 2, 5 and 2 ft, respectively. These dispersivity values were determined asoptimum by trial-and-error calibration of the transport model with the measured dataŽ .tritium discharged to Fourmile Branch acquired from WSRC. An effective porosity of25% was utilized for the transport simulations in this study. The apparent molecular

Ž . y06 2 Ž .diffusion coefficient D8 was 1.0=10 ft rday. The distribution coefficient Kd

was assumed equal to zero which translates to a retardation coefficient equal to 1.0,which is a common assumption for tritium. The decay coefficient was calculated using:ls ln2rt , where t is the half life in days. The half-life for tritium is 12.3 years.1r2 1r2

The transport model had zero initial concentration and zero mass flux boundaryconditions for all boundary inflow nodes. After solving for steady-state flow, flux valueson the boundary were utilized for determination of boundary conditions. All nodes with

Ž .positive fluxes inflow on the boundary were assigned a mass flux of zero and aŽ .volumetric flux equal to the value determined by the flow model. Flach et al. 1996

investigated the waste forms deposited in the OBG and developed a spatially andtemporally variable tritium discharge model. The Flach model was used to produce timedependent mass flux boundary conditions at 90 nodes, which are in the vadose zonebelow the OBG. The leaching constant in the Flach model was used to calibrate thetransport model to the measured tritium discharge data. The calibrated source model wasused as the mean source.

Page 28: Data fusion modeling for groundwater systemsmmc2.geofisica.unam.mx/cursos/geoest/Articulos... · 2015. 10. 23. · Data Fusion Modelling DFM has value added as an engineering decision

( )D.W. Porter et al.rJournal of Contaminant Hydrology 42 2000 303–335330

Fig. 8. Source scale factor histogram.

4.9. Tritium transport uncertainty Õia Monte Carlo analysis

To quantify the effects of recharge, conductivity, and source uncertainty on tritiumtransport, a Monte Carlo transport simulation was used to compute the distribution oftritium concentration at specified locations. For each realization, the DFM estimate of

Ž Ž . .the flow states nodal heads, log conductivities, anisotropy, ln K and recharge wasv

used as the sample mean, and a random perturbation consistent with the DFM errorŽcovariance was added. This is done by backsolving using the information array and a

.Gaussian random vector as the right-hand-side. Thus the flow state sample realizationshave the mean and covariance expected from the flow modeling. The nodal headsamples were not used directly in the transport integration because any flow massimbalance will cause instability. Thus the head samples were input to VAM3DF tore-compute the steady-state heads. It was observed that the re-computed heads did notdiffer significantly from the sample values, probably because the information arraymodels the flow equation accurately within the limits of a locally linear model.

Source uncertainty was treated as linear scaling of the nominal source time profile.Thus the effect of source uncertainty was incorporated in post processing using randomsamples for the scaling factor.

4.10. Monte Carlo results

Using 1000 Monte Carlo realizations of flow and transport parameters as describedabove, the distribution of tritium discharge to the Fourmile Branch was computed.

Fig. 9. Horizontal conductivity histogram.

Page 29: Data fusion modeling for groundwater systemsmmc2.geofisica.unam.mx/cursos/geoest/Articulos... · 2015. 10. 23. · Data Fusion Modelling DFM has value added as an engineering decision

( )D.W. Porter et al.rJournal of Contaminant Hydrology 42 2000 303–335 331

Fig. 10. Tritium discharge to Fourmile Branch with recharge, conductivity, and source uncertainty.

Source uncertainty was incorporated by scaling the mean source. The source scale factorhad a clipped normal distribution between 0.5 and 1.5 with Ms1.0 and SDs0.25Ž .Fig. 8 . Since the transport model solves a linear transport equation with linearboundary conditions, the source uncertainty was incorporated by scaling the output

Ž .concentrations. Fig. 9 shows a histogram of ln K at a location east of the F-Areah

effluent and 38 ft below the surface. Note that the conductivity uncertainty is significant.To construct breakthrough curves for a given confidence bound, the discharge realiza-tions were sorted for each time, and a discharge was output corresponding to the givenconfidence bound. Breakthrough curves were determined using confidence bounds of95%, 50%, and 5% and were normalized with respect to the maximum of the 95%confidence bound. Fig. 10 shows the tritium discharge curves superimposed on indepen-dent field data, providing validation for the confidence bounds.

5. Application summary

The DFM flow model is generally consistent with the previous model of Flach et al.Ž .1996 that was obtained using the trial-and-error method. The models have similarpathlines, recharge, and trends in conductivity. The conductivity field for DFM issomewhat smoother, reflecting spatial correlation parameters input to the fusion process-ing. The spatial correlation parameters are based on hydrogeological judgement and onthe fit to the data. The benefits of DFM come mainly from the ability to rapidly combinediverse sources of information to quantify and reduce uncertainty. The specific benefitsdemonstrated in the work being reported here are the following.

Ø Validated quantification of statistical uncertainty in tritium breakthrough curvesdue to uncertainty in recharge, hydraulic conductivity, and source terms. This shows thatuncertainty can be quantified at complex sites such as the OBG, opening the possibilityfor a more precise level of risk assessment at other sites. Further, this shows that the

Page 30: Data fusion modeling for groundwater systemsmmc2.geofisica.unam.mx/cursos/geoest/Articulos... · 2015. 10. 23. · Data Fusion Modelling DFM has value added as an engineering decision

( )D.W. Porter et al.rJournal of Contaminant Hydrology 42 2000 303–335332

uncertainty can be quantified in a model that is of sufficient detail to be used forremediation design so safety margins can be quantified.

Ø Rapid updating of the model. DFM reduced the time spent on calibration.Modeling conceptual iteration is accelerated in order to get the best final model,particularly with the use of GUI and visualization tools. This gives the hydrogeologistmore control over the modeling process, freed from much of the manual manipulation ofthe model required by trial-and-error. Further, this means that DFM could be used foron-line model-based monitoring and adjustment of remediation. DFM calibration runsfor the OBG converge in less than 1 h on a Pentium Pro PC for a 3D model with morethan 15,000 nodes. Run time is approximately linear with the number of nodes.

Ø Better RMS fit to the data. The RMS head error for DFM was 1.3 ft after editing asmall number of statistical outliers. The comparable RMS for trial-and-error was 3.0 ft.The model also produced a significantly better stream flow prediction, even thoughmeasurement of stream flow was not included in the calibration.

6. Conclusions

DFM builds on previous hydrogeological work to provide a mature methodologyimplemented and deployed in software tools for data integration, model calibration, andgeostatistical quantification of uncertainty. Key points are as follows.

Ø DFM solves the causality problem with Kalman filtering to produce a computa-tionally practical spatial state estimation methodology for model calibration.

Ø DFM provides a solution to the problem often referred to in the hydrogeologicalliterature of combining point data for hydraulic head and conductivity to take advantageof the considerable data worth that these data sets often possess.

Ø DFM provides a system identification theory for the determination of geostatisticalstructure based on maximum likelihood methods.

Ø DFM has been applied to several field-scale groundwater systems. An applicationis presented at the DOE SRS where DFM quantified breakthrough curve uncertainty andproduced a better data fit with less time spent relative to a conventional calibration.

Ø DFM is implemented in a software system complete with a GUI and visualizationand installed for a users group at the SRS.

Solving the causality problem for spatial estimation lays the groundwork for furtherextensions in hydrogeology and in other fields with similar spatial estimation andidentification problems. One of the possible extensions is the integration with manage-ment optimization methods for environmental remediation, particularly for quantifyingsafety margins and managing uncertainty. DFM opens the possibility for real-timemonitoring and optimization through rapid model updating for expedited cleanup.

Acknowledgements

This work was performed for the Department of Energy. We are deeply appreciativeof the support that we received from Caroline Purdy at DOE Headquarters, without

Page 31: Data fusion modeling for groundwater systemsmmc2.geofisica.unam.mx/cursos/geoest/Articulos... · 2015. 10. 23. · Data Fusion Modelling DFM has value added as an engineering decision

( )D.W. Porter et al.rJournal of Contaminant Hydrology 42 2000 303–335 333

whom this work would never have been done. We are indebted to John Steele at theŽ .Westinghouse Savannah River Company WSRC for having the management vision to

Žmake the Old Burial Ground application happen. We wish to thank John Haselow who.was with WSRC at the beginning of this work and is now with Haselow Engineering

for his important technical support. We wish to express our appreciation for theexcellent work performed by the technical staff at Coleman Research — Dan Chapman,Bill Hatch, Linda Hughes, Paul Kraus, Tim Mann, Jay Silverman, and Bill Yancey —and by the technical staff of HydroGeoLogic — Sorab Panday and Sree Tangella. Thiswork follows in the footsteps of departed colleagues and friends Jerry Bierman and JimVandergraft. The information filter ideas for Data Fusion Modeling came from livelydiscussions between Jerry Bierman, Dave Porter, and Bruce Gibbs, and Jim Vandergraftprovided critical optimization ideas for Bayesian estimation.

References

Bennett, A.F., 1992. Inverse Methods in Physical Oceanography. Cambridge Univ. Press, Cambridge.Bierman, G.J., 1977. Factorization Methods for Discrete Sequential Estimation. Academic Press, New York.

Ž .Bierman, G.J., Porter, D.W., 1988. Decentralized Tracking via New Square Root Information Filter SRIFConcepts. Business and Technological Systems, BTS-34-87-32.

Bierman, G.J., Belzer, M.R., Vandergraft, J.S., Porter, D.W., 1990. Maximum likelihood estimation usingŽ .square root information filters. IEEE Trans. Autom. Control 35 12 , 1293–1298.

Caines, P.E., 1978. Stationary linear and nonlinear system identification and predictor set completeness. IEEEŽ .Trans. Autom. Control AC 23 4 , 583–594.

Carrera, J., Glorioso, L., 1991. On geostatistical formulations of the groundwater flow inverse problem. Adv.Ž .Water Resour. 14 5 , 273–283.

Carrera, J., Neuman, S.P., 1986a. Estimation of aquifer parameters under transient and steady state conditions:Ž .1. Maximum likelihood method incorporating prior information. Water Resour. Res. 22 2 , 199–210.

Carrera, J., Neuman, S.P., 1986b. Estimation of aquifer parameters under transient and steady state conditions:Ž .2. Uniqueness, stability, and solution algorithms. Water Resour. Res. 22 2 , 211–227.

Carrera, J., Neuman, S.P., 1986c. Estimation of aquifer parameters under transient and steady state conditions:Ž .3. Application to synthetic and field data. Water Resour. Res. 22 2 , 228–242.

Chellappa, R., Jain, A., 1991. Markov Random Fields: Theory and Application. Academic Press, Boston.Clark, J.J., Yuille, A.L., 1990. Data Fusion for Sensory Information Processing Systems. Kluwer Academic

Publishers, Boston.Cooley, R.L., 1982. Incorporation of prior information on parameters into nonlinear regression groundwater

Ž .flow models. Water Resour. Res. 18 4 , 965–976.Courtier, P., Derber, J., Errico, R., Louis, J.F., Vukicevic, T., 1993. Important literature on the use of adjoint,

Ž .variational methods and the Kalman filter in meteorology. Tellus 45A 5 , 342–357.Cressie, N., 1991. Statistics for Spatial Data. Wiley, New York.Dagan, G., 1985. Stochastic modeling of groundwater flow by unconditional and conditional probabilities: the

Ž .inverse problem. Water Resour. Res. 21 1 , 65–72.Daniel, M.M., Willsky, A.S., 1997. A multiresolution methodology for signal-level fusion and data assimila-

Ž .tion with applications to remote sensing. Proc. IEEE 85 1 , 164–180.Dempster, A.P., Laird, N.M., Rubin, D.B., 1977. Maximum likelihood from incomplete data via the EM

Ž .algorithm. J. R. Stat. Soc., Ser. B. 39 1 , 1–22.Duncan, D.B., Horn, S.D., 1972. Linear dynamic recursive estimation from the viewpoint of regression

Ž .analysis. J. Am. Stat. Assoc. 67 340 , 815–821.Eppstein, M.J., Dougherty, D.E., 1996. Simultaneous estimation of transmissivity values and zonation. Water

Ž .Resour. Res. 32 11 , 3321–3336.

Page 32: Data fusion modeling for groundwater systemsmmc2.geofisica.unam.mx/cursos/geoest/Articulos... · 2015. 10. 23. · Data Fusion Modelling DFM has value added as an engineering decision

( )D.W. Porter et al.rJournal of Contaminant Hydrology 42 2000 303–335334

Flach, G.P., Hamm, L.L., Harris, M.K., Thayer, P.A., Haselow, J.S., Smits, A.D., 1996. Groundwater FlowŽ .and Tritium Migration from the SRS Old Burial Ground to Fourmile Branch U , Westinghouse Savannah

River Company. Technical Report No. WSRC-TR-96-0037.Freeze, R.A., Massmann, J., Smith, L., Sperling, T., James, B., 1990. Hydrogeological decision analysis: 1. A

Ž .framework. Ground Water 28 5 , 738–766.Ž .Gelhar, L.W., 1986. Stochastic subsurface hydrology from theory to applications. Water Resour. Res. 22 9 ,

1355–1455.Geman, S., Geman, D., 1984. Stochastic relaxation, Gibbs distributions, and the Bayesian restoration of

Ž .images. IEEE Trans. Pattern Anal. Mach. Intell. PAMI 6 6 , 721–741.Ghil, M., 1989. Meteorological data assimilation for oceanographers: Part 1. Description and theoretical

framework. Dyn. Atmos. Oceans 13, 171–218.Golub, G.H., Van Loan, C.F., 1989. Matrix Computations. The Johns Hopkins Univ. Press, Baltimore.Hall, D.L., 1992. Mathematical Techniques in Multisensor Data Fusion. Artech House, Boston.Hamm, L.L., Aleman, S.E., Jones, W.F., Flach, G.P., Haselow, J.S., Huyakorn, P.S., Panday, S., Birdie, T.,

1995. FACT: Subsurface Flow and Contaminant Transport Documentation and User’s Guide, Westing-house Savannah River Company. Technical Report No. WSRC-TR-95-0223.

Ž .Hill, M.C., 1992. A Computer Program MODFLOWP for Estimating Parameters of a Transient, Three-Di-mensional, Ground-Water Flow Model Using Nonlinear Regression. U.S. Geol. Surv. Rep. 91-484.

Hoeksema, R.J., Kitinidis, P.K., 1984. An application of the geostatistical approach to the inverse problem inŽ .two-dimensional groundwater modeling. Water Resour. Res. 20 7 , 1003–1020.

Huyakorn, P.S., Panday, S., 1995. VAM3DF: Variably Saturated Analysis Model in Three-dimensions for theData Fusion System: Documentation and User’s Guide, Version 1.0. HydroGeoLogic, Virginia, 278 pp.

Ž .Jain, A.K., 1981. Advances in mathematical models for image processing. Proc. IEEE 69 5 , 502–528.Journal, A.G., 1991. Mining Geostatistics. Academic Press, New York.Kalman, R.E., 1960. A new approach to linear filtering and prediction problems. Trans. ASME, Ser. D 82,

35–45.Kightly, J.R., Forsyth, P.A., 1991. MAT1 Iterative Sparse Matrix Solver: User’s Guide.Kitanidis, P.K., Vomvoris, E.G., 1983. A geostatistical approach to the inverse problem in groundwater

Ž . Ž .modeling steady state and one-dimensional simulations. Water Resour. Res. 19 3 , 677–690.Ž .Kung, S., 1991. Systolic array designs for Kalman filtering. IEEE Trans. Signal Process. 39 1 , 171–182.

LaVenue, A.M., RamaRao, B.S., de Marsily, G., Marietta, M.G., 1995. Pilot point methodology for automatedcalibration of an ensemble of conditionally simulated transmissivity fields: 2. Application. Water Resour.

Ž .Res. 31 3 , 495–516.Lawson, C.L., Hanson, R.J., 1974. Solving Least Squares Problems. Prentice-Hall, Englewood Cliffs, NJ.Levy, L.J., Porter, D.W., 1992. Large-scale system performance, prediction with confidence from limited field

Ž .testing using parameter identification. Johns Hopkins APL Tech. Dig. 13 2 , 300–308.McLaughlin, D., 1995. Recent advances in hydrologic data assimilation. U.S. Natl. Rep. Int. Union Geod.

Geophys. 1991–1994. Rev. Geophys. 33, 977–984.McLaughlin, D., Townley, L.R., 1996. A reassessment of the groundwater inverse problem. Water Resour.

Ž .Res. 32 5 , 1131–1161.Paige, C.C., Saunders, M.A., 1982. LSQR: an algorithm for sparse linear equations and sparse least squares.

Ž .ACM Trans. Math. Software 8 1 , 43–71.Poeter, E.P., McKenna, S.A., 1994. Geostatistical Simulation and Inverse Flow Modeling to Reduce

Uncertainty Associated with Flow and Transport Predictions, Proceedings of the 1994 Ground WaterModeling Conference, Fort Collins, CO. International Ground Water Modeling Center Publication.

Porter, D.W., 1991. Quantitative Data Fusion: A DistributedrParallel Approach to Surveillance, Tracking, andNavigation using Information Filtering, Fifth Joint Service Data Fusion Symposium, Johns HopkinsUniversityrApplied Physics Laboratory, October.

Press, W.H., Teukolsky, S.A., Vetterling, W.T., Flannery, B.P., 1992. Numerical Recipes in FORTRAN: TheArt of Scientific Computing. Cambridge Univ. Press, Cambridge.

RamaRao, B.S., LaVenue, A.M., de Marsily, G., Marietta, M.G., 1995. Pilot point methodology for automatedcalibration of an ensemble of conditionally simulated transmissivity fields: 1. Theory and computational

Ž .experiments. Water Resour. Res. 31 3 , 475–493.

Page 33: Data fusion modeling for groundwater systemsmmc2.geofisica.unam.mx/cursos/geoest/Articulos... · 2015. 10. 23. · Data Fusion Modelling DFM has value added as an engineering decision

( )D.W. Porter et al.rJournal of Contaminant Hydrology 42 2000 303–335 335

Ž .Sezan, M.I., Tekalp, A.M., 1990. Survey of recent developments in digital image restoration. Opt. Eng. 29 5 ,393–404.

Sun, N.Z., Yeh, W.W.G., 1992. A stochastic inverse solution for transient groundwater flow: parameterŽ .identification and reliability analysis. Water Resour. Res. 28 12 , 3269–3280.

Vandergraft, J.S., 1985. Efficient Optimization Methods for Maximum Likelihood Parameter Estimation.Proceedings of the 24th Conference on Decision and Control, Ft. Lauderdale, FL.

Van Geer, F.C., Te Stroet, C.B.M., Yangxiao, Z., 1991. Using Kalman filtering to improve and quantify theuncertainty of numerical groundwater simulations: 1. The role of system noise and its calibration. Water

Ž .Resour. Res. 27 8 , 1987–1994.Whittle, P., 1954. On stationary processes in the plane. Biometrika 41, 434–449.Yangxiao, Z., Te Stroet, C.B.M., Van Geer, F.C., 1991. Using Kalman filtering to improve and quantify the

uncertainty of numerical groundwater simulations: 2. application to monitoring network design. WaterŽ .Resour. Res. 27 8 , 1995–2006.

Zimmerman, D.A., de Marsily, G., Gotway, C.A., Marietta, M.G., Axness, C.L., Beauheim, R.L., Bras, R.L.,Carrera, J., Dagan, G., Davies, P.B., Gallegos, D.P., Galli, A., Gomez-Hernandez, J., Grindrod, P.,´ ´Gutjahr, A.L., Kitanidis, P.K., Lavenue, A.M., McLaughlin, D., Neuman, S.P., RamaRao, B.S., Ravenne,C., Rubin, Y., 1998. A comparison of seven geostatistically based inverse approaches to estimate

Ž .transmissivities for modeling advective transport by groundwater flow. Water Resour. Res. 34 6 ,1373–1413.

Ž .Zou, S., Parr, A., 1995. Optimal estimation of two-dimensional contaminant transport. Ground Water 33 2 ,319–325.