identification of physical systems (applications to condition monitoring, fault diagnosis, soft...

22
4 Estimation of Random Parameter 4.1 Overview In this chapter, estimation of a random parameter is considered. The measurement model relating the random parameter and the measurement is assumed to be linear. The measurement is corrupted by an additive zero-mean white noise. The estimate of the unknown random parameter is obtained by minimizing mean-square error (MSE), which is the expectation of the error between the parameter and its estimate in the mean-square error sense. The estimator is termed as the minimum mean-square (MMSE) estimator. The term MMSE specifically refers to estimation in a Bayesian setting with quadratic cost function. Unlike non-random parameter estimation schemes, such as the maximum likelihood, the Bayesian scheme deals with estimation of unknown parameters which are treated as random variables. The PDF including the mean values of the unknown parameters is assumed to be known a priori. If the PDF is assumed to be Gaussian, then the mean and the covariance of the unknown parameters are assumed to known. The Bayesian approach, based directly on Bayes’ theorem, provides a prob- abilistic setting so that the a priori known PDF of the random parameters is incorporated into the estimator. One may question the requirement of the a priori knowledge of the PDF, especially the mean. Consider the problem of estimating an unknown parameter which varies randomly. The unknown random parameters may be modeled (probabilistically) by a PDF. For example if the PDF is assumed to Gaussian, the mean and the covariance may be estimated by performing a number of experiments using, say, the maximum likelihood method. Using this a priori knowledge, the estimate of the actual parameter is obtained using a Bayesian-based MMSE approach. Thus the MMSE estimator incorporates the a priori knowledge of the probabilistic model and the a posteriori information provided by the measurement. This approach of fusing the information provided by the model and the measurement forms the backbone of the Kalman filter estimation scheme. The derivation of MMSE and its properties is based on seminal works of [1, 2]. 4.2 Minimum Mean-Squares Estimator (MMSE): Scalar Case Let the measurement model be y = ax + v (4.1) Identification of Physical Systems: Applications to Condition Monitoring, Fault Diagnosis, Soft Sensor and Controller Design, First Edition. Rajamani Doraiswami, Chris Diduch and Maryhelen Stevenson. © 2014 John Wiley & Sons, Ltd. Published 2014 by John Wiley & Sons, Ltd.

Upload: maryhelen

Post on 05-Jan-2017

213 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Identification of Physical Systems (Applications to Condition Monitoring, Fault Diagnosis, Soft Sensor and Controller Design) || Estimation of Random Parameter

4Estimation of Random Parameter

4.1 OverviewIn this chapter, estimation of a random parameter is considered. The measurement model relating therandom parameter and the measurement is assumed to be linear. The measurement is corrupted byan additive zero-mean white noise. The estimate of the unknown random parameter is obtained byminimizing mean-square error (MSE), which is the expectation of the error between the parameterand its estimate in the mean-square error sense. The estimator is termed as the minimum mean-square(MMSE) estimator. The term MMSE specifically refers to estimation in a Bayesian setting with quadraticcost function. Unlike non-random parameter estimation schemes, such as the maximum likelihood, theBayesian scheme deals with estimation of unknown parameters which are treated as random variables.The PDF including the mean values of the unknown parameters is assumed to be known a priori. Ifthe PDF is assumed to be Gaussian, then the mean and the covariance of the unknown parametersare assumed to known. The Bayesian approach, based directly on Bayes’ theorem, provides a prob-abilistic setting so that the a priori known PDF of the random parameters is incorporated into theestimator.

One may question the requirement of the a priori knowledge of the PDF, especially the mean.Consider the problem of estimating an unknown parameter which varies randomly. The unknown randomparameters may be modeled (probabilistically) by a PDF. For example if the PDF is assumed to Gaussian,the mean and the covariance may be estimated by performing a number of experiments using, say, themaximum likelihood method. Using this a priori knowledge, the estimate of the actual parameter isobtained using a Bayesian-based MMSE approach. Thus the MMSE estimator incorporates the a prioriknowledge of the probabilistic model and the a posteriori information provided by the measurement.This approach of fusing the information provided by the model and the measurement forms the backboneof the Kalman filter estimation scheme.

The derivation of MMSE and its properties is based on seminal works of [1, 2].

4.2 Minimum Mean-Squares Estimator (MMSE): Scalar CaseLet the measurement model be

y = ax + v (4.1)

Identification of Physical Systems: Applications to Condition Monitoring, Fault Diagnosis, Soft Sensor andController Design, First Edition. Rajamani Doraiswami, Chris Diduch and Maryhelen Stevenson.© 2014 John Wiley & Sons, Ltd. Published 2014 by John Wiley & Sons, Ltd.

Page 2: Identification of Physical Systems (Applications to Condition Monitoring, Fault Diagnosis, Soft Sensor and Controller Design) || Estimation of Random Parameter

168 Identification of Physical Systems

where x is a zero-mean Gaussian process with variance, 𝜎2x , v is a zero-mean white Gaussian noise with

variance, 𝜎2v , a is some constant, and x, v, and y are all scalars. The stochastic processes x and v are

uncorrelated. We will obtain the expressions for the following:

∙ The Minimum Mean-Squared Error (MMSE) estimate x̂ of x.∙ The estimation error.

Consider the problem of estimating the MMSE estimate x̂ from the following minimization of themean-squared estimation error (MMSE):

minx̂

{E[(x − x̂)2]} (4.2)

The solution to this problem is obviously x̂ = x since with this choice the performance measureE[(x − x̂)2] = 0, which is the minimum value:

x̂ = x (4.3)

However, the above solution is meaningless. If we know x why should we estimate it! In other words theproblem is ill posed. Let us now restrict the MMSE estimator to be x̂ = 𝛼 where 𝛼 is a constant, whichis to be determined from the minimization problem (4.2). Substituting x̂ = 𝛼 the minimization problembecomes:

min𝛼

{E[(x − x̂)2] = E[x2] − 2𝛼E[x] + 𝛼2} (4.4)

Differentiating J with respect to 𝛼 and setting it to zero yields

x̂ = E [x] (4.5)

This estimator makes sense if the measurement y is unavailable: with no other information about theunknown x available, the best estimate is x̂ = E[x].

4.2.1 Conditional Mean: Optimal EstimatorLet us now consider the practical case where the measurement y is available so that the estimator isrestricted as a function only of the measurement y. This restriction, namely x̂ is a function only of y, maybe imposed simply by modifying the performance measure to be a conditional expectation of the meansquared estimation error given y. The modified MMSE error problem becomes:

minx̂

{E[(x − x̂(y))2|y]} (4.6)

Expanding the conditional expectation measure yields:

E[(x − x̂(y))2|y] = E[x2 + x̂2(y) − 2xx̂(y)|y] (4.7)

Simplifying yields

E[(x − x̂(y))2|y] = E[x2|y] + E[x̂2(y)|y] − 2E[xx̂(y)|y] (4.8)

Page 3: Identification of Physical Systems (Applications to Condition Monitoring, Fault Diagnosis, Soft Sensor and Controller Design) || Estimation of Random Parameter

Estimation of Random Parameter 169

Since x̂(y) and x̂2(y) are functions only of y, using E[𝜑(y)|y] = 𝜑(y) where 𝜑(y) is a function only of y,the performance measure simplifies to:

E[(x − x̂(y))2|y] = E[x2|y] + x̂2(y) − 2x̂(y)E[x|y] (4.9)

The MMSE estimator x̂(y) is obtained by differentiating J with respect to x̂(y) and setting to zero:

𝛿

𝛿x{E[x2|y] + x̂2(y) − 2x̂(y)E[x|y]} = 0 (4.10)

Differentiating and simplifying yields

E[x̂(y)] = E[E[x|y]] = E[x] (4.11)

x̂(y) = E [ x| y] (4.12)

This shows that the conditional mean E[x|y] is an optimal estimator x̂(y) of x when we restrict theestimator to be a class of all estimators which are a function of the measurement. Let us verify whetherthe MMSE estimator x̂(y) is unbiased. Taking the expectation on both sides of Eq. (4.12) and usingE[E[x|y]] = E[x] we get

E [x̂(y)] = E [E [ x| y]] = E [x] (4.13)

Since E[x̂(y)] = E[x], the MMSE estimator x̂(y) is an unbiased estimator of x. Further the minimummean-square error E[(x − x̂(y))2|y] is the conditional variance x̂(y) of x given y.

4.3 MMSE Estimator: Vector Case

Linear Measurement ModelWe will assume the measurement model relating the unknown M × 1 parameter x =[x(0) x(1) x(2) . x(N − 1)]T , and the N × 1 measurement y = [y(0) y(1) y(2) . y(N − 1)]T

to be linear given by:

y = Hx + v (4.14)

where v is N × 1 zero-mean measurement noise with covariance 𝚺v and H is a N × M matrix.

Performance MeasureLet x̂ be M × 1 minimum mean-squared error estimator of x with a resulting estimation error x − x̂. Fornotational convenience, the dependence on y in x̂(y) is suppressed and is indicated by x̂. Taking a cuefrom the previous section to ensure that the estimator is a function only of the measurement y, therebyavoiding an impractical solution, the performance measure is chosen to be a conditional mean-squaredestimation error rather than a mean-squared estimation error. The conditional mean-squared estimationerror measure is a M × M positive semi-definite matrix denoted J(y) given by

J(y) = E[(x − x̂)(x − x̂)T |y] (4.15)

An estimate x̂(y) is obtained from the minimization of J(y):

minx̂

J(y) (4.16)

Page 4: Identification of Physical Systems (Applications to Condition Monitoring, Fault Diagnosis, Soft Sensor and Controller Design) || Estimation of Random Parameter

170 Identification of Physical Systems

The definition of minimization of a positive definite matrix performance measure J(y) indicated by Eq.(4.16) is given below:

Definition A positive definite matrix J0 is the minimum value of the positive definite matrix J if thefollowing matrix inequality holds:

J0 − J ≤ 0 for all J (4.17)

where the notation J0 − J ≤ 0 implies that J0 − J is a negative semi-definite matrix.

Expanding the conditional MMSE (4.15) we get

E[(x − x̂)(x − x̂)T |y] = E[xxT |y] + E[x̂x̂T |y] − E[xx̂T |y] − E[x̂xT |y] (4.18)

Using the property of conditional expectation E[x̂|y] = x̂, namely E[x̂xT |y] = x̂E[xT |y]; E[xx̂T |y] =E[x|y]x̂T we get

E[(x − x̂)(x − x̂)T |y] = E[xxT |y] + x̂x̂T − E[x|y]x̂T − x̂E[xT |y] (4.19)

The problem of minimization of the conditional expectation Eq. (4.16) becomes

minx̂

{E[xxT |y] + x̂x̂T − E[x|y]x̂T − x̂E[xT |y]} (4.20)

There are two approaches to minimization of the above expression:

∙ Completing the square approach.∙ Vector calculus approach.

Completing the Square ApproachLet us first employ the complete square approach. The last three terms on the right-hand side is a vectorversion of a2 − 2ab where a and b respectively takes the role of x̂ and E[x|y(E[xT |y]); a2 and 2ab therole of x̂x̂T and E[x|y]x̂T + x̂E[xT |y] respectively.

Adding and subtracting b2 so as to complete the square we get, a2 − 2ab = a2 − 2ab + b2 − b2 =(a − b)2 − b2. We will use this approach to our problem.

Adding and subtracting E[x|y]E[xT |y] to the expression for J in Eq. (4.19) we get

J = E[xxT |y] + x̂x̂T − E[x|y]x̂T − x̂E[xT |y]

+E[x|y]E[xT |y] − E[x|y]E[xT |y](4.21)

Completing the square similar to ((a − b)2 − b2) on the right-hand side we get

J = E[xxT |y] + (x̂ − E[x|y])(x̂ − E[x|y])T − E[x|y]E[xT |y] (4.22)

The minimization problem (4.20) reduces to finding an estimate x̂ which minimizes the expression onthe right-hand side of Eq. (4.22). The performance measure J will be minimum if we choose x̂ so thatthe second term on the right-hand side is zero. Hence the MMSE estimate, x̂ is

x̂ = E[x|y] (4.23)

Page 5: Identification of Physical Systems (Applications to Condition Monitoring, Fault Diagnosis, Soft Sensor and Controller Design) || Estimation of Random Parameter

Estimation of Random Parameter 171

Vector Calculus ApproachConsider the minimization of the conditional covariance J given by Eq. (4.19). The optimal estimate x̂is the solution of the partial derivative of J:

𝛿

𝛿x̂{J} = 𝛿

𝛿x̂{E[xxT |y] + x̂x̂T − E[x|y]x̂T − x̂E[xT |y]} = 0 (4.24)

where a partial differentiation of a matrix using the notation𝛿

𝛿x̂{J} = 0 implies

𝛿

𝛿x̂i

{J} = 0 , i = 1, 2, 3,… , M (4.25)

Using this definition yields

𝛿

𝛿x̂i

{J} = 𝛿

𝛿x̂i

{x̂x̂T } − 𝛿

𝛿x̂i

{E[x|y]x̂T } − 𝛿

𝛿x̂i

{x̂E[xT |y]} = 0 (4.26)

Simplifying we get

𝛿

𝛿x̂{J} = 2x̂ − 2E[x|y] = 0 (4.27)

Hence the MMSE estimator is given by

x̂ = E[x|y] (4.28)

Unbiased EstimatorLet us verify whether x̂ = E[x|y] is unbiased. Taking expectation of Eq. (4.23) yields

E[x̂] = E[E[x|y]] (4.29)

Since E[E[x|y]] = E[x] we get

E[x̂] = E[x] (4.30)

4.3.1 Covariance of the Estimation ErrorSince x̂ = E[x|y] is unbiased, the performance measure J = E[(x − x̂)(x − x̂)T |y] is the conditionalcovariance of the estimation error, and its minimum value J0 is obtained by substituting x̂ = E[x|y] inthe expression for J given in Eq. (4.22). The minimum conditional covariance is:

J0 = E[xxT |y] − E[x|y]E[xT |y] (4.31)

The covariance of the estimator x̂ = E[x|y], denoted cov(x̂) = E[(x − x̂)(x − x̂)T ], is obtained by takingthe expectation of J0. The covariance cov(x̂) becomes:

cov(x̂) = E[J0] = E[E[xxT |y]] − E[E[x|y]E[xT |y]] (4.32)

Simplifying using E[E[xxT |y]] = E[xxT ] we get

cov(x̂) = E[xxT ] − E[E[x|y]E[xT |y]] (4.33)

Page 6: Identification of Physical Systems (Applications to Condition Monitoring, Fault Diagnosis, Soft Sensor and Controller Design) || Estimation of Random Parameter

172 Identification of Physical Systems

xˆ−x x

yx̂

Figure 4.1 The orthogonal relation between the estimate and the estimation error

4.3.2 Conditional Expectation and Its PropertiesThe unconditional and conditional expectations satisfy the following relation [1]:

E[xgT (y)] = E[E[x|y]gT (y)] (4.34)

For a scalar case we get

E[xg(y)] = E[E[x|y]g(y)] (4.35)

Orthogonality PrincipleThe MMSE estimator x̂ = E[x|y] of x is orthogonal to the estimation error x − E[x|y]:

E[(x − E[x|y])(E[x|y])T ] = 0 (4.36)

Proof: Expanding Eq. (4.36) yields

E[(x − E[x|y])(E[x|y])T ] = E[x(E[x|y])T ] − E[E[x|y](E[x|y])T ] (4.37)

Using Eq. (4.34) on the first term becomes

E[x(E[x|y])T ] = E[E[x|y](E[x|y])T ] (4.38)

Substituting Eq. (4.38) in Eq. (4.37) we get Eq. (4.36). For the scalar case the result Eq. (4.36) simplifiesto the following orthogonal relation:

E[(x − E[x|y])E[x|y]] = 0 (4.39)

The orthogonal relation between the MMSE estimator x̂ and the estimation error x − x̂ is shown inFigure 4.1.

4.4 Expression for Conditional MeanWe will derive an expression for conditional mean when (i) the measurement is linear and (ii) x and vare independent. Consider the minimum mean-squared estimator x̂ given by Eq. (4.28). Expressing interms of the conditional PDF yields:

x̂ = E[x|y] =

∫−∞

x f (x|y) dx (4.40)

Page 7: Identification of Physical Systems (Applications to Condition Monitoring, Fault Diagnosis, Soft Sensor and Controller Design) || Estimation of Random Parameter

Estimation of Random Parameter 173

where f (x|y) can be expressed either in terms of (i) joint PDF f (x, y) and marginal PDF fy(y), or (ii) theconditional PDF f (y|x) and the marginal PDFs fx(x) and fy(y) as:

f (x|y) =f (x, y)fy(y)

=f (y |x)fx(x)

fy(y)(4.41)

Since x and v are independent, using the linear model (4.14) the expression for the conditional PDFf (y|x) and joint PDF f (x, y) in terms of noise PDF fv(v) become

f (y|x) = fv(y − Hx)

f (x, y) = fv(y − Hx)fx(x)(4.42)

Substituting for f (y|x) in Eq. (4.41) the f (x|y) becomes

f (x|y) =fv(y − Hx)fx(x)

fy(y)(4.43)

Substituting for f (x|y) in the expression for x̂ in Eq. (4.40) we get:

x̂ = E[x|y] =

∞∫−∞

x fv(y − Hx)fx(x)dx

fy(y)(4.44)

where the marginal PDF fy(y) of y is:

fy(y) =

∫−∞

fx(x)fv(y − Hx)dx (4.45)

The determination of the MMSE estimator x̂ is computationally burdensome as it involves integrationoperations given by Eqs. (4.44) and (4.45). We will consider two special cases where it may be moreconvenient to compute the estimator:

∙ The PDF of the unknown parameter x is Gaussian.∙ All the PDFs are Gaussian.

4.4.1 MMSE Estimator: Gaussian Random VariablesThe Gaussian PDF models random variables which are clustered close to the mean as it has very thintails representing “good data” with negligible outliers. A non-Gaussian PDF, such as a Laplacian or aCauchy with thicker tails, models “bad data” with many outliers. The bad data points are due to faultyinstruments, unexpected failure of communication links, intermittent faults, and modeling errors, to namea few.

The measurement y = Hx + v is commonly used to model an unknown random parameter x clusteredclose to the mean (“good signal”) corrupted by measurement noise v which contain outliers (“bad noise”).As a result the measurement random variable y will be “bad measurement.” Hence in practice it may bereasonable to assume the PDF of x to be Gaussian and that of y to be non-Gaussian. The expression forthe MMSE estimator, namely the conditional expectation of x given y, is considerably simplified thanksto the property of a Gaussian PDF of x as the ratio between the gradient of the PDF and the PDF itselfis linear function of x. The MMSE estimator, as well as the covariance of the estimation error is only afunction of the PDF of the measurement y.

Page 8: Identification of Physical Systems (Applications to Condition Monitoring, Fault Diagnosis, Soft Sensor and Controller Design) || Estimation of Random Parameter

174 Identification of Physical Systems

4.4.2 MMSE Estimator: Unknown is Gaussian and MeasurementNon-Gaussian

4.4.2.1 Scalar Case

We will first consider the scalar version and then extend to the vector case. A scalar version for theexpression for the conditional mean given by Eq. (4.44) is:

E[x|y] =

∞∫−∞

x fv(y − ax)fx(x) dx

fy(y)(4.46)

The Gaussian PDF of x is:

fx(x) = 1

𝜎x

√2𝜋

exp{− x2

2𝜎2x

}(4.47)

Differentiating with respect to x yields:

f ′x (x) = − x2

𝜎2x

(1

𝜎x

√2𝜋

exp{− x2

2𝜎2x

})= − x

𝜎2x

fx(x) (4.48)

where f ′x (x) = ddx

fx(x). Using Eq. (4.48), substituting for x fx(x) in the integrand of the expression on the

right-hand side of Eq. (4.46) yields

E[x|y] =−𝜎2

x

∞∫−∞

fv(y − ax)f ′x (x) dx

fy(y)(4.49)

Lemma 4.1 If (i) y = hx + v, (ii) x and v are zero-mean random variables, (iii) x and v are independent,(iv) the PDFs fx(x), fy(y), and fv(v) are symmetric about the mean and absolutely continuous, and x is a

Gaussian random variable with PDF fx(x) = 1√2𝜋 𝜎2

x

exp{− x2

2𝜎2x

}, then

x̂ = E[x|y] = −𝜎2x h

(f ′y (y)

fy(y)

)(4.50)

where f ′y (y) =dfy(y)

dy

Proof: The proof is given in the Appendix.

Variance of the MMSE EstimatorThe conditional variance of the MMSE estimator x̂ given by Eq. (4.31) becomes

var(x̂|y) = E[(x − x̂)2|y] (4.51)

Page 9: Identification of Physical Systems (Applications to Condition Monitoring, Fault Diagnosis, Soft Sensor and Controller Design) || Estimation of Random Parameter

Estimation of Random Parameter 175

Expanding using Eq. (4.31)

var(x̂|y) = E[x2|y] − E[x̂2|y] (4.52)

Since x̂ is a function only of y, E[x̂2|y] = x̂2. Hence

var(x̂|y) = E[x2|y] − 𝜎4x h2

(f ′y (y)

fy(y)

)2

(4.53)

The variance of the MMSE estimator is obtained by taking expectation of the conditional variance (4.53)and is given by

var(x̂) = 𝜎2x − 𝜎4

x h2E⎡⎢⎢⎣(

f ′y (y)

fy(y)

)2⎤⎥⎥⎦ (4.54)

We will express the var(x̂) in terms of the Fisher information IF:

IF = E⎡⎢⎢⎣(

𝜕 ln(fy(y; x)

)𝜕x

)2⎤⎥⎥⎦ (4.55)

Using the linear model y = hx + v we get

IF = h2E⎡⎢⎢⎣(

f ′y (y)

fy(y)

)2⎤⎥⎥⎦ (4.56)

Using the expression for IF , var(x̂) can be expressed in terms of IF

var(x̂) = 𝜎2x − 𝜎4

x IF (4.57)

4.4.2.2 Vector Case

We will now consider the vector case

Lemma 4.2 If (i) y = Hx + v, (ii) x and v are zero-mean random variables, (iii) x and v are independent,(iv) the PDFs fx(x), fy(y), and fv(v) are symmetric about the mean and absolutely continuous, and x is a

Gaussian random variable with PDF fx(x) = 1√(2𝜋)N|𝚺x| exp

{−1

2x𝚺−1

x xT}

, then

E[x|y] = −ΣxHT

(f ′y (y)

fy(y)

)(4.58)

where f ′y (y) =df y(y)

dy=

[dfy(y)

dy(0)

dfy(y)

dy(1)

dfy(y)

dy(2).

dfy(y)

dy(N − 1)

]T

Proof: The proof is given in the Appendix.

Page 10: Identification of Physical Systems (Applications to Condition Monitoring, Fault Diagnosis, Soft Sensor and Controller Design) || Estimation of Random Parameter

176 Identification of Physical Systems

Covariance of the MMSE EstimatorUsing the expression for the covariance of the MMSE estimator (4.33) we get

cov(x̂) = 𝚺x − 𝚺xHT E

⎡⎢⎢⎣(

f ′y (y)

fy(y)

)(f ′y (y)

fy(y)

)T⎤⎥⎥⎦ H𝚺x (4.59)

We will express the cov(x̂) in terms of the Fisher information matrix IF given by

IF = E⎡⎢⎢⎣(

𝜕 ln(fy(y; x)

)𝜕x

)(𝜕 ln

(fy(y; x)

)𝜕x

)T⎤⎥⎥⎦ (4.60)

Using the linear model y = Hx + v we get

IF = HT E⎡⎢⎢⎣(

f ′y (y)

fy(y)

)(f ′y (y)

fy(y)

)T⎤⎥⎥⎦ H (4.61)

Expressing cov(x̂) in terms of IF we get

cov(x̂) = 𝚺x − 𝚺xI−1F 𝚺x (4.62)

Remarks

1. The MMSE estimator x̂ is in general a nonlinear function of the measurement y.2. The estimator x̂ is a function of the PDF of the measurement fy(y) and the covariance of the random

parameter 𝚺x.3. The covariance of the MMSE estimator cov(x̂) is a function of the PDF of the measurement fy(y) and

the covariance of the random parameter 𝚺x.

4. The ratio of the gradient of the PDF to the PDF,f ′y (y)

fy(y), plays a crucial role in x̂ and its covariance

cov(x̂).5. The covariance of the MMSE estimator is a function of the Fisher information.

4.4.3 The MMSE Estimator for Gaussian PDFIn most engineering applications, the PDFs of random variables are assumed to be Gaussian and themodel is assumed to be linear.

Most physical systems are nonlinear. However, a linearized model about an operating point is employedsuccessfully in many applications, such as mechatronic, process control, power, robotic, and aero-spacesystems. A linearized model captures nonlinear system behavior for small variations about an operatingpoint.

A random process is a signal which is corrupted by an unwanted random waveform termed noise.The noise is generated in all electronic circuits and devices as a result of random fluctuations in currentor voltage caused by the random movement of electrons. This noise is termed thermal noise or Johnsonnoise. The noise corrupting the signal at the macroscopic level is caused by the superposition of largenumbers of thermal noise generated at the atomic level. In view of Central Limit Theorem, it may bejustified to assume the noise corrupting the signal is Gaussian.

Page 11: Identification of Physical Systems (Applications to Condition Monitoring, Fault Diagnosis, Soft Sensor and Controller Design) || Estimation of Random Parameter

Estimation of Random Parameter 177

Thanks to advances in computers, microprocessors, digital signal processors, and dedicated processors,it is possible to store and process large data. In applications such as estimation, fault detection andisolation, and pattern classification, the sum of a large number of independent and identically distributedmeasurements are employed. The sum of i.i.d. random variables is Gaussian in view of Central LimitTheorem.

The other important reason for assuming a linear model and a Gaussian PDF is that solutions to manyengineering problems in analysis, estimation, control, detection, and isolation become mathematicallytractable. A closed form solution may be obtained, and in many cases the solution takes a linear form.For example, a controller, a filter, and an estimator are linear, while a discriminant function used indetection, isolation, and pattern classification is linear.

In this section we will give expressions of joint and conditional PDFs, expectation, conditionalexpectation, and conditional mean.

4.4.3.1 The Joint and Conditional Gaussian PDF

Let x = [x(0) x(1) x(2) . x(N − 1)]T and y = [ y(0) y(1) y(2) . y(M − 1) ]T be two Nx1 andMx1 random variables. Let 𝝁x = E[x] be Nx1 mean vector and 𝚺x = E[(x − 𝝁x)(x − 𝝁x)T ] be NxNcovariance of x; 𝝁y = E[y] is Mx1 mean and 𝚺y = E[(y − 𝝁y)(y − 𝝁y)

T ] is MxM covariance of y; 𝚺xy =E[(x − 𝝁x)(y − 𝝁y)

T ] is NxM covariance and 𝚺yx = 𝚺Txy = E[(y − 𝝁x)(x − 𝝁x)

T ] is MxN cross-covariance.

Let z =[

xy

]be a (N + M)x1 joint random variable formed of x and y. The joint PDF of z denoted

fxy(x, y) is:

fxy (x, y) = 1√(2𝜋)N ||𝚺z

|| exp{−1

2

(z − 𝝁z

)T 𝚺−1z

(z − 𝝁z

)},−∞ ≤ z ≤ ∞ (4.63)

where

𝝁z =[𝝁x

𝝁y

]is (N + M) x1; 𝚺z = E

[(z − 𝜇z)(z − 𝝁z)

T]

is (N + M) x (N + M) ; 𝚺z =[𝚺x 𝚺xy

𝚺yx 𝚺y

];

𝚺−1z =

[(𝚺x − 𝚺xy𝚺−1

y 𝚺yx)−1 −𝚺−1

x 𝚺xy(𝚺y − 𝚺yx𝚺−1x 𝚺xy)

−1

−(𝚺y − 𝚺yx𝚺−1x 𝚺xy)

−1𝚺yx𝚺−1x (𝚺y − 𝚺yx𝚺−1

x 𝚺xy)−1

]|𝚺z| = |𝚺x||𝚺y − 𝚺yx𝚺−1

x 𝚺xy|The conditional PDF f (x|y), which plays an important role in estimation of random parameter, may beexpressed in terms of the joint and the marginal PDFs as:

f (x|y) =fxy(x, y)

fy(y)(4.64)

The expression for f (x|y) becomes

f (x|y) = 1√(2𝜋)N |||𝚺x|y||| exp

{−1

2(x − 𝝁x|y)T𝚺−1

x|y(x − 𝝁x|y)}

,−∞ ≤ x ≤ ∞ (4.65)

The conditional PDF f (x|y) is also multivariate Gaussian with mean 𝝁x|y and conditional covariance 𝚺x|y.

Page 12: Identification of Physical Systems (Applications to Condition Monitoring, Fault Diagnosis, Soft Sensor and Controller Design) || Estimation of Random Parameter

178 Identification of Physical Systems

4.4.3.2 Conditional Mean and Conditional Variance

The conditional mean is affine in y (linear with an additive constant term) and its closed form expressionis given by

x̂ = E[x|y] = 𝝁x + 𝚺xy𝚺−1y (y − 𝝁y) (4.66)

The conditional covariance cov(x̂|y) of the estimation error is

cov(x̂|y) = E[(x − E[x|y])(x − E[x|y])T ] = 𝚺x − 𝚺xy𝚺−1y 𝚺yx (4.67)

Remarks In the case when x y are Gaussian random variables, the following holds:

1. Conditional mean E[x|y] is a function of y.2. Conditional covariance cov(x̂|y) is not a function of y, that is cov(x̂) = cov(x̂|y).3. Conditional PDF: bivariate Gaussian

E[x|y] = 𝜇x|y = 𝜇x +cov (x, y)

𝜎2y

(y − 𝜇y

)= 𝜇x + 𝜌

𝜎x

𝜎y

(y − 𝜇y

)(4.68)

var (x) = var (x|y) = 𝜎2x

(1 − 𝜌2

)= 𝜎2

x −cov2 (x, y)

𝜎2y

(4.69)

4.4.4 Illustrative Examples

Example 4.1 Scalar Gaussian random variablesWe will assume a linear model y = ax + v where x and v are zero-mean Gaussian random variables withvariances 𝜎2

x and 𝜎2v with PDFs

fx(x) = 1

𝜎x

√2𝜋

exp{− x2

𝜎2x

}(4.70)

fv(v) = 1

𝜎v

√2𝜋

exp{− v2

𝜎2v

}(4.71)

As the random variables are zero mean and Gaussian, the MMSE estimate x̂ is given by

x̂ = E[x|y] = 𝛼y where 𝛼 =E[xy]E[y2]

(4.72)

Substituting y = ax + v in the expression for 𝛼, and noting that x and v are uncorrelated gives;

x̂ =E[xy]E[y2]

y = E[x(ax + v)]E[(ax + v)2]

y = aE[x2]a2E[x2] + E[v2]

y =a𝜎2

x

a2𝜎2x + 𝜎2

v

y (4.73)

Thus the estimate x̂ becomes

x̂ =

(a𝜎2

x

a2𝜎2x + 𝜎2

v

)y =

( aSNRa2SNR + 1

)y where SNR =

𝜎2x

𝜎2v

(4.74)

Page 13: Identification of Physical Systems (Applications to Condition Monitoring, Fault Diagnosis, Soft Sensor and Controller Design) || Estimation of Random Parameter

Estimation of Random Parameter 179

The expression for the variance of the estimation error is

E[(x − x̂)2] = E[(x − 𝛼y)2

]where 𝛼 =

a𝜎2x

a2𝜎2x + 𝜎2

v

(4.75)

Substituting for y we get

E[(x − x̂)2] = E[(x − 𝛼 (ax + v))2

]= 𝜎2

x (1 − a𝛼)2 + 𝛼2𝜎2v (4.76)

Substituting for 𝛼 yields

E[(x − x̂)2] = (1 − a𝛼)2𝜎2x + 𝛼2𝜎2

v where 𝛼 =a𝜎2

x

a2𝜎2x + 𝜎2

v

(4.77)

Simplifying we get:

E[(x − x̂)2] =𝜎2

x 𝜎2v

a2𝜎2x + 𝜎2

v

(4.78)

Numerical example

a = 1; 𝜎2x = 0.2; 𝜎2

v = 1

Performance of the MMSE estimator is shown in Figure 4.2. Subfigures (a), (b), (c), and (d) showrespectively the measurements y, the random parameter x, the random parameter and its MMSE estimate

0 50 100-2

0

2

4

6A: Measurement

Measure

ments

0 50 1000.5

1

1.5

B: Parameter

Para

mete

r

Experiments

0 50 1000

0.5

1

1.5

2C: MMSE

Estim

ate

s

MMSE

Parameter

0 50 100-0.5

0

0.5D: MMSE error

Err

or

Experiments

Figure 4.2 Performance of the MMSE estimator

Page 14: Identification of Physical Systems (Applications to Condition Monitoring, Fault Diagnosis, Soft Sensor and Controller Design) || Estimation of Random Parameter

180 Identification of Physical Systems

x̂, and the parameter estimation error x − x̂. The theoretical mean and the standard deviations of theestimation error are shown. The estimation error variance is E[(x − x̂)2] = 0.1667 and the standarddeviation is 0.4082.

Example 4.2 ML estimate, scalar caseObtain the maximum likelihood estimate of an unknown deterministic parameter 𝜃 given a measurementmodel

y = A𝜃 + v

where y is a 2 × 1 vector of measurement 𝜃 is a scalar, v is a 2 × 1 zero-mean white Gaussian noise withcovariance,

𝚺v =[

1 00 2

]and A is a 2 × 1 vector given by

A =[

12

]Solution

>The log-likelihood function l(𝜽|y) = ln fy(y;𝜽) is given by

l(𝜽|y) = 𝛼 −(y − A𝜃)T𝚺−1

v (y − A𝜃)

2

where 𝛼 is a constant. The optimal estimate is computed by maximizing the log-likelihood function:differentiate with respect to 𝜃 and set it to zero. The optimal estimate becomes

�̂� =(AT𝚺−1

v A)−1

AT𝚺−1v y

Substituting for A and 𝚺v we get

�̂� = 13

[1 1

]y

The Fisher information matrix is (AT𝚺−1v A) = 3

Example 4.3 ML and MMSE, vector caseCompute an estimate of the unknown 2 × 1 parameter vector 𝜽 using

∙ Maximum likelihood approach.∙ Minimum mean squared Approach.

The measurement model is:

y = A𝜽 + v (4.79)

Page 15: Identification of Physical Systems (Applications to Condition Monitoring, Fault Diagnosis, Soft Sensor and Controller Design) || Estimation of Random Parameter

Estimation of Random Parameter 181

where y is a 2 × 1 vector of measurement, 𝜃 is 2 × 1 random parameter with mean 𝜇𝜃 = 1 with variance

𝜎𝜃 = 0.2, v is a 2 × 1 zero-mean white Gaussian noise with covariance 𝜎v =[

1 00 2

], and A is a 2 × 2

matrix given by A =[

1 22 1

]. We will assume that 𝜃 is uncorrelated with measurement noise v.

a) Maximum likelihood estimateThe log-likelihood function l(𝜽|y) is given by

l(𝜽|y) = 𝛼 −(y − A𝜽)T𝚺−1

v (y − A𝜽)

2(4.80)

where 𝛼 is a constant. The optimal estimate is computed by maximizing the log-likelihood function:differentiate with respect to 𝜽 and set it to zero. The optimal estimate �̂�ML becomes:

�̂�ML =(AT𝚺−1

v A)−1

AT𝚺−1v y (4.81)

The covariance of the estimation error is:

cov(𝜽) = I−1F (4.82)

Where IF = HT𝚺−1v H.

Substituting for A and Σ we get

�̂�ML = 13

[−1 2

2 −1

]y (4.83)

b) Minimum Mean Square Error estimatorUsing Eq. (4.66) the MMSE becomes

�̂� = E[𝜽|y] = 𝜇𝜃 + 𝚺𝜃y𝚺−1y (y − 𝝁y) (4.84)

where 𝝁y = A𝝁𝜃 . Using the measurement model (4.79) the cross-covariance matrix 𝚺𝜃y becomes

𝚺𝜃y = E[(𝜽 − 𝝁𝜃 )(y − 𝝁y)T ] = E[(𝜽 − 𝝁𝜃 )(A𝜽 + v − 𝝁y)

T ] (4.85)

Since 𝝁y = A𝝁𝜃 , 𝚺𝜃 = E[(𝜽 − 𝝁𝜃 )(𝜽 − 𝝁𝜃 )T ] and E[𝜽vT ] = 0 we get:

𝚺𝜃y = E[(𝜽 − 𝝁𝜃 )(A(𝜽 − 𝝁𝜃 ) + v)T ] = E[(𝜽 − 𝝁𝜃 )(𝜽 − 𝝁𝜃 )T ]AT = 𝚺𝜃 AT (4.86)

Consider the covariance 𝚺y = E[(y − 𝝁y)(y − 𝝁y)T ]. Using E[𝜽vT ] = 0 and 𝝁y = A𝝁𝜃 we get:

𝚺y = E[(A𝜽 + v − 𝝁y)(A𝜽 + v − 𝝁y)T ] = AE[(𝜽 − 𝝁𝜃 )(𝜽 − 𝝁𝜃 )T ]AT + E[vvT ] (4.87)

Using the definition 𝚺𝜃 = E[(𝜽 − 𝝁𝜃 )(𝜽 − 𝝁𝜃 )T ] and 𝚺v = E[vvT ] we get:

𝚺y = A𝚺𝜃 AT + 𝚺v (4.88)

Using Eqs. (4.85) and (4.86), the MMSE estimate Eq. (4.84) becomes:

�̂� = E[𝜽|y] = 𝝁𝜃 + 𝚺𝜃 AT (AΣ𝜃 AT + 𝚺v)−1(y − A𝝁𝜃 ) (4.89)

Page 16: Identification of Physical Systems (Applications to Condition Monitoring, Fault Diagnosis, Soft Sensor and Controller Design) || Estimation of Random Parameter

182 Identification of Physical Systems

0 50 100-5

0

5

10A: Measurement

Measure

ments

0 50 100

0.6

0.8

1

1.2

1.4

B: Parameter

Para

mete

r

Experiments

0 50 100

0

1

2

C: MMSE

Estim

ate

s

0 50 100

0

1

2

D: MLE

Estim

ate

s0 50 100

-2

-1

0

1

2E: MMSE error

Err

or

Experiments

0 50 100-2

-1

0

1

2F: ML error

Err

or

Experiments

Figure 4.3 Minimum mean square error and maximum likelihood estimators.

Using Eqs. (4.86) and (4.88), the variance of the estimate becomes:

cov(�̂�) = 𝚺𝜃 − 𝚺𝜃 AT (A𝚺𝜃 AT + 𝚺v)−1A𝚺𝜃 (4.90)

Results of evaluation:Performance of MMSE and ML estimators are shown in Figure 4.3. Subfigures (a), (b), (c), (d), (e),and (f) show respectively the measurements y(1) and y(2), the random parameter 𝜃, parameter 𝜃 andits MMSE estimate �̂�, parameter 𝜃 and its ML estimate �̂�ML, the parameter estimation error 𝜃 − �̂� withMMSE scheme, and the parameter estimation error 𝜃 − �̂�ML with ML scheme. The theoretical mean andthe standard deviations of the estimation error are shown. The standard deviations of the MMSE estimateand that of the ML estimate are given in Table 4.1. The standard deviations of the MMSE estimate andthe ML estimate were computed using Eqs. (4.90) and (4.82).

Comments The MMSE and the MLE schemes are designed to handle random and non-randomunknown parameters respectively.

The performance of the MMSE scheme is superior to that of the MLE approach when the unknownparameter is random.

Using ML estimation scheme to estimate a random parameter (assuming it to be a non-randomparameter), gives a poor estimation error performance.

Table 4.1 Standard deviation

MMSE scheme MLE scheme

0.3536 0.5774

Page 17: Identification of Physical Systems (Applications to Condition Monitoring, Fault Diagnosis, Soft Sensor and Controller Design) || Estimation of Random Parameter

Estimation of Random Parameter 183

The MMSE gives superior performance as it fuses the information provided by the model and themeasurement data. The MMSE scheme requires, however, a priori knowledge of the mean value of theparameter to be estimated.

The performance MMSE estimator for the vector case, Example 4.3, is superior to that of the scalarcase, Example 4.1. In the vector case the scalar random parameter is estimated using two measurementswhile a single measurement is employed in the scalar case. The standard deviations of the estimators forthe vector and the scalar cases were respectively 0.3536 and 0.4082.

Example 4.4 True or falseState whether the following statements are true or false:

1. The maximum likelihood estimate will always be a linear function of the measurement independentof the PDF of the measurement noise.

2. The maximum likelihood estimate will be a linear function of the measurement if and only if theprobability density function of measurement noise is Gaussian.

3. The minimum mean-squared error estimate of the unknown random parameter is always a linearfunction of the measurement independent of the probability density function of the measurementnoise.

4. The minimum mean-squared error estimate of the unknown random parameter will be a linear functionof the measurement if the probability density functions and the random parameter and measurementare both Gaussian.

Solution

1. False.2. True.3. False.4. True.

4.5 SummaryThe measurement model relating the random parameter and the measurement is assumed to be MinimumMean Squares Estimator (MMSE):

Linear Measurement Model

y = Hx + v

Performance Measure

J(y) = E[(x − x̂)(x − x̂)T |y]

An estimate x̂(y) is obtained from the minimization of J(y):

minx̂

J(y)

x̂ = E[x|y]

Covariance of the Estimation Error

cov(x̂) = E[xxT ] − E[E[x|y]E[xT |y]]

Page 18: Identification of Physical Systems (Applications to Condition Monitoring, Fault Diagnosis, Soft Sensor and Controller Design) || Estimation of Random Parameter

184 Identification of Physical Systems

Conditional Expectation and its Properties

E[xgT (y)] = E[E[x|y]gT (y)]

Orthogonality PrincipleThe MMSE estimator x̂ = E[x|y] of x is orthogonal to the estimation error x − E[x|y]:

E[(x − E[x|y])(E[x|y])T ] = 0

Expression for Conditional Mean

x̂ = E[x|y] =

∞∫−∞

x fv(y − Hx)fx(x)dx

fy(y)

fy(y) =

∫−∞

fx(x)fv(y − Hx) dx

MMSE Estimator: Gaussian Random VariablesIf (i) y = Hx + v, (ii) x and v are zero-mean random variables, (iii) x and v are independent, (iv) the PDFsfx(x), fy(y), and fv(v) are symmetric about the mean and absolutely continuous and (v) x is

fx(x) = 1√(2𝜋)N|Σx| exp

{−1

2xΣ−1

x xT}

, then

E[x|y] = −𝚺xHT

(f ′y (y)

fy(y)

)

Covariance of the MMSE Estimator

cov(x̂) = 𝚺x − 𝚺xI−1F 𝚺x

Conditional Mean and Conditional Variance

x̂ = E [x|y] = 𝝁x + 𝚺xy𝚺−1y

(y − 𝝁y

)cov(x̂|y) = E[(x − E[x|y])(x − E[x|y])T ] = 𝚺x − 𝚺xy𝚺−1

y 𝚺yx

4.6 Appendix: Non-Gaussian Measurement PDFThe random parameter x is Gaussian while the measurement y is non-Gaussian. We will assume that allthe PDFs are symmetrical, and absolutely continuous. We will assume that all random variables are zeromean: 𝝁x = 0; 𝝁y = 0.

4.6.1 Expression for Conditional ExpectationThe conditional expectation is:

x̂ = E[x|y] =

∫−∞

x f (x|y) dx (4.91)

Page 19: Identification of Physical Systems (Applications to Condition Monitoring, Fault Diagnosis, Soft Sensor and Controller Design) || Estimation of Random Parameter

Estimation of Random Parameter 185

where f (x|y) can be expressed either in terms of (i) joint PDF f (x, y) and marginal PDF fy(y), or (ii) theconditional PDF f (y|x) and the marginal PDFs fx(x) and fy(y) as:

f (x|y) =f (x, y)fy(y)

=f (y|x) fx(x)

fy(y)(4.92)

where x and y are related by the linear measurement model is y = Hx + v. Since x and v are independent,using the linear model, the expression for conditional PDF f (y|x) and joint PDF f (x, y) in terms of noisePDF fv(v) become

f (y|x) = fv(y − Hx) (4.93)

f (x, y) = fv(y − Hx)fx(x) (4.94)

Substituting for f (y|x), f (x|y) becomes

f (x|y) =fv(y − Hx)fx(x)

fy(y)(4.95)

Substituting for f (x|y) in the expression for x̂ we get:

E[x|y] =

∞∫−∞

x fv(y − Hx)fx(x)dx

fy(y)(4.96)

where the marginal PDF fy(y) of y is:

fy(y) =

∫−∞

fx(x)fv (y − Hx)dx (4.97)

4.6.2 Conditional Expectation for Gaussian x and Non-Gaussian yWe will derive the expression for the conditional expectation for the scalar x and y first and then generalizeto the vector case Mx1 x and Nx1y.

Scalar x and y

Lemma 4.1 If (i) y = hx + v, (ii) x and v are zero-mean random variables, (iii) x and v are independent,(iv) the PDFs fx(x), fy(y), and fv(v) are symmetric about the mean and absolutely continuous, and x is a

Gaussian random variable with PDF fx(x) = 1√2𝜋 𝜎2

x

exp{− x2

2𝜎2x

}, then

E[x|y] = −𝜎2x h

f ′y (y)

fy(y)(4.98)

where f ′y (y) =dfy(y)

dy

Proof: The conditional expectation for the scalar case is

E[x|y] = 1fy(y)

∫−∞

x fv (y − hx) fx (x) dx (4.99)

Page 20: Identification of Physical Systems (Applications to Condition Monitoring, Fault Diagnosis, Soft Sensor and Controller Design) || Estimation of Random Parameter

186 Identification of Physical Systems

Differentiating the Gaussian PDF of the zero-mean random variable x, we get

dfx (x)

dx= d

dx

(1√

2𝜋 𝜎2x

exp{− x2

2𝜎2x

})= − x

𝜎2x

fx (x) (4.100)

Hence

xfx(x) = −𝜎2x f ′x (x) (4.101)

where f ′x (x) =dfx(x)

dx. Substituting xfx(x) = −𝜎2

x f ′x (x) in the integrand on the right-hand side of the expres-

sion for conditional expectation (4.99) we get

E[x|y] = −𝜎2

x

fy(y)

∫−∞

fv (y − hx) f ′x (x) dx (4.102)

The partial differentiation of the joint density function f (x, y) = fv(y − Hx)fx(x) gives

𝛿f (x, y)𝛿x

= fv (y − hx) f ′x (x) +𝛿fv (y − hx)

𝛿xfx (x) (4.103)

Since𝛿fv(y − hx)

𝛿x= −h

𝛿fv(y − hx)

𝛿ywe get

𝛿f (x, y)𝛿x

= fv (y − hx) f ′x (x) − h𝛿fv (y − hx)

𝛿yfx (x) (4.104)

Substituting for fv(y − hx)f ′x (x) in the expression for conditional expectation Eq. (4.102) becomes

E[x|y] = −𝜎2

x

fy(y)

∫−∞

(𝛿f (x, y)

𝛿x

)dx −

𝜎2x h

fy(y)

∫−∞

(𝛿fv (y − hx)

𝛿yfx (x)

)dx (4.105)

Since f (x, y) is even and symmetrical about the origin x = 0 and y = 0,𝛿f (x, y)

𝛿xis symmetric and odd

about the origin. Hence the first term on the right vanishes and we get

E[x|y] = −𝜎2

x h

fy(y)

∫−∞

(𝛿fv (y − hx)

𝛿yfx (x)

)dx (4.106)

Since fy(y) =∞∫

−∞fx(x)fv(y − Hx dx,

dfy(y)

dy=

∫−∞

𝛿fv(y − Hx)

𝛿yfx(x)dx, we get

E[x|y] = −𝜎2x h

f ′y (y)

fy(y)(4.107)

where f ′(y) =dfy(y)

dy.

Page 21: Identification of Physical Systems (Applications to Condition Monitoring, Fault Diagnosis, Soft Sensor and Controller Design) || Estimation of Random Parameter

Estimation of Random Parameter 187

Lemma 4.2 If (i) y = Hx + v, (ii) x and v are zero-mean random variables, (iii) x and v are independent,(iv) the PDFs fx(x), fy(y), and fv(v) are symmetric about the mean and absolutely continuous, and x is a

Gaussian random variable with PDF fx(x) = 1√(2𝜋)N|𝚺x| exp

{−1

2x𝚺−1

x xT}

, then

E[x|y] = −ΣxHT

f ′y (y)

fy(y)(4.108)

where f ′y (y) =dfy(y)

dy=

[dfy(y)

dy(0)

dfy(y)

dy(1)

dfy(y)

dy(2)⋅

dfy(y)

dy(N − 1)

]T

Proof: The conditional expectation is

E[x|y] = 1fy(y)

∫−∞

x fv (y − Hx) fx(x)dx (4.109)

Differentiating the Gaussian PDF of the zero-mean random variable x, we get

𝛿fx(x)

𝛿x= 𝛿

𝛿x

⎛⎜⎜⎜⎝1√

(2𝜋)N ||𝚺x|| exp

{−1

2x𝚺−1

x xT}⎞⎟⎟⎟⎠ = −𝚺−1

x xfx(x) (4.110)

Hence

xfx(x) = −𝚺x

𝛿fx(x)

𝛿x(4.111)

where𝛿fx(x)

𝛿x=

[𝛿fx(x)

𝛿x(0)

𝛿fx(x)

𝛿x(1)

𝛿fx(x)

𝛿x(2)⋅

𝛿fx(x)

𝛿x(N − 1)

]T

.

Using Eq. (4.111), the integrand on the right-hand side of the expression for conditional expectationEq. (4.109) becomes

E[x|y] = −𝚺x1

fy(y)

∫−∞

fv (y − Hx)𝛿fx(x)

𝛿xdx (4.112)

From the partial differentiation of the joint density function f (x, y) = fv(y − Hx)fx(x) we get

𝛿f (x, y)𝛿x

= fv(y − Hx)𝛿fx(x)

𝛿x+

𝛿fv(y − Hx)

𝛿xfx(x) (4.113)

Since𝛿fv(y − Hx)

𝛿x= −HT 𝛿fv(y − Hx)

𝛿y, and rearranging we get

fv(y − Hx)𝛿fx(x)

𝛿x=

𝛿f (x, y)𝛿x

+ HT 𝛿fv (y − Hx)

𝛿yfx(x) (4.114)

Page 22: Identification of Physical Systems (Applications to Condition Monitoring, Fault Diagnosis, Soft Sensor and Controller Design) || Estimation of Random Parameter

188 Identification of Physical Systems

The expression for conditional expectation Eq. (4.112) becomes

E[x|y] = −𝚺x

fy(y)

∫−∞

𝛿f (x, y)𝛿x

dx −𝚺xHT

fy(y)

∫−∞

𝛿fv (y − Hx)

𝛿yfx(x) dx (4.115)

Since the joint multivariate PDF f (x, y) is symmetric and even about the origin, its partial derivative𝛿f (x, y)

𝛿xis symmetric and an odd function of x about the origin. Hence the first term on the right vanishes

and we get

E[x|y] = −𝚺xH

T

fy(y)𝛿

𝛿y

∫−∞

fv (y − Hx) fx(x)dx (4.116)

Using the expression for fy(y) using Eq. (4.97) we get

E[x|y] = −𝚺xHT 1

fy(y)

𝛿fy(y)

𝛿y(4.117)

References

[1] Mendel, J. (1995) Lessons in Estimation Theory in Signal Processing, Communications and Control, PrenticeHall, New Jersey.

[2] Olofsson, P. (2005) Probability, Statistics and Stochastic Process, John Wiley and Sons, New Jersey.

Further Readings

Haykin, S. (2001) Adaptive Filter Theory, Prentice Hall, New Jersey.Kay, S.M. (1993) Fundamentals of Signal Processing: Estimation Theory, Prentice Hall, New Jersey.Mix, D.F. (1995) Random Signal Processing, Prentice Hall, New Jersey.