interpositional transfer function for 3d-sound generation*diniz/papers/ri53.pdf · sound generation...

16
Interpositional Transfer Function for 3D-Sound Generation* F. P. FREELAND, L. W. P. BISCAINHO, AES Member, AND P. S. R. DINIZ Universidade Federal do Rio de Janeiro, LPS–PEE/COPPE & DEL/POLI, 21945-970, Rio de Janeiro, RJ, Brazil The interpolation of head-related transfer functions (HRTFs) for 3D-sound generation through headphones is addressed. HRTFs, which represent the paths between sound sources and the ears, are usually measured for a finite set of source locations around the listener. Any other virtual position requires interpolation procedures on these measurements. In this work a definition of the interpositional transfer function (IPTF) and an IPTF-based method for HRTF interpolation, recently proposed by the authors, are reviewed in detail. The formulation of the interpolation weights is generalized to cope with azimuth measurement steps which change with elevation. It is shown how to obtain a set of reduced-order IPTFs using tools such as balanced model reduction and spectral smoothing. Detailed comparisons between the accuracy and the efficiency of the IPTF-based and the bilinear interpolation methods (the latter using the HRTFs directly) are provided. A practical set of IPTFs is built and applied to a computationally efficient interpolation scheme, yielding results perceptually similar to those attainable by the bilinear method. Therefore the proposed method is promising for real-time generation of spatial sound. 0 INTRODUCTION The interest in so-called 3D-sound techniques [1] has increased recently. Besides cinema and home theater, they can be found in applications ranging from a simple video game to sophisticated virtual reality simulators. Also, there is an increasing level of interaction between users and the virtual environment in which they are immersed. Contrasting with cases when the location and/or move- ment of sound sources is predefined (as in a moving pic- ture exhibition), which allow off-line digital signal pro- cessing, interactive systems require real-time processing [2] and may impose severe restrictions on the computa- tional complexity. The present work addresses the problem of spatial sound generation employing reduced computational com- plexity when the 3D sound for a single virtual source is supplied binaurally through headphones. In this case the direction of arrival of the sound can be controlled by fil- tering the original monaural signal through a proper set of previously measured head-related transfer functions (HRTFs) [3], [4]. These functions model the paths from the virtual sound source to the listener’s ears. In practical systems only a finite number of HRTFs is available. They are usually obtained from their corre- sponding head-related impulse responses (HRIRs) mea- sured under controlled conditions on a spherical surface of constant radius only for a predefined set of elevation angles and azimuth angles with respect to the ears of a target head at its center. In order to place the sound source at an arbitrary position on this surface, it is neces- sary to estimate the HRTF associated with the desired source location from the set of measured ones via an in- terpolation scheme. In particular, a realistic rendering of moving sound requires a large set of HRTFs so that no noticeable discontinuity occurs along its path. As a con- sequence, in real-time systems the computational cost to perform HRTF interpolation must be kept to a minimum without degrading the final perceptual result. In this sense the tradeoff between accuracy and system complexity is critical when designing 3D audio systems. The target here is to render faithful location of a virtual sound source moving rapidly under user control. One could think of applications such as virtual reality simulators, video games, or even films in which the perspective can be controlled by the user. When dealing with a single source, 1 a straightforward way to perform HRTF interpolation is the bilinear ap- *Manuscript received 2003 October 8; revised 2004 June 29. This work was presented in part at the AES 22nd International Conference in Espoo, Finland, 2002 June 15–17. 1 Some methods, such as [5], [6], provide alternatives to the HRTF-based model. They split the overall model into two parts, one related to the spectral characteristics, the other to the spatial ones. This interpretation can be more efficient than the direct use of HRTFs when dealing with a multisource case. PAPERS J. Audio Eng. Soc., Vol. 52, No. 9, 2004 September 915

Upload: others

Post on 19-Aug-2020

5 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Interpositional Transfer Function for 3D-Sound Generation*diniz/papers/ri53.pdf · sound generation employing reduced computational com-plexity when the 3D sound for a single virtual

Interpositional Transfer Function for3D-Sound Generation*

F. P. FREELAND, L. W. P. BISCAINHO, AES Member, AND P. S. R. DINIZ

Universidade Federal do Rio de Janeiro, LPS–PEE/COPPE & DEL/POLI, 21945-970, Rio de Janeiro, RJ, Brazil

The interpolation of head-related transfer functions (HRTFs) for 3D-sound generationthrough headphones is addressed. HRTFs, which represent the paths between sound sourcesand the ears, are usually measured for a finite set of source locations around the listener. Anyother virtual position requires interpolation procedures on these measurements. In this worka definition of the interpositional transfer function (IPTF) and an IPTF-based method forHRTF interpolation, recently proposed by the authors, are reviewed in detail. The formulationof the interpolation weights is generalized to cope with azimuth measurement steps whichchange with elevation. It is shown how to obtain a set of reduced-order IPTFs using tools suchas balanced model reduction and spectral smoothing. Detailed comparisons between theaccuracy and the efficiency of the IPTF-based and the bilinear interpolation methods (thelatter using the HRTFs directly) are provided. A practical set of IPTFs is built and applied toa computationally efficient interpolation scheme, yielding results perceptually similar to thoseattainable by the bilinear method. Therefore the proposed method is promising for real-timegeneration of spatial sound.

0 INTRODUCTION

The interest in so-called 3D-sound techniques [1] hasincreased recently. Besides cinema and home theater, theycan be found in applications ranging from a simple videogame to sophisticated virtual reality simulators. Also,there is an increasing level of interaction between usersand the virtual environment in which they are immersed.Contrasting with cases when the location and/or move-ment of sound sources is predefined (as in a moving pic-ture exhibition), which allow off-line digital signal pro-cessing, interactive systems require real-time processing[2] and may impose severe restrictions on the computa-tional complexity.

The present work addresses the problem of spatialsound generation employing reduced computational com-plexity when the 3D sound for a single virtual source issupplied binaurally through headphones. In this case thedirection of arrival of the sound can be controlled by fil-tering the original monaural signal through a proper set ofpreviously measured head-related transfer functions(HRTFs) [3], [4]. These functions model the paths fromthe virtual sound source to the listener’s ears.

In practical systems only a finite number of HRTFs isavailable. They are usually obtained from their corre-

sponding head-related impulse responses (HRIRs) mea-sured under controlled conditions on a spherical surface ofconstant radius only for a predefined set of elevationangles � and azimuth angles � with respect to the ears ofa target head at its center. In order to place the soundsource at an arbitrary position on this surface, it is neces-sary to estimate the HRTF associated with the desiredsource location from the set of measured ones via an in-terpolation scheme. In particular, a realistic rendering ofmoving sound requires a large set of HRTFs so that nonoticeable discontinuity occurs along its path. As a con-sequence, in real-time systems the computational cost toperform HRTF interpolation must be kept to a minimumwithout degrading the final perceptual result. In this sensethe tradeoff between accuracy and system complexity iscritical when designing 3D audio systems. The target hereis to render faithful location of a virtual sound sourcemoving rapidly under user control. One could think ofapplications such as virtual reality simulators, videogames, or even films in which the perspective can becontrolled by the user.

When dealing with a single source,1 a straightforwardway to perform HRTF interpolation is the bilinear ap-

*Manuscript received 2003 October 8; revised 2004 June 29.This work was presented in part at the AES 22nd InternationalConference in Espoo, Finland, 2002 June 15–17.

1Some methods, such as [5], [6], provide alternatives to theHRTF-based model. They split the overall model into two parts,one related to the spectral characteristics, the other to the spatialones. This interpretation can be more efficient than the direct useof HRTFs when dealing with a multisource case.

PAPERS

J. Audio Eng. Soc., Vol. 52, No. 9, 2004 September 915

Page 2: Interpositional Transfer Function for 3D-Sound Generation*diniz/papers/ri53.pdf · sound generation employing reduced computational com-plexity when the 3D sound for a single virtual

proach [7], [1], which consists of computing the HRTFassociated with a given point on the reference sphere as aweighted mean of the measured HRTFs associated withthe four nearest points that circumscribe the desired point.

Consider that a set of HRIRs was measured for fixed-angle steps �grid and �grid relative to azimuth and eleva-tion, respectively. An estimate of the HRIR at an arbitrarycoordinate (�, �) on the spherical surface, as shown in Fig.1, can be obtained by

h�k� = �1 − c���1 − c��ha�k� + c��1 − c��hb�k�+ c�c�hc�k� + �1 − c��c�hd�k� (1)

where ha(k), hb(k), hc(k), and hd(k) are the HRIRs associ-ated with the four points nearest to the desired position.The parameters c� and c� are computed as

c� =C�

�grid=

� mod �grid

�grid, c� =

C�

�grid=

� mod �grid

�grid(2)

where C� and C� are the relative angular positions definedin Fig. 1.

Low-cost interpolation methods have been the subjectof several works [8]–[11], which usually focus on thereduction of HRTF orders. Surprisingly, the way theHRTF interpolation is carried out is not frequently ad-dressed. In a recent work [12] the authors propose analternative interpolation procedure, similar in reasoning tothe bilinear method, but based on a new auxiliary transferfunction, called interpositional transfer functions (IPTFs),which leads to lower complexity than that of the bilinearinterpolation without perceptible losses in the modelingaccuracy.

Since the IPTFs can be represented by models of re-duced order without implying substantial deviations in theoverall model, the proposed method exhibits reduced com-putational complexity.

For the sake of completeness, the present work starts byreviewing the definition and use of IPTFs for the interpo-lation of HRTFs [12]. The computation of the interpola-tion weights is generalized for the case in which the azi-muth measurement steps are elevation dependent. Then apractical procedure to obtain a complete set of low-orderIPTFs is described. An interpolation scheme based on low-order IPTFs is carefully assessed against the bilinearmethod.

In the next section the IPTF is defined. In Section 2 theIPTF-based interpolation method is presented in detail.Comparisons between the performances of the IPTF-basedand the bilinear interpolations are presented in Section 3.In Section 4 the simplification of IPTFs is achievedthrough balanced model reduction [13] with the aid of aspectral smoothing tool [14]. Section 5 describes and dis-cusses the experimental results. Conclusions are drawn inSection 6.

1 INTERPOSITIONAL TRANSFER FUNCTION (IPTF)

A practical set of HRTFs is composed of pairs of func-tions, each pair corresponding to one predefined locationof the sound source. As seen in Fig. 2(a) the sound pathsfrom the sound source to the listener’s nearer and fartherears are referred to as ipsi- and contralateral HRTFs, re-spectively. Fig. 2(b) shows the alternative representationfor the HRTF pair proposed in [11]. In that work theauthors show that it is possible to represent the pair of ipsi-and contralateral HRTFs associated with a given positionas depicted in Fig. 2, where the ratio between contra- andipsilateral HRTFs is called interaural transfer function(IATF)2. The ipsilateral HRTF is chosen to be modeleddirectly since it conveys no information about sound dif-fraction on the head, thus allowing lower order modelingthan the contralateral function [11]. In addition, the IATFis a ratio between two similar transfer functions, whichallows for low-order modeling. Therefore the IATF-basedconfiguration yields a more efficient realization of thecontralateral HRTF.

Intuitively one can expect that HRTFs from near posi-tions to the same ear should be very similar. This obser-vation inspires the idea of defining an IPTF, which allowsthe calculation of the HRTF associated with a certain po-sition as a cascade of the HRTF from a contiguous posi-tion and an incremental error function of reduced order.

The IPTF can be defined as

IPTFi,f =HRTFf

HRTFi(3)

2In the original work the interaural transfer function is calledITF. Here the notation IATF is adopted to avoid ambiguity.

Fig. 1. Bilinear interpolation—graphical interpretation. Fig. 2. HRTF modeling. (a) Direct. (b) IATF-based.

FREELAND ET AL. PAPERS

J. Audio Eng. Soc., Vol. 52, No. 9, 2004 September916

Page 3: Interpositional Transfer Function for 3D-Sound Generation*diniz/papers/ri53.pdf · sound generation employing reduced computational com-plexity when the 3D sound for a single virtual

where HRTFi and HRTFf are the HRTFs associated withthe initial and final points, respectively, as shown in Fig.3. Note that all the vectors were drawn opposite to thedirection of true sound paths to make the generation ofHRTFf starting from HRTFi clearer.

In order to guarantee the stability of the IPTFs it suf-fices that HRTFf and HRTFi have minimum phase. Thischaracteristic does not seem to affect the perception of thedirection of arrival of the generated sound [15]. On theother hand, the difference between the time delays of leftand right HRTFs—called interaural time difference (ITD)[9]—is crucial for the perceptual determination of sounddirection. As a consequence, the inherent delays of theoriginal HRTFs must be preserved, since they representthe propagation time from the sound source to the ears. Ina practical implementation these initial delays could beextracted from the impulse responses and saved, whereasthe remaining responses could be transformed into theirminimum-phase versions. Then the saved delays would beassociated back to the minimum-phase transfer function.An alternative approach consists of deriving the mini-mum-phase versions of HRTFf and HRTFi, and thenevaluating the relative delays of the original and mini-mum-phase versions based only on the linear componentof their excess of phase [2].

The availability of an efficient approximation for theIPTFs is assumed here. This topic will be specificallydiscussed in Section 4.

2 IPTF-BASED INTERPOLATION OF HRTFs

The IPTF-based interpolation follows the same basicprinciple as the bilinear method. However, in the formermethod just three HRTFs are used for interpolating an-other one. In addition, two of them can be indirectly ap-proximated using reduced-order IPTFs, which further con-tributes to reduce the computational cost of the synthesisprocedure. Due to the three-function implementation, thespherical surface must be divided into triangular regions.

Once the triangular regions are chosen, the HRTF for agiven point P will be calculated using the HRTFs relatedto the three vertices of that region. Fig. 4 depicts twodifferent situations in the same quadrangular region of thereference sphere grid for which measured HRTFs areavailable for points A, B, C, and D, whereas the HRTFsassociated with point P are to be estimated. From Fig. 4(a)one can see that, given the arbitrary position P to be de-scribed, the vertices of the P region are denoted by A, B,

and C, with A and B having the same elevation. Thecoordinates of the nearest point (A in this example) will beassociated with the original HRTF, whereas the coordi-nates of the other two points will be related to the IPTF-approximated HRTFs. Now the HRTF associated with Pcan be estimated through a weighted mean of the HRTFsassociated with A, B, and C. It is important to note that allthe points inside the shaded region are interpolated usingthe HRTFs related to these three points, of which two areapproximations and the nearest one is an original HRTF.Another example is shown in Fig. 4(b), where anotherpoint P is inside a different ABC region. Notice that thevertices were relabeled to allow referring to the verticeswith the same elevation in the shaded region as A and B,like before. In this new figure the nearest vertex is C,which will be associated with an original HRTF.

According to the assignment of these vertices, an HRTFrelated to P can be described by

HRTFP = wAHRTFA + wBHRTFB + wCHRTFC (4)

where

wC =��

��grid(5)

wB =��

��grid(6)

wA + wB + wC = 1 (7)

and the required angular distances are shown in Fig. 5 anddefined by

�� = �P − �A (8)

�� = �P − �X (9)

��A = �P − �A (10)

��AC = �C − �A (11)

��grid = �B − �A (12)

��grid = �C − �A. (13)

Since

��

��grid=

��A − ��

��AC(14)

Fig. 4. Details of reference spherical surface. A, B, C, and D haveknown HRTFs; P is the position requiring an interpolated ap-proximation to the HRTF.Fig. 3. IPTF-computed HRTFf—graphical interpretation.

PAPERS IPTFs FOR 3D-SOUND GENERATION

J. Audio Eng. Soc., Vol. 52, No. 9, 2004 September 917

Page 4: Interpositional Transfer Function for 3D-Sound Generation*diniz/papers/ri53.pdf · sound generation employing reduced computational com-plexity when the 3D sound for a single virtual

one can get

�� = ��A −��

��grid��AC . (15)

Denoting by I the vertex nearest to P and by F� and F�

the other two vertices that form the triangular regionaround P, according to Fig. 5, the IPTF-based interpolationcan be written as

HRTFP = �HRTFI + �HRTFF�+ �HRTFF�

= HRTFI�� + �IPTFI,F�+ �IPTFI,F�

� (16)

where IPTFI,F�and IPTFI,F�

are defined by Eq. (3). Thecoefficients �, �, and � are chosen according to the as-signment of the labels I, F�, and F�. As an example, in Fig.5 A, B, and C are equivalent to I, F�, and F�, respectively.Therefore Eq. (16) becomes

HRTFP = wAHRTFA + wBHRTFB + wCHRTFC

= HRTFA�wA + wBIPTFA,B + wCIPTFA,C�. (17)

It is worth mentioning that this triangular configurationis closely related to the formulation of the vector-basedamplitude panning method [16] for virtual source position-ing, in the specific case of three loudspeakers. This issuewill not be discussed here.

The block diagram of Fig. 6 illustrates the interpolationprocedure for both the left and right channels (identified

by superscripts l and r, respectively). In this diagram theHRTF and IPTF blocks represent the transfer functionswithout their inherent delays. In a simpler approximationprocedure, the delay of HRTFI

l,r, DIl,r, is merged with those

of the IPTFs, dI,F�

l,r , and dI,F�

l,r , so that the inherent delay andthe minimum-phase version of the desired HRTF can beinterpolated separately [3]. The resulting delay

�l,r = �DIl,r + ��DI

l,r + dI,F�

l,r � + ��DIl,r + dI,F�

l,r � (18)

= �� + � + ��DIl,r + �dI,F�

l,r + �dI,F�

l,r (19)

= DIl,r + �dI,F�

l,r + �dI,F�

l,r (20)

can be implemented using fractional delay filters [17].It is worth mentioning that the use of low-order IPTFs

is the core of the simplification attainable by the IPTF-based interpolation. The gain in computational efficiencythat results from applying a triangular configuration di-rectly to the measured HRTFs is trivial and, with no fur-ther consideration, amounts roughly to 25% compared tothe quadrangular geometry associated with the bilinearmethod. In fact, three-point linear combination methodswere attempted earlier [2]. In addition to the triangularformulation, the possibility of designing IPTFs using low-order models is what makes the proposed method appeal-ing with regard to computational complexity. This issue isdiscussed in greater detail in the following sections.

3 IPTF-BASED VERSUS BILINEAR METHOD

In this section the performance of the IPTF-based in-terpolation method is compared to that of the bilinearmethod by evaluating their frequency response for a de-fined range of azimuth and elevation angles.

Initially only differences in geometry and in weight cal-culations are considered, that is, the IPTF-based interpo-lation takes the IPTFs in Eq. (3) as the direct ratio betweenthe corresponding HRTFs without any simplification,which can be considered as a triangular interpolation ap-plied to the HRTFs, in the form

HRTFP � �HRTFI + �HRTFF�+ �HRTFF�

. (21)Fig. 5. Angular distances used to obtain parameters wA, wB,and wC.

Fig. 6. Block diagram of IPTF-based interpolation procedure.

FREELAND ET AL. PAPERS

J. Audio Eng. Soc., Vol. 52, No. 9, 2004 September918

Page 5: Interpositional Transfer Function for 3D-Sound Generation*diniz/papers/ri53.pdf · sound generation employing reduced computational com-plexity when the 3D sound for a single virtual

The underlying idea behind the tests consists in con-fronting the spatial frequency response surfaces (SFRSs)[18], [19] of the results from both interpolation methods.Each SFRS shows, for a given frequency, the dependenceof the magnitude response on the changes in elevation andazimuth.

The equations of the bilinear interpolation are originallydefined for equally spaced grids in both � and �. In orderto provide a fair comparison between the two methods, therange of evaluation of the bilinear interpolation was re-stricted to −20° < � < 20°, where the available set ofHRTFs is spaced uniformly (see Table 1).

All simulations in this work used a set of 128-sample-long measured HRIRs sampled at 44.1 kHz, originatedfrom [3]. They are available for elevation steps of 10° andvariable azimuth steps, as shown in Table 1.

Figs. 7–10 provide a visual comparison of performancesof the IPTF and bilinear methods. The first and secondcolumns show the SRFSs related to the bilinear (|HB(�, �,�)|) and the IPTF-based (|HIPTF(�, �, �)|) interpolations,respectively. Different magnitude scales were adopted ac-cording to the dynamic range at each frequency bin. Onlyleft-ear HRTFs are considered, since the opposite func-tions can be inferred by symmetry. The frequency bins arechosen according to [18].

To improve readability, the third column illustrates thedifference in dB between the two SFRSs in the same row,given by

���, �, �� = 20 log10� |HIPTF��, �, ��||HB��, �, ��| �. (22)

This time a lighter region means small differences (nearzero) between correspondent SFRSs. The same range waschosen for all figures in this column in order to make thecomparison between different frequency bins easier.

In spite of the occurrence of more noticeable differencesaround the azimuth angle of 90° (the region more stronglyaffected by head diffraction), the equivalence between thetwo methods is confirmed by the small differences ob-served. It can also be noticed that the worst cases arerelated to low-energy regions of the SFRSs.

Of course, the mere use of the IPTFs, computed as thedirect ratio between the corresponding HRTFs without anysimplification, does not yield an improvement in compu-tational complexity. Since the purpose of this work is topropose efficient spatial sound generation, the next step isto investigate whether the IPTF model can be reduced, asexpected.

Assuming that each HRTF directly used in the interpo-lation procedure requires N multiplications per sample(mps), the complexity of the bilinear interpolation is

Cbilinear � 4N + 12 mps. (23)

On the other hand, if each approximated IPTF requires Mmps, the IPTF-based interpolation scheme requires

CIPTF-based � 2M + N + 7 mps. (24)

Therefore the latter exhibits lower computational com-plexity if

M 3N + 5

2. (25)

For the HRTFs used in the simulations [3] (which have N� 128), that means M 194. On the other hand, it is alsopossible to take advantage of efficient approximations ofHRTFs [9], which allow for model-order reduction. Forexample, assuming N � 33, which is the lowest orderattained in [9], making M < 52 would suffice to turn theIPTF-based interpolation less expensive computationally.

It should be noticed that since the IPTF-based interpo-lation procedure utilizes a triangular grid, the number ofregions visited by the virtual sound source is increased. Asa result, if any special procedure is necessary at the edgecrossings3, it can result in additional computational load.Of course, at high sampling rates (such as 44.1 kHz) and fornot too fast movements, this overhead becomes irrelevant.

The remaining issue is how much can the IPTFs besimplified so that the approximation errors are acceptable?The next section describes some tools to approximate acomplete set of IPTFs.

4 ORDER REDUCTION OF IPTFs

4.1 Balanced-Model TruncationIn [8] the authors apply the balanced-model truncation

or reduction (BMR) to decrease the order of HRTFs. Sincethe measured HRTFs are in FIR form, the specializedBMR algorithm that approximates FIR filters by low-orderIIR filters described in [13] is used. In the present work theIPTF is the ratio of two HRTFs, and thus it is originally anIIR system. Therefore the simpler formulation of [13] can-not be used. Nevertheless, the balanced-model reductionmethod, in its general formulation, is applicable to the IIRcase and is found to be effective in reducing the order ofthe IPTFs. In the following, a review of the balanced-model reduction is presented.

The state-space representation of a system in which thegrammians of its observability and controllability matrices(see the discussion leading to Eqs. (29) and (30) below)are both diagonal and equal is called a balanced model.The main advantage of this type of model is the possibilityof ordering the system states based on their significance tothe representation of the system characteristics. This is avery important feature, which allows model simplificationby direct truncation, that is, by discarding its “less impor-tant” states.

3More on this topic in Section 5.

Table 1. Azimuth measurement steps.

Elevation �(deg)

Azimuth resolution �grid

(deg)

−20 to 20 5−30 and 30 6−40 and 40 45/750 860 1070 1580 30

PAPERS IPTFs FOR 3D-SOUND GENERATION

J. Audio Eng. Soc., Vol. 52, No. 9, 2004 September 919

Page 6: Interpositional Transfer Function for 3D-Sound Generation*diniz/papers/ri53.pdf · sound generation employing reduced computational com-plexity when the 3D sound for a single virtual

Consider the state-space representation of an IPTF,

x�k + 1� = Ax�k� + bu�k�

y�k� = cTx�k� + du�k�(26)

where A, b, c, and d are the transition matrix, input vector,output vector, and feedforward coefficient, respectively.x(k) is the state vector at instant k, and u(k) and y(k) are theinput and output signals at instant k, respectively.

Fig. 7. Comparison of SFRSs.—I. (a) Bilinear. (b) IPTF-based. (c) Difference in dB.

FREELAND ET AL. PAPERS

J. Audio Eng. Soc., Vol. 52, No. 9, 2004 September920

Page 7: Interpositional Transfer Function for 3D-Sound Generation*diniz/papers/ri53.pdf · sound generation employing reduced computational com-plexity when the 3D sound for a single virtual

Define

Wo � [c ATc . . . (AT)i

c . . . ]T (27)

and

Wc � [b Ab . . . Aib . . . ] (28)for i � 1, 2, 3, . . . , as the observability and controllabilitymatrices of the system, respectively. Their respectivegrammian matrices are given by

P � WcWcT (29)

and

Q � WoTWo. (30)

If rank [A] � J, the rank-J Hankel matrix

H = WoWc = �h1 h2 · · ·

h2 h3

···· · ·� (31)

Fig. 8. Comparison of SFRSs—II. (a) Bilinear. (b) IPTF-based. (c) Difference in dB.

PAPERS IPTFs FOR 3D-SOUND GENERATION

J. Audio Eng. Soc., Vol. 52, No. 9, 2004 September 921

Page 8: Interpositional Transfer Function for 3D-Sound Generation*diniz/papers/ri53.pdf · sound generation employing reduced computational com-plexity when the 3D sound for a single virtual

can be constructed, where

hi � cT Ai−1 b, i � 1, 2, 3, . . . (32)

are the samples of the system impulse response.Now H can be decomposed as

H � V1�V2 (33)

where � is the matrix containing the singular values�1, . . . , �J of H along the first J positions of its main

diagonal, by convention arranged in descending order oftheir magnitudes [20]. Let ∑ be the upper left J × J sub-matrix of �.

Since the square of each singular value of H, �i2, is

equal to a corresponding eigenvalue of the product PQ[21], [13], denoted by i, for i � 1, 2, 3, . . . , then

PQ � K∑2K−1. (34)

Fig. 9. Comparison of SFRSs—III. (a) Bilinear. (b) IPTF-based. (c) Difference in dB.

FREELAND ET AL. PAPERS

J. Audio Eng. Soc., Vol. 52, No. 9, 2004 September922

Page 9: Interpositional Transfer Function for 3D-Sound Generation*diniz/papers/ri53.pdf · sound generation employing reduced computational com-plexity when the 3D sound for a single virtual

Hence one can show that the grammians Pb and Qb of thedesired balanced system must be such that

Pb = Qb = � = diag�� 1, . . . , � J�

= diag��1, . . . , �J�.(35)

Then the transformation that turns the original system (A,b, c, and d) into its balanced form (Ab, bb, cb, and d)through

Ab � T−1 AT, bb � T−1 b, cb � TT c (36)

is

T � S−1 U∑1/2 (37)

where S results from decomposing the matrix Q into theform

Q � STS. (38)

Matrices U and ∑ can be computed from the followingdecomposition:

SPST � U∑2UT (39)

where

UUT � I. (40)

Note that Eq. (39) holds since SPST � SPQS−1 is sym-metric and has the same eigenvalues as the product PQ.

Consider the case where, according to some error cri-terion, only the j < J larger singular values are judged to besufficient to describe the IPTF. In this case matrix ∑ canbe written in the form

� =��1 0

0 �2� (41)

where 0 are null matrices and

�1 = diag��1, . . . , �j� (42)

�2 = diag��j+1, . . . , �J�. (43)

Accordingly the balanced-model matrices can be de-composed into

Ab =�A1,1 A1,2

A2,1 A2,2�, bb =�b1

b2�, cb =�c1

c2�

(44)

such that (A1,1, b1, c1, and d) are the state-space matricesof the subsystem of order j that can approximate the origi-nal IPTF accurately.

The last step, if desired, consists of converting the state-space representation back to the transfer function form.

4.2 Spectral SmoothingThe design of IPTFs through the balanced-model reduc-

tion method has some peculiarities. Since the target fre-quency response results from the ratio of two HRTFs, itmay include a number of sharp spectral peaks and valleysat some locations. As a consequence, numerical problemsmay arise during the simplification procedure.

On the other hand, psychoacoustics shows that fre-quency discrimination in human hearing varies with fre-quency in such a way that it becomes coarser as the fre-quency increases. To some extent the logarithmic westernmusical scale is a consequence of this phenomenon [22],[23]. Furthermore, measured HRIRs tend to be less reli-able at higher frequencies where, due to the sound attenu-ation, measurements have lower signal-to-noise ratios.

Considering the aforementioned facts, one can think ofapplying a spectral smoother to the HRTF response insuch a way that the degree of smoothness increases towardthe higher frequencies of IPTFs. For instance, thesmoothed spectral response at any frequency f can be com-puted as the average spectral response within the rangedelimited by the frequencies

f1 =f

�K, f2 = f �K (45)

which are geometrically equidistant from f by a factor of√K. In the DFT domain with the dc frequency indexed as

Fig. 10. Comparison of SFRSs—IV. (a) Bilinear. (b) IPTF-based. (c) Difference in dB.

PAPERS IPTFs FOR 3D-SOUND GENERATION

J. Audio Eng. Soc., Vol. 52, No. 9, 2004 September 923

Page 10: Interpositional Transfer Function for 3D-Sound Generation*diniz/papers/ri53.pdf · sound generation employing reduced computational com-plexity when the 3D sound for a single virtual

n � 0, if f is the frequency associated with the spectral binn, the corresponding indices of the approximated limitfrequencies can be obtained by

n1 = � n

�K�, n2 = �n �K� (46)

respectively, denoting ��� as the lowest integer greater thanor equal to (�) and ��� as the greatest integer lower than orequal to (�). The value of K � 2a gives a window width ofa octaves. For example, using K � 21/3 one defines aone-third-octave range [9], [24]. However, a fixed valuefor K may not smooth the high-frequency range of theHRTFs sufficiently without degrading the low-frequencyrange. Then a linearly varying value K(n) given by

K�n� = 1 − n1 − K

L�2 − 1(47)

was defined, where L is the FFT length. A constant K �21/6 was chosen, resulting in a one-sixth-octave windowwidth at n � L/2 − 1.

Two smoothing methods are considered, namely, thesquare magnitude [9] and the complex frequency response[14] methods, where in each case an average is computedon a sliding window. In the former, smoothing is attainedby computing the arithmetic average of the squared mag-nitude response in the window range and taking the squareroot of the result, that is,

|Hs�n�| =� 1

n2 − n1 + 1 �l=n1

n2

|Ho�l�|2 (48)

where Ho(n) and Hs(n) are, respectively, the original andthe smoothed frequency responses. The latter methodtakes into account the phase information by computing themean over the complex frequency response, which can beexpressed as

Hs�n� =1

n2 − n1 + 1 �l=n1

n2

Ho�l�. (49)

These two methods are based on the arithmetic mean,which treats peak and valley occurrences differently. Analternative solution is to replace the arithmetic mean bythe geometric mean in Eq. (49). Thus Eq. (49) becomes

Hs�n� =n2−n1+1��

l=n1

n2

Ho�l�. (50)

In each case n1 and n2 are expressed by Eq. (46).When using spectral smoothing in the design of low-

order IPTFs it was verified that the smoother design givenby Eq. (50) is more suitable. Smoothing was applied to thespectrum of each HRTF during the minimum-phase trans-formation algorithm [4], [24] prior to log-magnitude op-eration (see [25, p. 784]).

4.3 Order Reduction AlgorithmThe techniques mentioned in the preceding are the basic

tools to build a complete set of low-order IPTFs. In the endthese IPTFs are used together with measured HRTFs toimplement an efficient moving sound system.

The procedure to obtain a complete set of IPTFs con-sists of three stages.

• Each HRTF on the reference spherical surface grid istransformed into its smoothed minimum-phase version.

• For each smoothed HRTF one IPTF is defined withrespect to each of its four contiguous HRTFs.

• Balanced-model reduction is applied to each previouslydefined IPTF in order to obtain a reduced-order modelby which the IPTF can be well represented.

In fact, this strategy implies redundancy since two IPTFscan be defined for each two contiguous HRTFs, dependingon the choice of HRTFf and HRTFi in Eq. (3). Obviouslyonly one low-order IPTF per HRTF pair needs to be saved.However, redundant functions allow the choice of thefunction that matches the response of the original HRTFratio best. This flexibility also helps to deal with cases thatexhibit numerical problems.

5 EXPERIMENTAL RESULTS

In Section 3 a graphical comparison between the per-formances of the IPTF-based and the bilinear interpolationmethods was provided. In the following, the results ofIPTF-based interpolation with reduced-order IPTFs arecompared with those obtained through IPTFs with no or-der reduction. Once more, several SFRSs and their corre-sponding differences are used for visualization. Thistime elevation angles from −40° to 90° are considered,thus covering the entire region with measured HRTFs.

Figs. 11–14 refer to left-ear HRTFs; right-ear sur-faces can be obtained by simply inverting the azimuthaxis. Each row contains the SFRSs related to interpolationthrough nonreduced-order (a) and tenth-order (b) IPTFs,as well as their differences in dB (c). The scales and fre-quency bins were chosen according to the criteria ex-plained in Section 3.

One can observe again that corresponding SFRSs arequite similar, exhibiting small differences. Even thoughsome discontinuities appear in these differences, they canbe considered negligible when compared to the overallenergy of each frequency bin, thus attesting to the reliabil-ity of low-order IPTF modeling. Like before, the worstcase tends to occur around the azimuth angle of 90°.

A complete set of tenth-order IIR IPTFs was obtainedvia the procedure described in Section 4.3, and all theexamples of model-order reduction shown refer to this set.Each IPTF in this set corresponds to M � 21, satisfyingthe efficiency bounds for N � 33 from Section 3 (M 52). Thus the proposed interpolation method is more com-putationally efficient than the bilinear method.

Since the complexity of the HRTFs around the refer-ence sphere varies with the location, one could considerusing IPTFs of variable orders—lower order whereverpossible and higher order wherever needed. As a con-sequence the processing time would vary with the posi-tion. In real-time systems the worst-case time should beconsidered.

Based on the aforementioned results, it is feasible to usethe IPTF-based interpolation in a simple moving sound

FREELAND ET AL. PAPERS

J. Audio Eng. Soc., Vol. 52, No. 9, 2004 September924

Page 11: Interpositional Transfer Function for 3D-Sound Generation*diniz/papers/ri53.pdf · sound generation employing reduced computational com-plexity when the 3D sound for a single virtual

system restricted only to angular positioning of a singlesound source. In this case the system receives successivesamples of the monaural signal as well as the informationrelated to the virtual source location on the referencesphere. Then based on considerations about timing andposition accuracy, the reference points where the virtualsource must be located are defined.

When performing HRTF interpolation through theIPTF-based method, the following procedure is followed.For each point P in the source trajectory, the nearest mea-sured HRTF is found as well as the two IPTFs that com-plete the HRTF triangle around P. Then the HRTF in P isinterpolated according to Eq. (16) and used to process theaudio signal.

Fig. 11. Comparison of SFRSs—I. (a) Using nonreduced IPTFs. (b) Using tenth-order IPTFs. (c) Difference in dB.

PAPERS IPTFs FOR 3D-SOUND GENERATION

J. Audio Eng. Soc., Vol. 52, No. 9, 2004 September 925

Page 12: Interpositional Transfer Function for 3D-Sound Generation*diniz/papers/ri53.pdf · sound generation employing reduced computational com-plexity when the 3D sound for a single virtual

Finally it must be remembered that using the proposedtechnique, switching between successive interpolatedHRTFs involves the switching of IIR filters whenever thesource crosses the boundary between triangles. With nospecial treatment this introduces spurious transients thatmay cause audible artifacts in the synthesized signal. IIRfilters involve the feedback of past outputs. When the filtercoefficients change, the saved outputs related to the oldfilter make no more sense to the new one, thus creating the

transients. This problem has been addressed successfullyin [26], which proposes the redundancy of old and newfilter states as a way to minimize these effects. This strat-egy should be included in the system.

Discussions on other time-varying issues, such as therate of position updates, will be addressed in future work.

A system with the characteristics described here wasimplemented. Performance evaluation through informallistening tests revealed that using the IPTF-based interpo-

Fig. 12. Comparison of SFRSs—II. (a) Using nonreduced IPTFs. (b) Using tenth-order IPTFs. (c) Difference in dB.

FREELAND ET AL. PAPERS

J. Audio Eng. Soc., Vol. 52, No. 9, 2004 September926

Page 13: Interpositional Transfer Function for 3D-Sound Generation*diniz/papers/ri53.pdf · sound generation employing reduced computational com-plexity when the 3D sound for a single virtual

lation with the designed set of tenth-order IPTFs yieldsgood results, comparable to those of a similar systembased on bilinear interpolation. Considering N � 33 [9]and tenth-order IPTFs (M � 21) from Eq. (23), the bilin-ear method requires 144 mps, whereas from Eq. (24) theproposed method achieves 82 mps, saving up to 62 mps, ormore than 40% of the computational effort required by thebilinear method. Notice, however, that comparing the tri-

angular interpolations performed directly on HRTFs andthrough IPTFs, the gain in computational complexity is 2× (N − M), which means 2 × 12 � 24 mps for the valuessuggested here. Therefore the use of IPTFs alone saves upto 24 mps over 106 mps, or more than 22%.4

4Some audio samples can be found at www.lps.ufrj.br/sound/journals/JAES04.

Fig. 13. Comparison of SFRSs—III. (a) Using nonreduced IPTFs. (b) Using tenth-order IPTFs. (c) Difference in dB.

PAPERS IPTFs FOR 3D-SOUND GENERATION

J. Audio Eng. Soc., Vol. 52, No. 9, 2004 September 927

Page 14: Interpositional Transfer Function for 3D-Sound Generation*diniz/papers/ri53.pdf · sound generation employing reduced computational com-plexity when the 3D sound for a single virtual

6 CONCLUSION

A recently proposed method for the interpolation ofhead-related transfer functions for binaural audio applica-tions [12] was reviewed and extended. It combines a tri-angular interpolation scheme with an auxiliary functioncalled interpositional transfer function, which can lead toefficient implementations. Extensive simulations com-pared IPTF-based and bilinear interpolations in terms ofaccuracy and computational complexity.

A direct comparison between the bilinear interpolationand the IPTF method indicates that the latter is more ef-ficient in terms of computational complexity. The IPTFmethod saves more than 40% of the operations required bythe bilinear one, thanks to the use of a triangular geometryand also a reduced-order model to represent the IPTFs—which alone accounts for up to 22% of the savings. For theorder-reduction task the balanced-model reduction andspectral smoothing methods were proposed. The IPTF-based interpolation method was used in a system that gen-erates moving spatial sound, and the results were encour-aging. Further investigations consist in proposing moreeffective ways to simplify the IPTFs and the realization ofsystematic listening tests to confirm the objective and in-formal subjective results obtained.

7 ACKNOWLEDGMENT

The authors wish to thank the CNPq Brazilian agencyfor funding this research and B. Gardner and K. Martin forkindly providing the set of HRTFs used in this work.

8 REFERENCES

[1] D. R. Begault, 3D Sound for Virtual Reality andMultimedia (Academic Press, Cambridge, MA, 1994).

[2] J. M. Jot, O. Warusfel, and V. Larcher, “DigitalSignal Processing Issues in the Context of Binaural andTransaural Stereophony,” presented at the 98th Conven-tion of the Audio Engineering Society, J. Audio Eng. Soc.(Abstracts), vol. 43, p. 396 (1995 May), preprint 3980.

[3] B. Gardner and K. Martin, “HRTF Measurements ofa KEMAR Dummy-Head Microphone,” Tech. Rep. 280,MIT Media Lab., Cambridge, MA (1994 May).

[4] C. I. Cheng and G. H. Wakefield, “Introduction toHead-Related Transfer Functions (HRTFs): Representa-tions of HRTFs in Time, Frequency, and Space,” J. AudioEng. Soc., vol. 49, pp. 231–249 (2001 Apr.).

[5] J. Chen, B. D. Van Veen, and K. E. Hecox, “Exter-nal Ear Transfer Function Modeling: A Beamforming Ap-proach,” J. Acoust. Soc. Am., vol. 92, pp. 1933–1944,(1992 Oct.).

[6] V. Larcher, O. Warusfel, J. M. Jot, and J. Guyard,“Study and Comparison of Efficient Methods for 3D Au-dio Spatialization Based on Linear Decomposition ofHRTF Data,” presented at the 108th Convention of theAudio Engineering Society, J. Audio Eng. Soc. (Ab-stracts), vol. 48, p. 351 (2000 Apr.), preprint 5097.

[7] L. Savioja, J. Huopaniemi, T. Lokki, and R. Vaan-anen, “Creating Interactive Virtual Acoustic Environ-ments,” J. Audio Eng. Soc., vol. 47, pp. 675–705 (1999Sept.).

[8] J. Mackenzie, J. Huopaniemi, V. Valimaki, and I.Kale, “Low-Order Modeling of Head-Related TransferFunctions Using Balanced Model Truncation,” IEEE Sig-nal Process. Letts., vol. 4, pp. 39–41 (1997 Feb.).

[9] J. Huopaniemi, N. Zacharov, and M. Karjalainen,“Objective and Subjective Evaluation of Head-RelatedTransfer Function Filter Design,” J. Audio Eng. Soc., vol.47, pp. 218–239 (1999 Apr.).

[10] C. Avendano, R. O. Duda, and V. R. Algazi,“Modeling the Contralateral HRTF,” in Proc. AES 16thInt. Conf. (Rovaniemi, Finland, 1999 Apr.), pp. 313–318.

[11] G. Lorho, J. Huopaniemi, N. Zacharov, and D.Isherwood, “Efficient HRTF Synthesis Using an InterauralTransfer Function Model,” in Proc. EUSIPCO’2000(Tampere, Finland, 2000 Sept.), pp. 2213–2216.

[12] L. W. P. Biscainho, F. P. Freeland, and P. S. R. Di-niz, “Using Inter-Positional Transfer Functions in 3D-Sound,” in Proc. IEEE Int. Conf. on Acoustics, Speech andSignal Processing (Orlando, FL, 2002 May).

[13] B. Beliczynsky, I. Kale, and G. D. Cain, “Approxi-

Fig. 14. Comparison of SFRSs—IV. (a) Using nonreduced IPTFs. (b) Using tenth-order IPTFs. (c) Difference in dB.

FREELAND ET AL. PAPERS

J. Audio Eng. Soc., Vol. 52, No. 9, 2004 September928

Page 15: Interpositional Transfer Function for 3D-Sound Generation*diniz/papers/ri53.pdf · sound generation employing reduced computational com-plexity when the 3D sound for a single virtual

mation of FIR by IIR Digital Filters: An Algorithm Basedon Balanced Model Reduction,” IEEE Trans. Signal Pro-cess., vol. 40, pp. 532–542 (1992 Mar.).

[14] P. D. Hatziantoniou and J. N. Mourjopoulos, “Gen-eralized Fractional-Octave Smoothing of Audio andAcoustic Responses,” J. Audio Eng. Soc., vol. 48, pp.259–280 (2000 Apr.).

[15] A. Kulkarni, S. K. Isabelle, and H. S. Colburn, “Onthe Minimum-Phase Approximation of Head-RelatedTransfer Functions,” in IEEE Proc. WASPAA’1995 (NewPaltz, NY, 1995 Oct.).

[16] V. Pulkki, “Virtual Sound Source Positioning Us-ing Vector Base Amplitude Panning,” J. Audio Eng. Soc.,vol. 45, pp. 456–466 (1997 June).

[17] T. I. Laakso, V. Valimaki, M. Karjalainen, andU. K. Laine, “Splitting the Unit Delay,” IEEE Signal Pro-cess. Mag., vol. 13, pp. 30–59 (1996 Jan.).

[18] C. I. Cheng and G. H. Wakefield, “Spatial Fre-quency Response Surfaces (SFRS’s): An Alternative Vi-sualization and Interpolation Technique for Head-RelatedTransfer Functions (HRTF’s),” in Proc. AES 16th Int.Conf. (Rovaniemi, Finland, 1999 Apr.), pp. 147–159.

[19] K. Hartung, J. Braasch, and S. J. Sterbing, “Com-parison of Different Methods for the Interpolation of

Head-Related Transfer Functions,” in Proc. AES 16th Int.Conf. (Rovaniemi, Finland, 1999 Apr.), pp. 319–329.

[20] G. Strang, Linear Algebra and Its Applications, 3rded. (Saunders, Orlando, FL, 1988).

[21] K. Glover, “All Optimal Hankel-Norm Approxi-mation of Linear Multivariable Systems and Their L�-ErrorBounds,” Int. J. Cont., vol. 39, pp. 1115–1117 (1984).

[22] J. G. Roederer, The Physics and Psychophysics ofMusic–An Introduction, 3rd ed. (Springer, New York,1994).

[23] D. Butler, The Musician’s Guide to Perception andCognition (Schirmer, New York, 1992).

[24] J. Huopaniemi and J. O. Smith, III, “Spectral andTime-Domain Preprocessing and the Choice of ModelingError Criteria for Binaural Digital Filters,” in Proc. AES16th Int. Conf. (Rovaniemi, Finland, 1999 Apr.), pp.301–312.

[25] A. V. Oppenheim and R. W. Shafer, Discrete-TimeSignal Processing (Prentice Hall, Englewood Cliffs, NJ,1989).

[26] V. Valimaki and T. I. Laakso, “Suppression ofTransients in Time-Varying Recursive Filters for AudioSignals,” in Proc. IEEE ICASSP’98, vol. 6 (Seattle, WA,1998 May), pp. 3569–3572.

THE AUTHORS

F. P. Freeland L. W. P. Biscainho P. S. R. Diniz

Fabio P. Freeland was born in Rio de Janeiro, Brazil. Hereceived an Electrical Engineering degree from the Fed-eral University of Rio de Janeiro (UFRJ) in 1999 and anM.Sc. degree from COPPE/UFRJ in 2001.

Since 1998 Mr. Freeland has been with the Signal Pro-cessing Laboratory (LPS/COPPE-UFRJ), where he devel-oped his undergraduate project on audio restoration andhis M.Sc. dissertation on spatial audio. He is currentlypursuing a Ph.D. degree from COPPE/UFRJ and teachesat UERJ. His research interests include digital audio pro-cessing, auditory scene synthesis, audio effects, amongothers.

Luiz Wagner Pereira Biscainho was born in Rio deJaneiro, Brazil. He received an M.Sc. degree in 1990 anda D.Sc. degree in 2000, both from the Federal Universityof Rio de Janeiro (UFRJ), Brazil.

From 1985 to 1993 Dr. Biscainho worked for the tele-communications industry. Since 1993 he has been with theDepartment of Electronics and Computer Engineering of

the Polytechnic (DEL/POLI) at UFRJ, where he is cur-rently an associate professor of telecommunications. Hismain research interests include digital signal processing ofmusic and audio. He is a member of the AES and IEEE.

Paulo S. R. Diniz was born in Niteroi, Brazil. He receivedan Electronics Engineering degree (cum laude) from theFederal University of Rio de Janeiro (UFRJ), Brazil, in1978, an M.Sc. degree from the COPPE/UFRJ in 1981,and a Ph.D. degree from the Concordia University, Mon-treal, Que., Canada, in 1984, all in electrical engineering.

Since 1979 Dr. Diniz has been with the Department ofElectronic Engineering (the undergraduate dept.) at UFRJ.He has also been with the Program of Electrical Engineer-ing (the graduate studies dept.), COPPE/UFRJ, since 1984,where he is a professor. He has served as undergraduatecourse coordinator and as chair of the Graduate Depart-ment. He is one of the three senior researchers and co-ordinators of the National Excellence Center in SignalProcessing.

PAPERS IPTFs FOR 3D-SOUND GENERATION

J. Audio Eng. Soc., Vol. 52, No. 9, 2004 September 929

Page 16: Interpositional Transfer Function for 3D-Sound Generation*diniz/papers/ri53.pdf · sound generation employing reduced computational com-plexity when the 3D sound for a single virtual

From 1991 to 1992 Dr. Diniz was a visiting researchassociate in the Department of Electrical and ComputerEngineering, University of Victoria, B.C., Canada. Healso holds a docent position at the Helsinki University ofTechnology, Finland. From 2002 January to 2002 June hewas a Melchor chair professor in the Department of Elec-trical Engineering, University of Notre Dame, NotreDame, IN. His teaching and research interests include ana-log and digital signal processing, adaptive signal process-ing, digital communications, wireless communications,multirate systems, stochastic processes, and electronic cir-cuits. He has published several refereed papers in some ofthese areas.

Dr. Diniz was the technical program chair of the 1995MWSCAS held in Rio de Janeiro, Brazil. He has been on thetechnical committee of several international conferencesincluding ISCAS, ICECS, EUSIPCO, and MWSCAS. Hehas served as vice president for region 9 of the IEEECircuits and Systems Society and as chair of the IEEEDSP technical committee. He has also served as associ-

ate editor for the following Journals: IEEE Transactionson Circuits and Systems II: Analog and Digital Signal Pro-cessing from 1996 to 1999, IEEE Transactions on SignalProcessing from 1999 to 2002, and Circuits, Systems andSignal Processing from 1998 to 2002. He was a distin-guished lecturer of the IEEE Circuits and Systems Societyfrom 2000 to 2001. Currently he is serving as a distinguishedlecturer of the IEEE Signal Processing Society and has re-ceived the 2004 Education Award from the IEEE Circuitsand Systems Society. He has also received the Rio de JaneiroState Scientist award from the governor of Rio de Janeiro.

Dr. Diniz wrote the books ADAPTIVE FILTERING:Algorithms and Practical Implementation, 2nd ed. (Klu-wer Academic Publishers, 2002) and DIGITAL SIGNALPROCESSING: System Analysis and Design, withE. A. B. da Silva and S. L. Netto (Cambridge UniversityPress, Cambridge, UK, 2002). He is a fellow of the IEEE(for fundamental contributions to the design and imple-mentation of fixed and adaptive filters and for electricalengineering education).

FREELAND ET AL. PAPERS

J. Audio Eng. Soc., Vol. 52, No. 9, 2004 September930