the anatomy of language: a review of 100 fmri studies...

27
Ann. N.Y. Acad. Sci. ISSN 0077-8923 ANNALS OF THE NEW YORK ACADEMY OF SCIENCES Issue: The Year in Cognitive Neuroscience The anatomy of language: a review of 100 fMRI studies published in 2009 Cathy J. Price Wellcome Trust Centre for Neuroimaging, UCL, London, UK Address for correspondence:Cathy Price, Wellcome Trust Centre for Neuroimaging, Institute of Neurology, UCL, 12, Queen Square, London WC1N 3BG, UK. c.price@fil.ion.ucl.ac.uk In this review of 100 fMRI studies of speech comprehension and production, published in 2009, activation is reported for: prelexical speech perception in bilateral superior temporal gyri; meaningful speech in middle and inferior temporal cortex; semantic retrieval in the left angular gyrus and pars orbitalis; and sentence comprehension in bilateral superior temporal sulci. For incomprehensible sentences, activation increases in four inferior frontal regions, posterior planum temporale, and ventral supramarginal gyrus. These effects are associated with the use of prior knowledge of semantic associations, word sequences, and articulation that predict the content of the sentence. Speech production activates the same set of regions as speech comprehension but in addition, activation is reported for: word retrieval in left middle frontal cortex; articulatory planning in the left anterior insula; the initiation and execution of speech in left putamen, pre-SMA, SMA, and motor cortex; and for suppressing unintended responses in the anterior cingulate and bilateral head of caudate nuclei. Anatomical and functional connectivity studies are now required to identify the processing pathways that integrate these areas to support language. Keywords: speech; comprehension; production; fMRI; language Introduction The aim of this paper was to provide an up to date and integrated review of functional imaging studies of language published in 2009. I included studies that (i) used fMRI, (ii) were published online be- tween January and mid-October 2009, and (iii) were concerned with the functional anatomy of speech comprehension and production in the healthy adult brain. This excluded papers that used MEG, EEG, TMS, or optical imaging; that were published in 2008 or before; or that focused on populations with acquired or developmental language disorders. I also excluded papers on orthographic processing of writ- ten words, sign language, and multilingualism: each of these topics requires its own review. Having selected more than 100 fMRI papers on speech comprehension or production, I categorized each paper according to the conclusions provided in their abstracts. The first categorization was speech comprehension or speech production. Within each of these main categories, I progressively grouped the papers according to whether their stimuli were prelexical, single words, or sentences . This process il- lustrated the wide range of language topics that are currently being investigated with fMRI. Although all papers published prior to 2009 were excluded, my assumption was that the 2009 papers would have incorporated the most salient conclusions from for- mer years, as well as assimilating their literature. Each of the 2009 papers listed in the references of this review article should therefore provide a list of related papers published prior to 2009. Given the number of language fMRI studies pub- lished in 2009, it was not possible to do justice to the contribution of each one. Furthermore, a review is limited to the reviewer’s own under- standing of the topic and how the conclusions of each paper fit together. What follows is the overall picture that I saw when reviewing the se- lected literature. It emphasizes consistent rather than atypical results. Only time and more studies will determine whether the conclusions drawn are correct. doi: 10.1111/j.1749-6632.2010.05444.x 62 Ann. N.Y. Acad. Sci. 1191 (2010) 62–88 c 2010 New York Academy of Sciences.

Upload: others

Post on 31-Jan-2020

24 views

Category:

Documents


4 download

TRANSCRIPT

Page 1: The anatomy of language: a review of 100 fMRI studies ...qibawiki.rsna.org/images/5/5c/DrMohamed-Price_etal_2010-20100420.pdf · The anatomy of language: a review of 100 fMRI studies

Ann. N.Y. Acad. Sci. ISSN 0077-8923

ANNALS OF THE NEW YORK ACADEMY OF SCIENCESIssue: The Year in Cognitive Neuroscience

The anatomy of language: a review of 100 fMRI studiespublished in 2009

Cathy J. PriceWellcome Trust Centre for Neuroimaging, UCL, London, UK

Address for correspondence: Cathy Price, Wellcome Trust Centre for Neuroimaging, Institute of Neurology, UCL, 12, QueenSquare, London WC1N 3BG, UK. [email protected]

In this review of 100 fMRI studies of speech comprehension and production, published in 2009, activation isreported for: prelexical speech perception in bilateral superior temporal gyri; meaningful speech in middle andinferior temporal cortex; semantic retrieval in the left angular gyrus and pars orbitalis; and sentence comprehensionin bilateral superior temporal sulci. For incomprehensible sentences, activation increases in four inferior frontalregions, posterior planum temporale, and ventral supramarginal gyrus. These effects are associated with the use ofprior knowledge of semantic associations, word sequences, and articulation that predict the content of the sentence.Speech production activates the same set of regions as speech comprehension but in addition, activation is reportedfor: word retrieval in left middle frontal cortex; articulatory planning in the left anterior insula; the initiation andexecution of speech in left putamen, pre-SMA, SMA, and motor cortex; and for suppressing unintended responses inthe anterior cingulate and bilateral head of caudate nuclei. Anatomical and functional connectivity studies are nowrequired to identify the processing pathways that integrate these areas to support language.

Keywords: speech; comprehension; production; fMRI; language

Introduction

The aim of this paper was to provide an up to dateand integrated review of functional imaging studiesof language published in 2009. I included studiesthat (i) used fMRI, (ii) were published online be-tween January and mid-October 2009, and (iii) wereconcerned with the functional anatomy of speechcomprehension and production in the healthy adultbrain. This excluded papers that used MEG, EEG,TMS, or optical imaging; that were published in2008 or before; or that focused on populations withacquired or developmental language disorders. I alsoexcluded papers on orthographic processing of writ-ten words, sign language, and multilingualism: eachof these topics requires its own review.

Having selected more than 100 fMRI papers onspeech comprehension or production, I categorizedeach paper according to the conclusions provided intheir abstracts. The first categorization was speechcomprehension or speech production. Within eachof these main categories, I progressively grouped

the papers according to whether their stimuli wereprelexical, single words, or sentences. This process il-lustrated the wide range of language topics that arecurrently being investigated with fMRI. Although allpapers published prior to 2009 were excluded, myassumption was that the 2009 papers would haveincorporated the most salient conclusions from for-mer years, as well as assimilating their literature.Each of the 2009 papers listed in the references ofthis review article should therefore provide a list ofrelated papers published prior to 2009.

Given the number of language fMRI studies pub-lished in 2009, it was not possible to do justiceto the contribution of each one. Furthermore, areview is limited to the reviewer’s own under-standing of the topic and how the conclusionsof each paper fit together. What follows is theoverall picture that I saw when reviewing the se-lected literature. It emphasizes consistent ratherthan atypical results. Only time and more studieswill determine whether the conclusions drawn arecorrect.

doi: 10.1111/j.1749-6632.2010.05444.x62 Ann. N.Y. Acad. Sci. 1191 (2010) 62–88 c© 2010 New York Academy of Sciences.

Page 2: The anatomy of language: a review of 100 fMRI studies ...qibawiki.rsna.org/images/5/5c/DrMohamed-Price_etal_2010-20100420.pdf · The anatomy of language: a review of 100 fMRI studies

Price The anatomy of language

Anatomical terms

Across studies, different anatomical terms have beenused for the same region. For example, the sameinferior frontal region has been variously referredto as Broca’s area, BA44, or pars opercularis. Theanatomical distinction between anterior and pos-terior superior temporal lobe activation also variesacross studies; and the same left temporo-parietalarea has been referred to as the supramarginal gyrus(SMG), planum temporale, Spt (Sylvian parieto-temporal), superior temporal gyrus (STG), and theinferior parietal lobe (IPL). I have therefore triedto use a more consistent terminology (see Fig. 1),reported the MNI or Talairach coordinates [x,y,z]of the effects I am referring to and estimated thesubregion being referred to using both MNI andTalairch templates. I divided inferior frontal activa-tions into four different regions: pars triangularis(pTr), pars orbitalis (pOr), dorsal pars opercularis(dpOp), and ventral pars opercularis (vpOp). Inthe superior temporal gyrus, anterior and posteriorparts were in relation to Heschl’s gyrus (HG).

Given the limited spatial resolution of fMRI andthe use of different spatial normalization templates,it was not always possible to determine the locationof reported coordinates, particularly when they fellat the border between two regions. For example, thelower bank of the sylvian fissure, posterior to HG isin the posterior planum temporale (pPT), and theupper bank of the same part of the sylvian fissure isin the ventral supramarginal gyrus (vSMG). Theseregions will not have the same connectivity or func-tions but it was difficult to tell when an activationfell in the lower or upper bank of the fissure becauseof their close proximity. One option is to refer to ac-tivation around the posterior part of the Sylvian fis-sure as Spt (Sylvian parieto-temporal), as in Hickokand colleagues.1 However, this de-emphasizes thefact that the lower and upper banks of the sylvianfissure will have different functions. I have there-fore labeled activation around Spt as pPT/vSMg toemphasize the convergence of two different regionswithout implying that they have the same function.

To provide summaries of the results, I markedthe approximate location of the reported coordi-nates on a two-dimensional template of the brain.This further reduced the spatial precision of the re-sults. Remarkably, however, the low-resolution car-toon illustrations were sufficient to highlight dis-

tinct spatial dissociations in the anatomical systemsthat support speech comprehension and produc-tion at the prelexical, word, and sentence levels (seeFig. 2). Future studies that dissociate activationswithin subject, rather than across study, are nowneeded to provide more accurate anatomical local-ization; and functional connectivity analyses withhigh temporal resolution are needed to determinehow the widely distributed sets of regions interactwith one another.

Review of speech comprehension

Auditory speech processing extracts meaningful in-formation from continuously changing acoustic in-puts. Many acoustic features predict the perceptualidentity (intelligibility) of a given auditory stimu-lus including the temporal dynamics of the timbre,pitch, and volume in the auditory waveform andthe integration of this information over short- andlong-time frames. The type of information that isextracted will depend on the expectations and in-tentions of the listener. For example, auditory inputcan be translated into meaningful messages (seman-tics), articulatory associations (speech production),spellings, and physical actions (locomotion) towardor away from the source of the sound (i.e., spatialprocessing).

As auditory speech information is ambiguous,comprehension can be facilitated by auditory cuesfrom prosody and visual cues from concurrent handgestures, lip movements, facial expressions, andwriting. These additional cues are particularly use-ful under adverse listening conditions. For example,co-speech hand gestures provide semantic cues rel-evant to the intended message. Lip reading (view-ing mouth movements) provides articulatory infor-mation and writing facilitates both semantic andarticulatory processing. Therefore, speech compre-hension, even at the single-word level, involves theintegration of multiple cues from auditory and vi-sual sources.

Speech comprehension at the sentence level is alsoambiguous because the meaning of a sentence can bedifferent from its parts, particularly for a metaphor-ical sentence such as “You are the sun in my sky.”Sentence comprehension is therefore constrainedby our prior knowledge of the world (semantic con-straints), what we are expecting to hear (context ef-fects), and how the words are typically combined

Ann. N.Y. Acad. Sci. 1191 (2010) 62–88 c© 2010 New York Academy of Sciences. 63

Page 3: The anatomy of language: a review of 100 fMRI studies ...qibawiki.rsna.org/images/5/5c/DrMohamed-Price_etal_2010-20100420.pdf · The anatomy of language: a review of 100 fMRI studies

The anatomy of language Price

Figure 1. Anatomical terms. The names and location of the regions referred to in the text are illustrated on acartoon sketch of the left hemisphere. This low-resolution sketch was deliberately chosen for all figures because it hadminimal anatomical details, which were not appropriate in the context of the poor spatial precision of the activationloci illustrated in Figure 2, see introductory text for further explanation.

(i.e., syntactic constraints/grammatical rules). Asthe meaning of the sentence emerges over time, au-ditory short-term memory is also required to holdindividual words in memory until a satisfactory

interpretation of the word combination has beenachieved. Each stage in the speech-processing hier-archy therefore involves the integration of bottom-up processing from auditory and visual inputs with

64 Ann. N.Y. Acad. Sci. 1191 (2010) 62–88 c© 2010 New York Academy of Sciences.

Page 4: The anatomy of language: a review of 100 fMRI studies ...qibawiki.rsna.org/images/5/5c/DrMohamed-Price_etal_2010-20100420.pdf · The anatomy of language: a review of 100 fMRI studies

Price The anatomy of language

Figure 2. Summary of left hemisphere activation foci. The approximate location of the left hemisphere activationfoci reported in the text. Right hemisphere foci were less consistent and not summarized here. Despite the poor spatialprecision, the relative distributions of the different types of effects are valid and informative. Top left panel: prelexicalprocessing3–5,7,8; top right panel: pseudowords13; middle left panel: black circles = words relative to pseudowords,13

white circles = words relative to rotated words,9,10 open circles = effect of familiarity with auditory objects14; middle

right panel: black circles = comprehension versus imcomprehensible sentences,19,21,23,24 open circles = sentenceswith versus without visual gestures.27–29 Lower left panel: black circles = semantic ambiguity,9,19,35–39 white cir-cles = syntactic ambiguity10,32,49,50,55; lower right panel: black circles = word retrieval,43,76,79,81–86 open circles = overtarticulation,83,87,96 white circles = covert articulation.72,73,84

Ann. N.Y. Acad. Sci. 1191 (2010) 62–88 c© 2010 New York Academy of Sciences. 65

Page 5: The anatomy of language: a review of 100 fMRI studies ...qibawiki.rsna.org/images/5/5c/DrMohamed-Price_etal_2010-20100420.pdf · The anatomy of language: a review of 100 fMRI studies

The anatomy of language Price

top-down predictions from our prior knowledge oflanguage and the context in which it is being used.

My review of the anatomy of speech comprehen-sion is divided into seven sections. These focus onprelexical auditory processing (prelexical phonemicprocessing); semantic recognition at the single-wordlevel (semantic processing of spoken words); se-mantic recognition at the sentence level (sentencecomprehension); semantic constraints at the sen-tence level (semantic constraints in sentence com-prehension); syntactic constraints at the sentencelevel (syntactic constraints); the role of subvocalarticulation during speech comprehension (subvo-cal articulation during speech comprehension), andthe influence of prosodic information (the role ofprosody in speech comprehension). I then (“Sum-mary of activation related to speech comprehen-sion”) sum up, and (“Left lateralization for speechcomprehension”) hypothesize that top-down pro-cessing has a strong influence on determining leftlateralization for language.

Prelexical phonemic processingPrelexical processing occurs prior to the semanticrecognition of the auditory stimulus. It includesacoustic analysis of the frequency spectrum and theintegration of auditory features over different timeframes. The result is phonemic categorization ofthe speech signal and the morphological parsing ofword stems and inflectional affixes.

Rauschecker and Scott2 review evidence that au-ditory inputs are processed along an antero-lateralgradient that starts at HG, progressing in the ante-rior direction toward the temporal pole. As audi-tory information progresses anteriorly and laterally,lower level auditory features are combined into in-creasingly complex and intelligible sounds. For ex-ample, middle and anterior parts of the superiortemporal sulci respond to prelexical stimuli, suchas consonant–vowel syllables, and there may evenbe subregions selective for particular speech-soundclasses, such as vowels.

Further evidence for vowel selective responses hasbeen presented in two recent studies.3,4 Leff and col-leagues3 compared activation for hearing familiarvowels versus single formants and tones. The for-mants had the same intonation pattern as the vowelsbut they were less acoustically complex because theylacked stop bursts or other formants. Relative to theformants and tones, vowel stimuli increased activa-

tion in lateral left HG, anteriorly [−62 −12 6] andposteriorly [−64 −24 10]. Concurrent activationwas also observed in the right superior temporalsulcus [−64 −10 −8]. Likewise, Britton and col-leagues4 identified activation in lateral and anteriorHG [−64 −7 5/59 −7 5] for hearing vowels versustones. These bilateral superior temporal activationsare likely to reflect differences in the acoustic com-plexity of vowel stimuli versus tone and formantstimuli.

To probe more speech selective responses, Leffand colleagues3 manipulated the acoustic quality oftheir stimuli within category (i.e., vowel, formant, ortone). Using a mismatch paradigm, the same stim-ulus was repeated for a few trials followed by a newstimulus in the same category. In this context, thenew stimulus results in an activation change thatcan be related to the processes that differ betweenthe new stimulus and the prior stimuli. The resultsrevealed a left-lateralized activation in the anteriorsuperior temporal lobe [−56 6 −10] when the vowelchanged and this effect was not observed for formantor tone changes. Myers and colleagues5 also used amismatch paradigm, repeating a phonemic stimu-lus and measuring activation in response to acousticdeviants. They found left dorsal pars opercularis ac-tivation at [−41 8 27] when the phonetic categoryof the new stimulus was different versus the same asits predecessors. These mismatch responses3,5 canbe explained in terms of “prediction error” whentop-down expectations from prior experience failto predict the bottom-up inputs.6 According to thisaccount, left anterior temporal activation for voweldeviants3 and left dorsal pars opercularis activationfor phonetic deviants5 reflect the influence of top-down predictions, with the location of the mismatchresponse depending on the type of stimulus and thetype of prediction.

Specht and colleagues7 also investigated speechselective responses by manipulating acoustic com-plexity within category. Using a sound morphingtechnique, white noise was gradually changed intoeither prelexical speech (consonant-vowels) or mu-sic (piano or guitar chords). Bilateral activation inthe anterior–superior temporal sulci at [−51 −3−9/60 −6 −3] was observed irrespective of whetherspeech or music emerged but left lateralized speechselective responses were identified more posteri-orly [−56 −36 12] and ventrally [−54 −18 −6].Likewise, Vaden and colleagues8 report an effect

66 Ann. N.Y. Acad. Sci. 1191 (2010) 62–88 c© 2010 New York Academy of Sciences.

Page 6: The anatomy of language: a review of 100 fMRI studies ...qibawiki.rsna.org/images/5/5c/DrMohamed-Price_etal_2010-20100420.pdf · The anatomy of language: a review of 100 fMRI studies

Price The anatomy of language

of phonological priming in superior temporal re-gions [−63 −30 3/45 −33 −3] that were posteriorand ventral to HG. Together these results illustratespeech selective responses in the posterior and ven-tral processing directions.

To summarize, prelexical processing of auditoryspeech increased activation in bilateral superiortemporal gyri but left lateralized responses were ob-served when the noise merged into speech versusmusic7 or when there was a mismatch between thestimulus presented and the stimulus expected.3,4

In both these contexts, left lateralization may bedriven by top-down predictions from prior experi-ence rather than bottom-up processing. For exam-ple, the emergence of speech over time allows theparticipant to predict the morphing of the stimulus;and mismatch responses have been associated withthe failure of top-down processing to predict thestimulus.6

The location of all left hemisphere coordinatesreported above for prelexical speech versus non-speech processing are illustrated in the top leftpanel of Figure 2. This illustrates that prelexicalactivation is most extensive in the direction ante-rior to HG. However, the results do not supporta straightforward gradient of increased speech se-lectivity as information progresses more anteriorly.Speech selectivity was observed in the ventral7,8

and posterior7 directions as well as the ante-rior direction.3 This suggests that there are mul-tiple prelexical processing pathways, with the levelof activation in each depending on the stimu-lus, task and top-down expectations from priorexperience.

Semantic processing of spoken wordsSemantic processing of auditory speech links heardsounds to what we know about their meanings. Thishas been probed by comparing familiar words to (i)closely matched but unintelligible spectrally rotatedwords9,10 and (ii) pseudowords that are intelligi-ble but without meaning.11–13 A study of auditoryfamiliarity, when the acoustic stimulus is held con-stant,14 is also reported in this section.

When participants listen to spoken sentences, rel-ative to spectrally rotated speech (which is unintel-ligible), bilateral mid-to-anterior superior temporalactivation has been reported by Friederici and col-leagues10 at [−58 −4 4/62 −4 −14] and by Obleserand Kotz9 at [−60 −8 −6/62 −6 −4] with the latter

authors9 also reporting activation in the left angu-lar gyrus [−48 −64 38]. It is important to note,however, that the bilateral mid-to-anterior superiortemporal activations observed for familiar wordsrelative to spectrally rotated speech are in very closeproximity to those reported as common to musicand prelexical speech by Specht and colleagues7

at [−51 −3 −9/60 −6 −3]. It may therefore bethe case that the activation reported for the com-parison of speech to spectrally rotated speech9,10

reflects prelexical differences between speech androtated speech rather than differences in semanticprocessing. In contrast, semantic retrieval involvesmore distributed regions. For example, Sharp andcolleagues15 have shown that when attention is di-rected to the meaning of single words versus theacoustic properties of spectrally rotated speech, ac-tivation is observed in a left-lateralized network in-cluding regions in the inferior temporal gyrus, an-terior fusiform, hippocampus, angular gyrus, parsorbitalis, superior and middle frontal gyri, and theright cerebellum.

The comparison of spoken words to pseudowordsprovides a tighter control for prelexical processingand speech intelligibility than the comparison ofspoken words to spectrally rotated speech. I foundthree studies11–13 published on line in 2009 thatcompared spoken words and pseudowords in thecontext of deciding if the stimulus was a word ornot (i.e., a lexical decision task). In a review of priorstudies, Davis and Gaskell13 report an impressivedouble dissociation in activation for familiar wordsversus pseudowords with activation for words dis-tributed in bilateral anterior middle temporal cor-tices, posterior temporal parietal cortices, and theprecuneus, with left-lateralized activation in thetemporal pole, posterior middle temporal cortex,anterior fusiform, pars orbitalis, middle frontal cor-tex, anterior cingulate and putamen, and the rightprecentral gyrus. The left hemisphere foci of theseresults are illustrated in the left middle panel ofFig. 2). This highlights word activations that are out-side the areas associated with prelexical processing,extending further in the anterior, ventral and pos-terior directions, consistent with the results of theprelexical studies discussed in section “Pre-lexicalphonemic processing” that suggest multiple speech-processing routes.

The meta-analysis of auditory words versus pseu-dowords reported by Davis and Gaskell13 also

Ann. N.Y. Acad. Sci. 1191 (2010) 62–88 c© 2010 New York Academy of Sciences. 67

Page 7: The anatomy of language: a review of 100 fMRI studies ...qibawiki.rsna.org/images/5/5c/DrMohamed-Price_etal_2010-20100420.pdf · The anatomy of language: a review of 100 fMRI studies

The anatomy of language Price

demonstrated increased activation for pseudowordsrelative to words within the superior temporal areasassociated with prelexical processing in the previ-ous section (compare top left and top right panelsin Fig. 2). Together the results suggest that pseu-dowords increased the demands on prelexical pro-cessing whereas real words increased activation insemantic areas. Indeed, many of the areas associatedwith word versus pseudoword processing13 have alsobeen associated with semantic processing of writtenwords and pictures of objects, as summarized in areview of 120 functional imaging studies by Binderand colleagues.16 Diaz and McCarthy17 have alsoreported greater anterior middle temporal lobe ac-tivation bilaterally at [−59 −14 −27/66 −10 −21]for written content words (with high semantic asso-ciations) versus written function words (that carryminimal or ambiguous semantic associations).

The second study that compared spoken wordsand pseudowords during lexical decision was re-ported by Kotz and colleagues.12 Each target stim-ulus (word or pseudoword) was primed with arhyming or nonrhyming word or pseudoword.Rhyming pseudowords compared to rhyming wordsincreased activation in bilateral superior temporalgyri as reported by Davis and Gaskell,13 whereasrhyming words compared to rhyming pseudowordsincreased activation in frontal and parietal regions.Kotz and colleagues12 focus their interpretation onthe observation that left inferior frontal activationat the level of the left pars orbitalis/pars triangularis[−56 28 6] was greater for words than pseudowords.

The third study to compare lexical decisions towords and pseudowords was reported by Kouiderand colleagues11 who report higher activation forwords than pseudowords in the precuneus, medialsuperior frontal gyrus, and the anterior cingulate.Although Kouider and colleagues11 do not presentany coordinates for these effects, it is interesting tonote that the same regions were also reported earlierfor words versus pseudowords13 and for phonologi-cal repetition of words.8 Activation in the precuneusand superior frontal gyrus have also been associatedwith semantic decisions on heard words,15,16 nar-ratives,18 metaphors,19 and semantic versus phono-logical word generation.20 In brief, activation forspoken words relative to pseudowords during lex-ical decision was identified in frontal and parietalareas that are also activated during semantic deci-sion tasks.

Finally, an alternative approach for identifyingsemantic recognition of auditory stimuli is to keepthe stimulus (and prelexical processing) constantand compare responses to the same stimuli beforeand after learning. Leech and colleagues14 used avideo game to train participants to associate novelacoustically complex, artificial nonlinguistic soundsto visually presented aliens. After training, left pos-terior superior temporal activation at [−54−37−1]increased with how well the auditory categories rep-resenting each alien had been learnt. The authorsemphasize that part of what makes speech special isthe extended experience that we have with it. Thiscould include acoustic familiarity, enhanced audio–visual associations, or auditory memory.

To summarize this section, semantic processingof familiar auditory stimuli activates a distributedset of regions that surround the ventral, anterior,and posterior borders of the perisylvian regionssupporting prelexical auditory speech processing(see middle left panel vs. top left panel in Fig. 2).The extended distribution of semantic activationssuggests that there are numerous pathways sup-porting speech perception and comprehension (seeRauschecker and Scott2). This might be envisionedas activity spreading out and around a spider’s web-like network, centered on HG.

Sentence comprehensionTo tap speech comprehension at the sentence level,activation has been compared for hearing grammat-ically correct sentences with plausible versus im-plausible meanings. This comparison is expectedto reveal semantic representations at the level ofword combinations while controlling for phono-logical and lexical familiarity, syntactic processing,and working memory. Activation in four key regionshas been reported: anterior and posterior parts ofthe left middle temporal gyrus,9,19,21 bilateral ante-rior temporal poles,21,22 left angular gyrus,9,19 andthe posterior cingulate/precuneus.18,19 Each of theseareas was also associated with semantic processingin the review of 120 studies reported by Binder andcolleagues.16 Surprisingly, however, activation forplausible relative to implausible sentences is not asextensive as that associated with semantic process-ing of single words (see left and right middle panelsin Fig. 2) but was primarily focused in anterior andposterior parts of the middle temporal gyrus, justventral to the superior temporal sulci.

68 Ann. N.Y. Acad. Sci. 1191 (2010) 62–88 c© 2010 New York Academy of Sciences.

Page 8: The anatomy of language: a review of 100 fMRI studies ...qibawiki.rsna.org/images/5/5c/DrMohamed-Price_etal_2010-20100420.pdf · The anatomy of language: a review of 100 fMRI studies

Price The anatomy of language

In the left anterior middle temporal locus, activa-tion has been reported by Mashal and colleagues19

at [−46 −6 −10] for semantically plausible morethan implausible sentences; by Adank and Devlin23

at [−58 −8 −4] when time compressed auditorysentences were understood versus not understoodduring a sentence verification task; by Devauchelleand colleagues24 at [−60 −12 −8/60 −8 −12] whenauditory sentence processing was primed by a sen-tence with the same meaning; by Obleser and Kotz9

at [−52 −6 −14] for sentences with meanings thatwere difficult versus easy to predict; and by Hubbardand colleagues25 at [−57 −12 −8] when spokenspeech was accompanied by beat gestures (rhyth-mic beating of hand) that enhance semantic mean-ing by providing intonation. These sentence-levelcomprehension effects are not specific to auditorywords because anterior middle temporal activationhas also been reported by Snijders and colleagues26

at [−56 −6 −16] for written sentences comparedto unrelated word sequences; and by Kircher andcolleagues27 at [−52 4 −8] when spoken sentenceswith abstract meanings were accompanied by visualgestures relative to speech only or gesture only. Thissuggests that the left anterior superior temporal sul-cus is involved in multimodal sentence processing.

Likewise, left posterior middle temporal activa-tion has been reported at [−54 −41 −1] by Mashaland colleagues,19 for semantically plausible versusimplausible sentences; more anteriorly at [−41 −310] by Rogalsky and Hickok21 when meaningful sen-tences were compared to lists of unrelated words;and more posteriorly at [−55 −52 2] by Adankand Devlin23 when time compressed auditory sen-tences were understood versus not understood dur-ing a sentence verification task. Activation in the leftposterior middle temporal cortex is also increasedwhen auditory sentences are accompanied by visualobservation of the speaker’s body movements. Forexample, Holle and colleagues28 reported activationat [−48 −39 3] when participants viewed a speakerdescribing manual object actions (e.g., “now I gratethe cheese”) while making meaningful manual ges-tures versus hearing speech only or seeing gesturesalone. This effect was stronger in adverse listeningconditions that benefit most from visual cues. Like-wise, Kircher and colleagues27 found activation at[−63 −43 −1] when spoken sentences with abstractmeanings were accompanied by visual cues fromgestures versus hearing speech only or seeing ges-

tures only; and Robins and colleagues29 found acti-vation at [−66 −31 4] when participants viewed theface and emotional expression of a person speakingsentences versus hearing the sentence or viewing theface alone. Seeing the lip movements of the speakeralso provides phonological cues,30 therefore it is notsurprising that the activation reported by Robinsand colleagues29 for speech and faces extended dor-sally [−45 −37 10] into the superior temporalgyrus.

The findings of Dick and colleagues31 further sug-gest that left posterior temporal activation in re-sponse to speech (spoken stories) and gesture re-flects two sources of semantic information ratherthan an active integration process. Specifically, theyfound that perceiving hand movements duringspeech increased activation in the left posteriortemporal regions irrespective of the semantic re-lationship between the hand movement and the ac-companying speech. Thus, it did not depend onwhether speech and gesture could be integratedinto the same meaning or not. Instead, the au-thors conclude that listeners attempt to find mean-ing, not only in the words that speakers produce,but also in the hand movements that accompanyspeech.

The role of the left and right temporal polesin sentence processing relative to unrelated lists ofwords has been reported by Rogalsky and Hickok21

at [−47 17 −18 and 52 18 −20]. Snijders and col-leagues26 also found bilateral temporal pole regionsat [−53 18 −30/54 20 −32] for written sentencescompared to unrelated word sequences but the co-ordinates are more ventral than those reported byRogalsky and Hickok.21 To distinguish auditory andvisual sentence processing requires a within subjectscomparison of (i) auditory sentences with plau-sible versus implausible meanings with (ii) visualsentences with plausible versus implausible mean-ings. All four of these conditions were included ina study reported by Richardson and colleagues,32

which found no interaction between plausible ver-sus implausible sentences presented in the auditoryversus visual modality. Likewise, a meta-analysisof semantic and sentence processing activation byVisser and colleagues22 found anterior temporallobe activation for both auditory and visual sen-tences and these authors discuss the role for theanterior temporal cortex in amodal combinatorialsemantics.

Ann. N.Y. Acad. Sci. 1191 (2010) 62–88 c© 2010 New York Academy of Sciences. 69

Page 9: The anatomy of language: a review of 100 fMRI studies ...qibawiki.rsna.org/images/5/5c/DrMohamed-Price_etal_2010-20100420.pdf · The anatomy of language: a review of 100 fMRI studies

The anatomy of language Price

Left angular gyrus activation has been less consis-tently reported during sentence processing but wasobserved by Mashal and colleagues19 at [−47 −5925] for sentences with plausible versus implausiblemeanings; and by Obleser and Kotz9 at [−46 −6438] for the comparison of heard sentences to un-intelligible spectrally rotated speech. Obleser andKotz9 suggest that activation in the angular gyrus fa-cilitates sentence comprehension via top-down ac-tivation of semantic concepts. Likewise, Carreirasand colleagues33 demonstrated a top-down rolefor the angular gyrus at [−48 −74 28] duringreading relative to object naming, and Brownsettand Wise34 highlight the role of the left angulargyrus in both speaking and writing. In the medialparietal lobes, Whitney and colleagues18 highlightthe importance of the right precuneus and bilat-eral posterior/middle cingulate cortices for narra-tive language comprehension, associating these re-gions with the processes involved in updating storyrepresentations.

In summary, the comparison of grammaticallycorrect sentences with comprehensible versus in-comprehensible meanings has been associated withactivation in anterior and posterior parts of the leftmiddle temporal gyrus,9,19,21 bilateral anterior tem-poral poles,21,22 the left angular gyrus,9,19 and theposterior cingulate/precuneus.18,19 Each region mayplay a different role in sentence semantics. For ex-ample, seeing concurrent hand gestures enhancedactivation for heard sentences in the posterior mid-dle temporal cortex consistent with auditory andvisual semantics converging at the posterior end ofthe semantic network. Inconsistent activation in theleft angular gyrus and medial parietal regions mayreflect task-dependent semantic retrieval strategies;and activation in anterior middle temporal lobe mayreflect the integration of multiple semantic con-cepts.

When the sentence comprehension activationsare illustrated on the cartoon brains in Figure 2,it is striking to see that temporal lobe activation forplausible versus implausible sentences (middle rightpanel) lies outside the areas associated with prelexi-cal processing (top left panel) or pseudowords (topright panel) but inside the ring of activation as-sociated with semantic processing of single words(middle left panel). This dissociation might reflecta task confound (e.g., lexical decision for words andplausibility judgments for sentences). Alternatively,

it raises the interesting possibility that activationin the superior temporal sulci for sentences withplausible versus implausible meanings reflects theconsolidation of multiple semantic concepts into anew and integrated concept. For example, the sen-tence “My memory is a little cloudy” has six wordsincluding three rich content words (memory, little,cloudy) to express one concept “forgotten.” Thissuggests that activation for accessing the semanticsof single words is spatially extensive but consoli-dates in the ventral banks of the superior temporalsulci when multiple meanings can be integrated intofewer concepts.

Semantic constraints in sentencecomprehensionThe previous section focused on activation forgrammatically correct sentences with plausible ver-sus implausible meanings. In contrast, this sectionconsiders the reverse of this comparison, that is,which areas are more activated by sentences withimplausible versus plausible meanings. The assump-tion here is that, when the sentence meaning isimplausible, ambiguous, or unconstrained, therewill be a greater need for semantic constraints thatwill be expressed by top-down processing from ourprior knowledge of the world and the experimentalcontext.

When sentence comprehension becomes moredifficult, activation is consistently observed in theleft pars opercularis. I am distinguishing dorsal ver-sus ventral pars opercularis according to whetheractivation is more or less than 20 mm above theACPC line. Dorsal pars opercularis activation hasbeen reported for grammatically correct sentenceswith implausible versus plausible meanings by Yeand Zhou35 at [−40 22 24] and in several studieswhen the meaning of the sentence is more difficultto extract. For example, by Bilenko and colleagues36

at [−52 14 23/47 14 29] for sentences with ambigu-ous versus unambiguous meanings; by Mashal andcolleagues19 at [−44 18 30] for sentences with novelmetaphoric versus literal meanings; by Willems andcolleagues37 at [−40 10 22] when speech was pre-sented with incongruent relative to congruent ges-tures or pantomimes (i.e., when there was interfer-ence at the comprehension level); and by Desai andcolleagues38 at [−54 10 31] when participants lis-tened to sentences with abstract versus visual or mo-toric meanings. However, left dorsal pars opercularis

70 Ann. N.Y. Acad. Sci. 1191 (2010) 62–88 c© 2010 New York Academy of Sciences.

Page 10: The anatomy of language: a review of 100 fMRI studies ...qibawiki.rsna.org/images/5/5c/DrMohamed-Price_etal_2010-20100420.pdf · The anatomy of language: a review of 100 fMRI studies

Price The anatomy of language

activation is not specific to semantic contexts. It hasalso been reported by Myers and colleagues5 at [−418 27] for prelexical stimuli when there is an unex-pected change in phonetic category and in otherstudies of linguistic and nonlinguistic sequencing(see next section).

More ventrally, pars opercularis activation hasbeen reported by Tyler and colleagues39 at [−5115 18] for grammatically correct sentences with im-plausible versus plausible meanings; by Obleser andKotz9 at [−60 12 16] for sentences with low versushigh semantic expectancy; by Desai and colleagues38

at [−51 19, 13] when participants listened to sen-tences with abstract versus visual or motoric mean-ings; by Turner and colleagues40 at [−50 14 18]when participants view silent videos of a speakermaking speech versus non-speech oral movements;and by Szycik and colleagues41 at [−49 14 18] whenheard speech is inconsistent with seen oral move-ments. These activations may reflect top-down pro-cessing that is attempting to predict the stimuli. Forexample, sentences with low semantic expectancy(e.g., “She weighs the flour”) are difficult to predictbefore completion of the sentence because a verbsuch as “weighed” can be associated with multi-ple objects; whereas the sentences with high seman-tic expectancy (e.g., “She sifts the flour”) are eas-ier to predict before sentence completion becausethere are not many nouns that are likely to followverbs like “sift.” In this context, activation in theleft ventral pars opercularis might reflect increaseddemands on predictions about the forthcoming se-quence of events. The examples in this section havepertained to semantic predictions. However, leftventral pars opercularis activation has also been as-sociated with verbal working memory at [−53 7 15]by Koelsch and colleagues42; and for articulatoryplanning. Specifically, Papoutsi and colleagues43 re-port activation at [−54 12 12] for the repetition andsubvocal rehearsal of pseudowords with low- ver-sus high-sublexical frequency; a manipulation thatdoes not involve semantic, syntactic, or phonolog-ical word form processing. Left ventral pars op-ercularis activation can therefore be consistentlyexplained in terms of the generation of semanticor articulatory sequences. Its function is consideredagain in sections “Articulation” and “Summary ofspeech production.”

The comparison of auditory sentences with plau-sible versus implausible meanings was also associ-

ated with left pars orbitalis, activation at [−39 27−6/−48 24 −9] by Tyler and colleagues39; and at[−52 32 −4] by Ye and Zhou.35 The same regionhas been reported by Davis and Gaskell13 at [−4027 −9/−48 24 −12] for single auditory words rela-tive to pseudowords; by Nosarti and colleagues44 at[−38 32 −6] for reading irregularly spelled wordsrelative to pseudowords; by Schafer and Constable45

at [−42 28 −11] for semantic relative to syntacticprocessing in written speech comprehension; andby Aarts and colleagues46 at [−40 22 −14] whenthe written words “right” or “left” are incongruentwith the direction of an arrow (right or left). Allthese studies suggest that left pars orbitalis activateswith increasing demands on semantic retrieval inthe context of semantic conflict that may be aris-ing at the single-word level or between the mean-ing of single words and sentences. Likewise, rightinferior frontal activation has also been reportedin the context of conflicting semantic information.For example, activation in the right pars opercu-laris [44 18 14/54 18 12] and triangularis [46 286] were reported by Snijders and colleagues26 whenthe meaning of a sentence was ambiguous versusunambiguous; and by Peelle and colleagues47 at [4414 22/44 4 28] when the meanings of a series ofsentences conflict with one another. Dick and col-leagues31 also report right inferior frontal activa-tion when participants listened to and watched astory teller using hand movements that were se-mantically incongruent relative to congruent withthe spoken speech. As with left inferior frontalactivation, right inferior frontal activation is notspecific to the auditory modality. It has also beenreported at [60 28 −4] by Schmidt and Seger48

when subjects read metaphors relative to literal sen-tences. Right inferior frontal activation may there-fore increase when there is incongruency betweenthe meaning of the words and the meaning of thesentence.

In summary, when the semantic content of sen-tences is difficult to extract, activation increases inthe left and right pars opercularis and orbitalis.This contrasts to the temporal and parietal regionsthat are more activated for sentences with plausi-ble meanings. The most likely explanation is thatactivation in the pars orbitalis reflects semanticcompetition at the single-word level and activationin the pars opercularis reflects top-down predic-tions on the plausible sequence of events. This is

Ann. N.Y. Acad. Sci. 1191 (2010) 62–88 c© 2010 New York Academy of Sciences. 71

Page 11: The anatomy of language: a review of 100 fMRI studies ...qibawiki.rsna.org/images/5/5c/DrMohamed-Price_etal_2010-20100420.pdf · The anatomy of language: a review of 100 fMRI studies

The anatomy of language Price

discussed in more detail in the following sectionsthat also attempt to clarify a distinction between thefunction of dorsal and ventral regions of the parsopercularis.

Syntactic constraintsSyntactic processing refers to the hierarchical se-quencing of words and their meanings with the ex-pected order of words depending on the languagespoken. Consequently, there are language-specificgrammatical rules about adjacent words (what typeof word is likely to follow another) and long-distanceword associations. Deriving the meaning of wordsbecomes particularly complex when one phrase isembedded within another. For example, a sentencesuch as “the man the dog chased was carrying aball” has one relative clause (“the dog chased”)nested within a simpler sentence “the man was car-rying the ball.” As the sentence becomes more com-plex, the demands on auditory short-term memoryincrease because items have to be held in mem-ory for longer until the meaning of the sentenceemerges.

Syntactic processing has been investigated bycomparing sentences with grammatical errors tosentences without grammatical errors; and for sen-tences with more versus less syntactically complexstructures. In both cases, the demands on syntacticprocessing are confounded by the differing demandson semantics because both grammatical errors andcomplex sentences make it more difficult to extractthe meaning of the sentence. It is therefore not sur-prising that the left pars opercularis areas that weremore activated for sentences with implausible ver-sus plausible meanings in the previous section arealso more activated for sentences with grammaticalerrors or complex structures. For example, left ven-tral pars opercularis activation has been reportedat [−62 18 12] by Friederici and colleagues10 whensentences had syntactic errors; and at [−47 12 18]by Raettig and colleagues49 when there were viola-tions in verb–argument structure. As discussed inthe previous section on semantic constraints, acti-vation in the left ventral pars opercularis has alsobeen associated with verbal working memory [−537 15] and predicting the sequence of semantic or ar-ticulatory events. It will be discussed again in section“Summary of speech production.”

Syntactic complexity has also been shown to in-crease auditory sentence activation in the left dorsal

pars opercularis at [−45 6 24], even when workingmemory demands are factored out.50 An effect ofsyntactic complexity at [−48 16 30] has also been re-ported by Newman and colleagues51 for written sen-tences; and activation at [−46 16 24] during writtensentences was more strongly primed by prior pre-sentation of the verbs than the nouns.52 Makuuchiand colleagues50 attributed dorsal left pars oper-cularis activation to the hierarchical organizationof sequentially occurring events. Hierarchical se-quencing of events is a core component of syntacticprocessing but it is not limited to linguistic stimuli.It can therefore be investigated using nonlinguisticstimuli and such studies have again highlighted theinvolvement of the left dorsal pars opercularis.53,54

For example, Bahlmann and colleagues53 found sig-nificantly higher activation in the dorsal pars oper-cularis at [−50 6 30] when sequences of nonlinguis-tic visual symbols could be predicted on the basis ofnonadjacent word dependencies compared to adja-cent word dependencies. Similarly, Tettamanti andcolleagues54 found dorsal left pars opercularis acti-vation at [−54 12 24] when participants learnt thenonrigid (variable) dependencies of items within asequence of unfamiliar colored shapes. The dorsalleft pars opercularis therefore appears to be involvedin sequencing events, irrespective of whether theyare linguistic or nonlinguistic.

The third region that has been associated withsyntactic errors or complexity lies on the border be-tween the pPT and the vSMG. Specifically, Raettigand colleagues49 reported activation at [−65 −4215] for sentences with grammatical errors (e.g., “Pe-ter has eat the apple” vs. “Peter has eaten the ap-ple”); Friederici and colleagues55 report activationat [−63 −42 21] for syntactically complex versusless complex sentences; and Richardson and col-leagues32 report increased activation at [−58 −4222] for sentences where the meaning depends onthe order of the subject and object (e.g., “The dogchased the horse”/“The horse chased the dog”) ver-sus sentences where the subject and object are notreversible (e.g., “The dog eats the bone”).

Activation in the vicinity of pPT and/or vSMGis not limited to difficult syntactic contexts. Theseareas are also activated when sentence compre-hension is made more challenging at the percep-tual or semantic level. In perceptually challengingcontexts, Dos Santos Sequeira and colleagues56 re-port activation at [−56 −42 24] when participants

72 Ann. N.Y. Acad. Sci. 1191 (2010) 62–88 c© 2010 New York Academy of Sciences.

Page 12: The anatomy of language: a review of 100 fMRI studies ...qibawiki.rsna.org/images/5/5c/DrMohamed-Price_etal_2010-20100420.pdf · The anatomy of language: a review of 100 fMRI studies

Price The anatomy of language

listened to consonant-vowel syllables in backgroundnoise. In semantically challenging contexts, Obleserand Kotz9 report activation at [−46 −36 18] forsentences with low versus high semantic expectancyparticularly when the stimuli had low intelligibility;and Hickok and colleagues1 report pPT/vSMG ac-tivation for individual subjects at [x = −47 to −61;y = −32 to −50; and z = 11 to 28] for hearingsentences that were syntactically correct but seman-tically meaningless because some words in the sen-tence had been replaced with pseudowords (“It isthe glandor in my nedderop”).

Why does left pPT/vSMG activate during difficultspeech comprehension conditions? Evidence fromstudies of auditory short-term memory42,43,57 andspeech production in the absence of auditory stim-uli32,58 suggests that it is involved in subvocal ar-ticulation. For example, Papoutsi and colleagues43

found increased left pPT/vSMG activation bilater-ally at [−56 −38 20/64 −32 10] when subjects ar-ticulated four versus two syllables during a task thatinvolved delayed repetition and subvocal rehearsalof pseudowords. Left pPT or vSMG activation hasalso been observed during auditory working mem-ory at [−44−38 21] by Koelsch and colleagues42 andat [−63 −34 19] by Buchsbaum and D’Esposito.57

One interpretation is that left pPT/vSMG activa-tion reflects a working memory rehearsal strategythat holds an auditory representation in memory.However, in Buchsbaum and D’Esposito,57 the taskdid not involve the online retention of a fixed setof items. Instead, left pPT/vSMG activation was ob-served when participants made correct word recog-nition decisions on constantly changing targets thatmay or may not have been presented before. Buchs-baum and D’Esposito57 therefore suggest that leftpPT/vSMG activation is involved in phonologicalretrieval processes (perhaps subvocal articulation)that occurs automatically during successful recog-nition of recently presented words.

Other studies have shown that left pPT/vSMG ac-tivation is not limited to speech production in thecontext of auditory stimuli or short-term memorytasks. Nor is it limited to the production of speechsounds. For example, Richardson and colleagues32

report activation at [−62 −40 14/50 −46 20] foralternatively articulating “one” and “three” in re-sponse to seeing (not hearing) the digits 1 and 3, andalso for nonverbal mouth movements versus handmovements; and Wilson and colleagues58 report ac-

tivation at [−64 −34 24] for naming pictures withlow- versus high-frequency names. For non-speechsounds, Koelsch and colleagues42 report activationat [−47 −42 24] when a sequence of auditory tonesis held in memory; and Yoncheva and colleagues59

report activation at [−42 −38 26] for a tone match-ing task (are tone-triplet pairs the same or not) aswell as for a rhyme-matching task (do word pairsrhyme or not) with no significant difference in ac-tivation for the rhyme versus tone comparisons.

The correspondence in the location of speechand non-speech effects has been demonstrated byKoelsh and colleagues42 for auditory working mem-ory tasks; by Yoncheva and colleagues59 duringmatching tasks; and by Richardson and colleagues32

for auditory speech comprehension and visuallycued mouth movements. Common activation fortwo tasks may or may not reflect a common underly-ing function. Indeed, Hickok and colleagues1 disso-ciated the spatial pattern of activation in pPT/vSMGfor speech comprehension and production, usingmultivariate pattern classification. This approachidentifies a significant difference in the pattern ofvoxels over a single region but it does not dissociatethe functional contribution of the pPT versus vSMGregions.

To conclude this section, activation for syntacti-cally difficult sentences increases in the left pars op-ercularis and the left pPT/vSMG.55 The same regionsare also activated when syntactic demands are heldconstant and sentence comprehension is made moredifficult. Therefore, activation in these regions is notspecific to syntactic processing. The dorsal pars op-ercularis appears to play a top-down role in sequenc-ing linguistic and nonlinguistic events, independentof working memory demands.50 The ventral parsopercularis is involved in verbal working memory42

and sequencing semantic and articulatory events.In contrast, activation in the pPT and vSMG can beexplained by subvocal articulation. The functionaldistinctions between these regions will be discussedfurther in section “Summary of speech production”but future studies will be required to dissociate thefunctions of pPT and vSMG.

Subvocal articulation during speechcomprehensionThe previous section concluded that left pPT/vSMGactivation during difficult speech comprehensionreflected subvocal articulation. The close link

Ann. N.Y. Acad. Sci. 1191 (2010) 62–88 c© 2010 New York Academy of Sciences. 73

Page 13: The anatomy of language: a review of 100 fMRI studies ...qibawiki.rsna.org/images/5/5c/DrMohamed-Price_etal_2010-20100420.pdf · The anatomy of language: a review of 100 fMRI studies

The anatomy of language Price

between comprehension and articulation is one ofthe unique features of speech because we can ac-curately reproduce speech sounds but we cannotaccurately reproduce non-speech sounds. This isnot surprising given that auditory speech, by def-inition, is constrained by the auditory signals thatour vocal tracts and articulators can produce. Evi-dence that heard speech activates areas involved inarticulating speech comes from common activationfor (i) auditory speech in the absence of articula-tion and (ii) articulation in the absence of audi-tory speech. Three areas meet these criteria. One isthe motor cortex,60,61 the others are the pPT andvSMG regions discussed in the previous section. Inall three of these regions, activation is consistentlyactivated during speech production but inconsis-tently activated during speech comprehension. Itwas quite striking that the majority of auditoryspeech perception and comprehension studies thatI reviewed did not report left pPT or vSMG activa-tion, irrespective of whether the stimuli were sen-tences,10,23,24,38 words,8,11,12,62–64 or prelexical con-sonant vowels.3,7 I only found reports of left pPTand/or vSMG activation in five studies of sentencecomprehension1,9,32,49,55 and one study of prelexicalperception.56 In all cases, pPT/vSMG activation wasidentified when speech comprehension was mademore difficult at the syntactic, semantic, or percep-tual level (see previous section). Thus, fMRI evi-dence suggests that an area associated with speechproduction is not consistently involved in speechcomprehension (when acoustic inputs are matched)but may play an essential role when speech compre-hension is difficult.

A similar conclusion has been reached for the roleof the motor cortex in speech perception. This is dis-cussed in two recent review papers.60,61 Devlin andAydelott60 discuss the degree to which functionalimaging data support the motor theory of speechperception. They emphasize that motor and premo-tor activation is not consistently reported in speechperception studies but it is observed when speechstimuli are degraded (e.g., embedded in noise) orminimal (e.g., syllables rather than words) and maytherefore be recruited to aid speech comprehen-sion, when speech perception is challenging. Scottand colleagues61 also conclude that motor activa-tion might facilitate speech perception in difficultlistening conditions. In addition, they suggest thatmotor activation during speech perception allows

people to coordinate their speech with others, bothin terms of turn taking and also in terms of id-iosyncratic characteristics of pronunciation and theuse of conceptual and syntactic structures. The mainpoint, however, in both reviews60,61 is that activationin motor cortex is not always necessary for phoneticencoding or speech comprehension.

To conclude this section, left motor and left pPTand left vSMG activations have been reported fortasks that challenge auditory comprehension andeach of these regions are also activated for speechproduction in the absence of auditory cues. Ac-tivation may therefore reflect subvocal articula-tion that facilitates difficult speech comprehension.More specifically, the overlap in the location of ac-tivation for speech production and difficult speechperception may reflect the use of the speech pro-duction system to make predictions during speechperception.

The role of prosody in speech comprehensionProsody refers to the patterns of stress and intona-tion in speech. These patterns carry emotional andnonverbal information that supplements the lexico-semantic information carried by the spoken words.Wiethoff and colleagues65 found that emotionalprosody increases activation in the amygdala. Thisis consistent with previous studies showing amyg-dala activation for facial expressions,66 and whensubjects read emotional words.67 Wittfoth and col-leagues68 found that when the emotional prosodyin heard sentences was incongruous with semanticcontent, activation increased in the right superiortemporal gyrus and sulcus and right dorsal anteriorcingulate cortex (ACC). Many previous studies haveassociated the right superior temporal sulcus withnonverbal emotional processing of human voicesand faces but Kreifelts and colleagues69 have recentlydemonstrated a functional subdivision of the supe-rior temporal lobes. They found maximum voicesensitivity in the trunk of the superior temporallobe and maximum face sensitivity in the posteriorterminal ascending branch. They also argue that anoverlap of these two regions at the bifurcation of thesuperior temporal cortex may support the forma-tion of an audiovisual emotional percept.

Although the earlier studies have investigated theimpact of emotional prosody in general, Ethoferand colleagues70 attempted to dissociate activationfor pseudowords spoken in five prosodic categories

74 Ann. N.Y. Acad. Sci. 1191 (2010) 62–88 c© 2010 New York Academy of Sciences.

Page 14: The anatomy of language: a review of 100 fMRI studies ...qibawiki.rsna.org/images/5/5c/DrMohamed-Price_etal_2010-20100420.pdf · The anatomy of language: a review of 100 fMRI studies

Price The anatomy of language

(anger, sadness, neutral, relief, and joy). Using mul-tivariate pattern analysis, they demonstrated thateach emotion had a specific spatial signature in theauditory cortex that generalized across speakers.

Summary of activation related to speechcomprehensionIn summary, the current speech comprehension lit-erature supports the following anatomical model.Acoustic processing of prelexical speech increasesactivation in the vicinity of bilateral HG with se-lectivity for speech emerging in lateral, anterior,ventral, and posterior regions of the superior tem-poral gyri and sulci. Semantic processing of singlewords extends even further in the anterior, ventral,and posterior directions into the middle and infe-rior temporal cortex. Remarkably, however, seman-tic processing of plausible relative to implausiblesentences was more focal than for words, and wasobserved in the ventral banks of the anterior andposterior superior temporal sulci, which lie betweenthe activations associated with prelexical processingin the superior temporal gyri and single-word se-mantics in the middle and inferior temporal gyri(see Fig. 2). This constricted pattern of activation forsentence semantics might reflect the consolidationof multiple words into a reduced set of concepts.

The observation that prelexical and semantic pro-cessing of spoken words extends in anterior, ven-tral, and posterior directions suggests that the samespeech inputs can follow multiple different path-ways. The location of prelexical activation wouldthen be determined by the task demands. For ex-ample, when an articulatory response is required,top-down expectations from motor programs maystabilize and enhance prelexical processing in thedorsal processing direction but when a lexical de-cision is required, top-down expectations fromsemantic knowledge may stabilize and enhanceprelexical activation in the ventral direction. Inthis way, the cortical networks supporting languagecomprehension are dynamically determined by thetask and context30 (see Fig. 3).

Several different sources of top-down inputs/constraints are distinguished in Figure 3, includ-ing semantic, syntactic (word sequences), articula-tion, auditory, visual (spelling), and spatial locationfor locomotion. The review reported in this sectionlocalized semantic constraints to the left pars or-bitalis, ventral pars opercularis, angular gyrus, and

precuneus. Word order constraints (grammatical se-quencing) were identified in the dorsal pars oper-cularis; and articulatory constraints were associatedwith activation in the left motor cortex,60 vSMGand pPT (see Fig. 3). For further discussion of theseregions (see section “Summary of speech produc-tion”).

Left lateralization for speech comprehensionDifferences in the role of the left and right temporallobes have been of particular interest because lesionstudies indicate that it is the left rather than righttemporal lobe that is needed for speech recogni-tion and production. In this context it is surprisingthat the comparison of speech and closely matchednon-speech stimuli result in bilateral temporal lobeactivation.9,10 The specialization and lateralizationfor temporal lobe speech function may therefore bedriven by nonacoustic differences between speechand non-speech stimuli. These nonacoustic differ-ences include our significantly greater experiencewith speech than non-speech stimuli,14 the result-ing categorical qualities of the perceptual represen-tation of speech,71 the ability to produce as well asto perceive speech,60,61 and the influence of spatialorienting.62

Why should the effect of auditory expertise14,23

be greater in the left than right temporal lobe? Here Iconsider two possibilities: (i) either familiarity withthe auditory stimulus increases bottom-up connec-tions from acoustic processing in bilateral supe-rior temporal lobes to representations of seman-tics, grammar, and articulation in the left frontalcortex; or (ii) familiarity results in more success-ful top-down predictions from left frontal regionsthat stabilize acoustic processing in the left temporallobe more than the right temporal lobe. Adank andDevlin23 support the latter hypothesis. They suggestthat the left-lateralized temporal lobe response theyobserved when time-compressed speech was rec-ognized might be driven by top-down processingfrom left-lateralized premotor activation. Kouiderand colleagues11 also highlight the influence of top-down inputs on lexical processing to explain whythey observed repetition suppression in left HGfor words but not pseudowords. Similarly, the left-lateralized anterior temporal lobe response that Leffand colleagues3 observed for vowel deviants couldreflect top-down mechanisms (see section “Prelexi-cal phonemic processing”).6

Ann. N.Y. Acad. Sci. 1191 (2010) 62–88 c© 2010 New York Academy of Sciences. 75

Page 15: The anatomy of language: a review of 100 fMRI studies ...qibawiki.rsna.org/images/5/5c/DrMohamed-Price_etal_2010-20100420.pdf · The anatomy of language: a review of 100 fMRI studies

The anatomy of language Price

Figure 3. Functional attributions. Top panel: Summary of functions attributed to language areas, see text for rationale.Future reviews are required to postulate and illustrate the multiple pathways that connect these areas. Below panel: aschematic summary of the balance between bottom-up and top-down processing during speech comprehension. Thehypothesis is that top-down constraints from frontal, motor, and parietal areas influence the location of activationfoci for bottom-up prelexical processing of auditory inputs.

76 Ann. N.Y. Acad. Sci. 1191 (2010) 62–88 c© 2010 New York Academy of Sciences.

Page 16: The anatomy of language: a review of 100 fMRI studies ...qibawiki.rsna.org/images/5/5c/DrMohamed-Price_etal_2010-20100420.pdf · The anatomy of language: a review of 100 fMRI studies

Price The anatomy of language

In conclusion, frontal lobe activations duringspeech comprehension are more consistently left-lateralized than temporal lobe activations. The pro-posal developed here is that lateralization in thetemporal lobes may be driven top-down from higherlevel processing in frontal areas. Moreover, differentparts of the frontal cortex may determine lateraliza-tion in different temporal regions. If this hypothesisis correct, then the more damaging effect of lefttemporal lobe lesions compared to right temporallobe lesions on language function might be a con-sequence of loss of top-down connections from an-terior speech areas. This could be investigated withfunctional connectivity studies.

Review of speech production

This section focuses on speech production, a com-plex multistage process that links conceptual ideasto articulation. The cognitive components of speechproduction have been described in many studies.Its roots are in the conceptual ideas that need tobe expressed. The words associated with these con-cepts must then be retrieved and sequenced withtheir appropriate morphological forms. At the sametime, competition from other words with similarmeanings needs to be suppressed. Word selectionis therefore a dynamic interaction between excita-tion of the intended words and inhibition of theunintended words.72–74

Sequencing is required at several different levels ofthe speech production process. At the sentence level,words need to be sequenced in a grammatically cor-rect order. Within a word, sequencing is required toassemble the correct combination of phonemes andsyllables. In addition, pitch, prosody, and a metri-cal structure need to be applied to speech output,75

which will depend on the conversational context.The combination of phonemes and syllables beingproduced is translated into a sequence of articu-latory plans (phonetic encoding). This is followedby the initiation and coordination of sequences ofmovements in the speech articulators, which in-clude the tongue, lips, and laryngeal muscles thatvibrate the vocal tract for vowel phonation andprosody. At the same time, respiration needs to becontrolled and the resulting auditory signal needsto be monitored and fed back to the motor systemfor online correction of speech production.

Given the many processes that support speechproduction, it is not surprising that many different

brain areas are involved. Each level of the classicalspeech output pipeline are addressed in the follow-ing sections: conceptual processing (controlling forword retrieval and articulation) (section “Concep-tual processing in speech production”); word re-trieval (controlling for conceptual processing andarticulation) (section “Word retrieval”); articula-tion (controlling for conceptual processing andword retrieval) (section “Articulation”); and au-ditory feedback (section “Monitoring speech out-put”).

Conceptual processing in speech productionIn a large-scale review of 120 functional neuroimag-ing studies of the neural systems that store and re-trieve semantic memories, Binder and colleagues16

identified seven left-lateralized regions associatedwith amodal semantic processing: (i) inferior frontalgyrus, (ii) ventral and dorsal medial prefrontal cor-tex, (iii) posterior inferior parietal lobe, (iv) middletemporal gyrus, (v) fusiform, (vi) parahippocam-pal gyri, and (vii) the posterior cingulate gyrus. Allthese areas have been associated with speech com-prehension in the previous section. Their contribu-tion to speech production depends on the task andthe type of semantic information that needs to beretrieved.

A demonstration that speech production andspeech comprehension access the same conceptualsystem was provided by de Zubicaray and McMa-hon.76 They found that activation in the left parsorbitalis [−51 24 −9] was reduced when picturenaming occurred in the context of semantically re-lated heard words versus unrelated or phonologi-cally related heard words. The authors suggest thatthe picture is recognized before the auditory wordand semantic processing of the picture facilitatesthe recognition of the auditory word. The coordi-nates of this effect [−51 24 −9] also correspond tothose reported by Tyler and colleagues39 at [−4824 −9] and by Ye and Zhou35 at [−52 32 −4]when heard sentences were semantically implausibleversus plausible. The left pars orbitalis is thereforeinvolved in semantic retrieval processes that sup-port both speech comprehension and productiontasks.

A direct comparison of semantic versus phono-logical word retrieval by Birn and colleagues20 hasillustrated the role of the medial superior frontal cor-tex at [−21 11 45] in the conceptual processes that

Ann. N.Y. Acad. Sci. 1191 (2010) 62–88 c© 2010 New York Academy of Sciences. 77

Page 17: The anatomy of language: a review of 100 fMRI studies ...qibawiki.rsna.org/images/5/5c/DrMohamed-Price_etal_2010-20100420.pdf · The anatomy of language: a review of 100 fMRI studies

The anatomy of language Price

support speech production. It is interesting to notethat activation in the medial superior frontal cortexhas also been reported by Tremblay and Gracco77

at [−21 19 57] for “volitional” versus stimulus-driven word selection. This raises the possibilitythat semantic word retrieval (e.g., producing wordsthat belong to a particular semantic category like“fruits”) requires more volitional effort than phono-logical word retrieval (e.g., producing words thatstart with the same letter). The point here is that in-creased activation for semantic tasks may reflect thedemands on executive strategies that are necessaryfor, but not limited to, semantic word retrieval.

In addition to the widely distributed set of re-gions that are typically associated with seman-tic/conceptual processing during speech productionand comprehension, many other sensory-motor ar-eas may play a role in accessing words with specificmeanings. This was illustrated in a study of verbalfluency by Hwang and colleagues,78 who found thatthe retrieval of words belonging to visual categoriesactivated extrastriate cortex (a secondary visual pro-cessing area); retrieval of words belonging to motorcategories activated the intraparietal sulcus and pos-terior middle temporal cortex; and retrieval of wordsbelonging to somato-sensory categories activatedpostcentral and inferior parietal regions. Thus, theydemonstrate that lexico-semantic processing dur-ing speech production is distributed across brainregions participating in sensorimotor processing.This is argued to be a consequence of the sen-sorimotor experiences that occurred during wordacquisition.

Hocking and colleagues72 also demonstrate howactivation during speech production depends on thetype of words that are being retrieved. These authorsfound bilateral hippocampal activation at [−30 −3−30/33 −6 −33] when pictures to be named wereblocked in terms of their semantic category; andbilateral anterior medial temporal activations whenthe objects to be named were blocked accordingto similar visual features. Whitney and colleagues79

also found left hippocampal activation [−28 −35 5]for spontaneous word production (free verbal asso-ciation) relative to category and letter fluency. Thesemedial temporal lobe activations during speech pro-duction tasks therefore appear to depend on thedemands placed on semantic retrieval even thoughmedial temporal lobe structures such as the hip-pocampus are more classically associated with the

initial acquisition of new words.80 To conclude, con-ceptual processing during speech production acti-vates the same set of regions that have been associ-ated with single-word comprehension.

Word retrievalTo investigate the processing that supports word re-trieval, speech production has been compared totasks that control for articulation. For example,Whitney and colleagues79 and Jeon and colleagues81

compared word generation tasks (find, select, andproduce a word related to the stimulus) to read-ing (produce the word or nonword specified by thestimulus). In both studies, the most significant andconsistent effects were observed in the left inferiorand middle frontal gyri, spanning both the pars op-ercularis (BA 44), the pars triangularis (BA45), andthe inferior frontal sulcus. The peak coordinates forthe left middle frontal activations in these two stud-ies are remarkably similar: at [−48 28 21] in Whitneyand colleagues79 and at [−51 25 25] in Jeon and col-leagues81 Two other studies also demonstrated thatthis left middle frontal area is involved in word re-trieval irrespective of the type of words that neededto be produced. Specifically, Heim and colleagues82

observed that the same left middle frontal region at[−44 28 22] was activated for generating words thatwere either semantically or phonologically relatedto a word; and de Zubicaray and McMahon76 foundan effect of semantic and phonological priming onpicture naming at [−48 21 18], which is again con-sistent with a generic role for this region in wordretrieval.

A second region that is consistently activated inword retrieval tasks is the left dorsal pars opercu-laris that was associated with top-down sequencingof linguistic and nonlinguistic events in the speechcomprehension section (“Syntactic constraints”).During speech production, left dorsal pars oper-cularis activation was located at [−40 17 25] byJeon and colleagues81 for word generation morethan reading; and at [−49 13 29] by Fridrikssonand colleagues83 for imitating or observing a speakerproducing nonsense syllables. As viewing a speakerproducing nonsense syllables does not involve se-mantics or articulation,83 speech production activa-tion in the left dorsal pars opercularis is not specificto semantic or syntactic processing. Indeed, as dis-cussed in the speech comprehension section earlier,the left dorsal pars opercularis is involved in the

78 Ann. N.Y. Acad. Sci. 1191 (2010) 62–88 c© 2010 New York Academy of Sciences.

Page 18: The anatomy of language: a review of 100 fMRI studies ...qibawiki.rsna.org/images/5/5c/DrMohamed-Price_etal_2010-20100420.pdf · The anatomy of language: a review of 100 fMRI studies

Price The anatomy of language

hierarchical sequencing of linguistic and nonlin-guistic events.37,50,53,54 This is required for speechproduction as well as speech comprehension. Theleft dorsal pars opercularis may therefore play a rolein the hierarchical sequencing of events during pro-duction as well as comprehension tasks.

A third region that is consistently activated inword retrieval tasks is the left ventral pars opercu-laris that was associated with predicting semanticor articulatory sequences in the speech compre-hension section (“Syntactic constraints”). Duringspeech production, this region has been reportedat [−57 9 9] by Zheng and colleagues84 for artic-ulating versus listening to the word “Ted”; and at[−54 12 12] by Papoutsi and colleagues43 for therepetition and subvocal rehearsal of pseudowordswith low versus high sublexical frequency. As thismanipulation does not involve semantic, syntactic,or phonological word form processing, the authorsattribute their activations to phonetic encoding (ar-ticulatory planning). Left ventral pars opercularisduring speech production has also been reportedin two studies by Heim and colleagues.82,85 In onestudy, Heim and colleagues82 found that left ventralpars opercularis activation at [−46 10 4] was greaterfor retrieving words that were phonologically ver-sus semantically related to a target. Another85 foundleft ventral pars opercularis activation at [−50 6 6]during picture naming in German was sensitive tothe gender of the noun being produced. This wasprobed by presenting pictures to be named with anauditory word and its gender determiner (e.g., “derTisch”—the table). The gender of the auditory wordand picture were either the same or different withmore activation when they were different. However,it is not clear from this study whether activation inthe left ventral pars opercularis was sensitive to thegender of the determiner (which would reflect syn-tactic processing) or to the phonological overlap inthe determiner being heard and the determiner be-ing retrieved (which could reflect articulatory pro-cesses). The explanation that is most consistent withall the other data being presented is that left ventralpars opercularis activation reflects the demands onarticulatory planning.

The distinction between ventral and dorsal parsopercularis has been rather arbitrary. I have been us-ing “dorsal” to refer to activation that is 20–30 mmabove the ACPC line and “ventral” when activationis 4–20 mm above the ACPC line but there may be

multiple functional subregions within each of theseregions. Moreover, some studies are at the border ofventral and dorsal pars opercularis activation. Forexample, Heim and colleagues86 report activation at[−50 8 19] when pictures to be named were blockedaccording to similarity in semantics or syntax (gen-der) versus phonology. The results were interpretedin terms of decreased demands when name retrievalwas primed by phonologically related responses andincreased demands when stimuli are presented withsemantic and syntactic competitors. However, with-out further studies it is not possible to determinewhether the function of this mid-pars opercularisregion corresponds to the properties of the dorsalor ventral regions.

In summary, word retrieval activates the left mid-dle frontal cortex and dorsal and ventral left parsopercularis. Left middle frontal activation was notassociated with speech comprehension, unless thetask required explicit semantic comparisons be-tween two words,16 therefore the left middle frontalcortex appears to be more involved in word retrievalthan word comprehension. In contrast, activation inthe dorsal and ventral regions of the pars opercularishas been reported for both speech production andcomprehension (see Fig. 2). The dorsal pars opercu-laris has been associated with sequencing linguisticand nonlinguistic events and the ventral pars op-ercularis has been associated with sequencing ar-ticulatory events.43 Generating or interpreting thesequence of events is essential for speech compre-hension and production at both the individual word(i.e., the sequence of phonemes) and sentence level(i.e., the sequence of words). The dorsal and ven-tral pars opercularis may contribute to sequencingby top-down constraints from prior knowledge ofwhat event typically follows another.

ArticulationWhen speech is produced, activation typically in-creases bilaterally in motor and premotor cortex, thecerebellum, the supplementary motor area (SMAand pre-SMA), the superior temporal gyri, thetemporo-parietal cortices (PPT/vSMG), and the an-terior insula, with left-lateralized activation in theputamen (see Brown et al.,87 for a recent review).None of these regions are dedicated to speech perse as illustrated by Chang and colleagues88 who re-port the same areas in the production of speechand the production of non-speech sounds from

Ann. N.Y. Acad. Sci. 1191 (2010) 62–88 c© 2010 New York Academy of Sciences. 79

Page 19: The anatomy of language: a review of 100 fMRI studies ...qibawiki.rsna.org/images/5/5c/DrMohamed-Price_etal_2010-20100420.pdf · The anatomy of language: a review of 100 fMRI studies

The anatomy of language Price

orofacial and vocal tract gestures (e.g., cough, sigh,kiss, snort, laugh, tongue click, whistle, cry) thathave no phonemic content. The only areas thatChang and colleagues88 found to activate for speechmore than vocal tract gestures were the ACC andbilateral caudate. Both these areas have been associ-ated with suppression of inappropriate responsesby Aarts and colleagues,46 Ali and colleagues,74

and Kircher and colleagues.89 Greater activation forspeech than non-speech production may thereforereflect the greater demands on response selectionin the context of more competitors that need tobe suppressed during the selection of speech versusnon-speech movements.

The functions of the anterior cingulate and cau-date nuclei during word selection are likely to be dif-ferent. There are many subregions in the ACC thatmay each have their own function during speechproduction.74,90,91 The head of the caudate has alsobeen implicated in sequencing temporal and cogni-tive processes,92 including the sequencing of nonlin-guistic events.53 It is therefore plausible that greateractivation in the head of the caudate nuclei forspeech relative to non-speech88 might contribute togreater precision in the sequencing of speech relativeto non-speech sounds.

Other studies have shown that activation in areascommon to speech and non-speech vocal produc-tion (i.e., the premotor, parietal, temporal, cerebel-lar, and putamen) increases with the length of thespoken utterance43 but is inversely related to thefamiliarity of the stimulus. For example, Papoutsiand colleagues43 found more activation throughoutthe speech production system for the repetition ofpseudowords with four syllables versus two sylla-bles; and Shuster93 found activation in this systemincreased for repetition of pseudowords comparedto the repetition of words. Interestingly, the mostextensive difference between pseudoword and wordrepetition in Shuster93 was located in the left ante-rior insula. This is consistent with a study by Moserand colleagues,94 who found increased activation inthe left anterior insula for repetition of pseudowordswith novel syllables relative to the repetition of pseu-dowords with native (i.e., familiar) syllables.

With respect to the function of the left ante-rior insula, the meta-analysis by Brown and col-leagues87 found that the frontal operculum andmedial-adjacent anterior insula were activated bysyllable singing as well as oral reading; and Koelsch

and colleagues42 found left anterior insula activa-tion for the rehearsal of tone (pitch) information[−32 19 3] as well as verbal information [−32 154]. Brown and colleagues87 speculate that the ante-rior insula is involved in generalized orofacial func-tions, including lip movement, tongue movement,and vocalization. The fact that activation is not de-pendent on whether speech is overt or covert;83 andnot reported to depend on the number syllables be-ing produced43 is consistent with previous claimsthat the anterior insula is involved in the planningrather than execution of articulation. In this con-text, increased activation in the left anterior insulafor unfamiliar speech sounds93,94 may simply reflectgreater demands on articulatory speech plans whenthey are unfamiliar versus familiar.

With respect to the function of the premotor cor-tex, Brown and colleagues87 distinguish areas thatcontrol larynx movements from those that controltongue and mouth movements. In Talairach coor-dinates, Brown and colleagues87 locate the larynxmotor cortex to [−40 −12 30/44 −10 30] and lo-cate tongue movements to part of the rolandic op-erculum at [−50 −9 23/59 −5 17], consistent withother studies of non-speech oral movements.95 Incontrast, Meister and colleagues96 found that themost dorsal part of the premotor cortex [−48 −1254/ −45 −12 60], above the larynx motor area, isactivated by finger tapping as well as articulation.This led Meister and colleagues96 to suggest thatthe dorsal premotor region plays a role in actionselection and planning within the context of arbi-trary stimulus–response mapping tasks. Thus, thereare at least three functionally distinct regions in thepremotor cortex. Of these three regions, Brown andcolleagues87 note that activation during speech istypically strongest in areas associated with phona-tion (i.e., the vibration of the vocal tracts). Thisis consistent with the findings of Fridriksson andcolleagues,83 who show peak activation for overt ar-ticulation relative to observation of articulation at[−58 −2 36/58 −10 34], which is in the vicinity ofthe larynx area rather than the mouth and tonguearea.

The direct comparison of producing versus ob-serving speech production83 identified the pre-SMA[2 6 60] and left putamen [−24 −6 6] as well asthe premotor cortex discussed earlier. This suggeststhat these regions are involved in the initiation orexecution of speech movements but activation is

80 Ann. N.Y. Acad. Sci. 1191 (2010) 62–88 c© 2010 New York Academy of Sciences.

Page 20: The anatomy of language: a review of 100 fMRI studies ...qibawiki.rsna.org/images/5/5c/DrMohamed-Price_etal_2010-20100420.pdf · The anatomy of language: a review of 100 fMRI studies

Price The anatomy of language

not specific to speech movements. For example,Tremblay and Gracco77 observed pre-SMA activa-tion at [−6 13 50] for both volitional and stimulusdriven mouth movements, irrespective of whetherthe response was speech or oral motor gestures; andBahlmann and colleagues53 report activation at [14 57] for learning the hierarchical organization ofsequentially occurring nonlinguistic visual symbols.

In summary, articulatory planning of orofacialmovements activates the left anterior insula.87 Thisarea is equally activated for overt articulation andaction observation83; and more activated when themotor plans are unfamiliar.93,94 The initiation andexecution of movement increases activation in bi-lateral premotor/motor cortex, the pre-SMA, andthe left putamen.83 At least three distinct areas inthe premotor/motor cortex support speech produc-tion. The most ventral is involved in tongue move-ments, a more dorsal region is involved in controlof the larynx, and the most dorsal region is involvedin planning both speech and finger tapping move-ments. Two areas were found to be more involvedwith articulation of speech than non-speech orofa-cial movements. These were the anterior cingulateand bilateral head of caudate. Both regions havebeen associated with the suppression of competingresponses (speech and non-speech), which suggeststhat the subtleties and precision involved in cor-rect speech production requires more suppressionof competition from nontargets (i.e., all the wordswe know with similar articulatory patterns). Thenext section discusses the role of the superior tem-poral gyri, pPT, vSMG, and cerebellum in speechproduction.

Monitoring speech outputThe final stage of speech production involves au-ditory and somato-sensory monitoring of the spo-ken response. This is crucial for online correctionof speech production, for example, when speakersmodify the intensity of their speech in noisy en-vironments, or when auditory feedback is altered(e.g., by delay on the telephone). There are threelevels at which this feedback can occur. One is fromhearing the sound of the spoken response (auditoryfeedback), the second is higher level “phonological”processing, and the third is from somato-sensory in-formation derived from the speech movements.72,97

Auditory processing of the spoken response ac-tivates bilateral superior temporal regions associ-

ated with speech perception. This has been ob-served when auditory processing of the spoken re-sponse was not controlled with a speech outputbaseline, for example, by Heim and colleagues86 andWilson and colleagues58 during picture naming andby Brown and colleagues87 during reading aloud.When speech output is controlled, however, supe-rior temporal activation has not been reported forword generation tasks.79,81

With respect to higher level phonological pro-cessing of the spoken output, three recent studieshave highlighted the involvement of the pPT and/orvSMG, when speech production is made more diffi-cult and therefore requires more auditory monitor-ing. In Hocking et al.,72 activation at [−54 −36 15]increased when pictures were named in the contextof semantic interference; and Abel and colleagues,73

showed that picture naming activation at [−56 −3715] was stronger for phonological than semantic in-terference. The argument here is that both semanticand phonological interference make picture namingmore error prone and pPT and/or vSMG activationis involved in monitoring the verbal response toreduce errors. The effect of verbal self-monitoringon left pPT/vSMG activation was directly demon-strated by Zheng and colleagues84 Their participantseither whispered or produced the word “Ted.” Dur-ing production, the auditory input was either (i)consistent with the speech output (i.e., the word“Ted”) or (ii) a masking noise that prevented par-ticipants hearing the sound of their own speech. LeftpPT/vSMG activation at [−66 −45 15] was greatestwhen participants spoke in the context of a maskrelative to the other conditions. This effect couldnot be explained by either speaking or hearing perse but indicates greater pPT/vSMG activation whenthere was conflict between the expected auditoryinput and the actual auditory input. This is consis-tent with the interaction between auditory inputsand auditory predictions that are generated by thespeech production system or conversely motor effer-ents and motor predictions that are generated fromthe auditory speech inputs, Zheng and colleagues.84

Linking back to the speech comprehension sec-tion, pPT/vSMG activation was attributed to sub-vocal articulation in perceptually or semanticallychallenging contexts. Increased subvocal articula-tion can also explain increased pPT/vSMG ac-tivation during speech production in the con-text of semantic,72 phonological,73 or acoustic

Ann. N.Y. Acad. Sci. 1191 (2010) 62–88 c© 2010 New York Academy of Sciences. 81

Page 21: The anatomy of language: a review of 100 fMRI studies ...qibawiki.rsna.org/images/5/5c/DrMohamed-Price_etal_2010-20100420.pdf · The anatomy of language: a review of 100 fMRI studies

The anatomy of language Price

interference.84 Specifically, pPT/vSMG activationcould reflect increased subvocal articulation fromsemantic and phonological competitors72,73 that isthen suppressed before motor output. There mayalso be increased subvocal articulation during pro-duction when no corresponding articulatory sig-nals are received from auditory processing of theoutput.84 In both cases, PPT and/or vSMG may beinvolved in suppressing articulatory errors.

With respect to somato-sensory feedback, in theabsence of auditory feedback, Peschke and col-leagues97 suggest a role for the postcentral gyrus.Consistent with this proposal, Zheng and col-leagues84 report activation in the left postcentralgyrus when auditory feedback during speech pro-duction was masked; and Hocking and colleagues72

report activation in the right postcentral gyrus dur-ing error-prone speech production. It is also relevantto note that, during the same conditions that evokedpostcentral activation, both Zheng and colleagues84

and Hocking and colleagues72 also report activationin the left cerebellum. This would be consistent witha role for both the postcentral gyri and the left cere-bellum in somato-sensory feedback during speechproduction.

In summary, auditory monitoring and feedbackduring speech production requires the integrationof auditory, articulatory, and somato-sensory sig-nals with the motor output. The dynamics of thissystem have been discussed by van de Ven and col-leagues,98 who used spatial and temporal indepen-dent components analysis to show the dynamic cou-pling between different functional networks andthereby demonstrated a negative correlation be-tween activation in the bilateral auditory cortex andthe supplementary motor area [−1 −10 58].

Summary of speech productionThe first thing to point out is that many of theareas associated with speech production were alsodiscussed in the previous section on speech compre-hension. The overlap in areas associated with con-ceptual processing is not surprising but the overlapin ventral and dorsal regions of the pars opercu-laris, pPT, and vSMG requires more consideration.Activation in both the dorsal and ventral pars op-ercularis has been associated with top-down expec-tations on the predicted sequence of events, whichis required at multiple different stages during bothspeech comprehension and speech production. The

dorsal pars opercularis is involved in sequencing lin-guisitic and nonlinguistic events and appears to beindependent of the demands on working memory.In contrast, the ventral pars opercularis is involvedin sequencing articulatory events and is involved inworking memory. It is therefore possible that in-creased ventral pars opercularis activation duringdifficult speech comprehension9,10,38 reflects top-down predictions about the possible or expectedarticulatory codes associated with the heard words.

Activation in the pPT/vSMG regions has alsobeen associated with subvocal articulation but theresponse properties in pPT/vSMG and the left ven-tral pars opercularis are distinct.42,49,84 For example,Raettig and colleagues49 reported ventral pars oper-cularis activation for errors in verb–noun agreement(“Peter has laughed the apple”) but pPT/vSMG ac-tivation for morphosyntactic errors (“Peter has eatthe apple”). One simplistic interpretation of this isthat morphosyntactic errors and pPT/vSMG acti-vation reflect a mismatch in the predicted articula-tion at the level of a single event. In contrast, verb–noun agreement errors and ventral pars opercularisactivation reflect a mismatch in the predicted ar-ticulation at the level of sequences of articulatoryevents. This might also explain why Koelsch andcolleagues42 reported ventral pars opercularis ac-tivation for memorizing lists of heard syllables atthe same time as singing, a combination of tasksthat would provide conflicting predictions aboutthe forthcoming sequence of events.

Areas that are more specific to speech produc-tion were the left middle frontal gyrus, left anteriorinsula, left putamen, bilateral head of caudate, an-terior cingulate, pre-SMA, SMA, motor cortex, andcerebellum. The left middle frontal gyrus has beenassociated with word retrieval when articulation iscontrolled; the left anterior insula been associatedwith articulatory planning irrespective of whetherthe sound is produced or not; and the left putamen,pre-SMA, and motor cortex have been associatedwith the initiation and execution of overt relative tocovert speech.

Despite the new insights into the neural process-ing of speech production, the number of functionalimaging studies of speech production has been lessthan the number of speech production studies.This is likely to be a consequence of speech pro-duction being more difficult to study than speechcomprehension. The difficulties studying speech

82 Ann. N.Y. Acad. Sci. 1191 (2010) 62–88 c© 2010 New York Academy of Sciences.

Page 22: The anatomy of language: a review of 100 fMRI studies ...qibawiki.rsna.org/images/5/5c/DrMohamed-Price_etal_2010-20100420.pdf · The anatomy of language: a review of 100 fMRI studies

Price The anatomy of language

production arise because it involves mouth move-ments that can cause artifacts in the imaging signalas well as increasing head movement. In previousyears, this led to a preference for speech produc-tion paradigms that used covert rather than overtresponses. New techniques for minimizing mouthmovements and measuring overt responses in thescanner have helped to overcome this previous lim-itation.43 There have also been considerable chal-lenges in studying speech production at the narra-tive rather than single-word level. This is because,during narrative speech production, it is very diffi-cult to constrain and control the speech productionrate, the order of words that are retrieved and thechoice of words relating to a semantic theme.

As the experimental and technical challenges as-sociated with studies of speech production are con-quered, many new questions can be addressed. Forexample, there has been remarkably little discus-sion of the cerebellum in any of the papers I re-viewed, even though it is clearly activated duringarticulation,87 working memory,42 word acquisi-tion,80 and auditory self-monitoring during speechproduction.72,84 Eickhoff and colleagues99 have alsosuggested that, during verbal fluency, activity in thecerebellum is fed to the left anterior insula. Therehas also been little discussion in the reviewed pa-pers on the degree to which activation is lateralizedduring speech production. My overall impression isone where motor, premotor, subcortical, and supe-rior temporal activation are bilaterally activated butmiddle frontal activation is left lateralized. In con-trast, Birn and colleagues20 reported that temporo-parietal activation was right-lateralized for auto-matic speech production (generating the names ofthe months) compared to semantic (category) andphonological (letter) speech production. Furtherstudies are clearly required to understand the deter-minants of lateralization and this is going to requirethe estimation of lateralization on a region by regionbasis.100

Future directions

This review has considered the functional anatomyof language but it has not drawn any conclusionsabout how the different regions are functionallyconnected or how the systems vary across indi-viduals. Many studies have started to investigatethe functional connectivity of the language sys-

tem either by exploring the anatomical connec-tions between regions,101 conducting functionalconnectivity analyses,82,98,99,102–107 or by com-bining fMRI data with neurophysiological tech-niques that have high temporal resolution includ-ing MEG,108,109 ERP,110–112 and TMS.113–115 There isalso a growing trend to use anatomically constrainedcomputational modeling.116–118 Inter-subject vari-ability in the language networks is also being ex-plored in a number of ways: for example, on thebasis of behavioral variations,58 age,39,63,64,119 mul-tilingualism,119–129 lesion studies,130 or using un-biased classification schemes.131,132 There is alsoincreasing evidence that individual differences inactivation are related to the underlying brain struc-ture100 and that brain structure predicts languageperformance.128,129,133,134 Therefore, future investi-gations can be embellished by the combination ofstructural, functional, and behavioral data.39,64

In addition to establishing the anatomical andfunctional connectivity of the language system,many more studies are still required to increasethe spatial resolution of the anatomical conclusionsthat are currently being drawn. For example, futurestudies are required to distinguish the contributionof different subregions in the pars opercularis.107

One relatively novel approach is to examine howthe pattern of activity within a specified area (ratherthan the amplitude within an area) varies with thestimuli and task.1,70,71,119 Speech selectivity can alsobe examined by cross-referencing between auditoryactivation in animals and humans.135 However, asthe cognitive function of any region depends on theareas that it interacts with, it may only be possibleto dissociate the functions of speech regions by ex-amining how they interact with other brain regions;and how these regional interactions are modulatedby task demands.

Acknowledgments

This work was funded by the Wellcome Trust. Theauthor thanks Joseph Devlin, Sue Ramsden, FionaRichardson, Alex Leff, Mohamed Seghier, DavidGreen, Goulven Josse, and Tom Schofield for theirhelpful comments.

Conflicts of interest

The author declares no conflicts of interest.

Ann. N.Y. Acad. Sci. 1191 (2010) 62–88 c© 2010 New York Academy of Sciences. 83

Page 23: The anatomy of language: a review of 100 fMRI studies ...qibawiki.rsna.org/images/5/5c/DrMohamed-Price_etal_2010-20100420.pdf · The anatomy of language: a review of 100 fMRI studies

The anatomy of language Price

References

1. Hickok, G., K. Okada & J.T. Serences. 2009. Area spt in

the human planum temporale supports sensory-motor

integration for speech processing. J. Neurophysiol. 101:

2725–2732.

2. Rauschecker, J.P. & S.K. Scott. 2009. Maps and streams

in the auditory cortex: nonhuman primates illuminate

human speech processing. Nat. Neurosci. 12: 718–724.

3. Leff, A.P., P. Iverson, T.M. Schofield, et al. 2009. Vowel-

specific mismatch responses in the anterior superior

temporal gyrus: an fMRI study. Cortex 45: 517–526.

4. Britton, B., S.E. Blumstein, E.B. Myers & C. Grindrod.

2009. The role of spectral and durational properties on

hemispheric asymmetries in vowel perception. Neu-

ropsychologia 47: 1096–1106.

5. Myers, E.B., S.E. Blumstein, E. Walsh & J. Eliassen.

2009. Inferior frontal regions underlie the perception

of phonetic category invariance. Psychol. Sci. 20: 895–

903

6. Schofield, T.M., P. Iverson, S.J. Kiebel, et al. 2009.

Changing meaning causes coupling changes within

higher levels of the cortical hierarchy. Proc. Natl. Acad.

Sci. USA 106: 11765–11770.

7. Specht, K., B. Osnes & K. Hugdahl. 2009. Detection

of differential speech-specific processes in the temporal

lobe using fMRI and a dynamic “sound morphing”

technique. Hum. Brain Mapp. 30: 3436–3444.

8. Vaden, K.I. Jr., L.T. Muftuler & G. Hickok. 2009. Phono-

logical repetition-suppression in bilateral superior tem-

poral sulci. Neuroimage 49: 1018–1023.

9. Obleser, J. & S.A. Kotz. 2009. Expectancy constraints

in degraded speech modulate the language com-

prehension network. Cereb. Cortex. doi: 10.1093/cer-

cor/bhp128.

10. Friederici, A.D., S.A. Kotz, S.K. Scott & J. Obleser.

2009. Disentangling syntax and intelligibility in audi-

tory language comprehension. Hum. Brain Mapp. doi:

10.1002/hbm.20878.

11. Kouider, S., V. de Gardelle, S. Dehaene, et al. 2009.

Cerebral bases of subliminal speech priming. Neuroim-

age 49: 922–929.

12. Kotz, S.A., A. D’Ausilio, T. Raettig, et al. 2010. Lexical-

ity drives audio-motor transformations in Broca’s area.

Brain Lang . 112: 3–11.

13. Davis, M.H. & M.G. Gaskell. 2009. A complemen-

tary systems account of word learning: neural and be-

havioural evidence. Phil. Trans. R. Soc. B. 364: 3773–

3800.

14. Leech, R., L.L. Holt, J.T. Devlin & F. Dick. 2009. Exper-

tise with artificial nonspeech sounds recruits speech-

sensitive cortical regions. J. Neurosci. 29: 5234–5239.

15. Sharp, D.J., J.E. Warren, M. Awad, et al. 2009. The neu-

ral response to changing semantic and perceptual com-

plexity during language processing. Hum. Brain Mapp.

doi: 10.1002/hbm.20871.

16. Binder, J.R., R.H. Desai, W.W. Graves & L.L. Conant.

2009. Where is the semantic system? A critical re-

view and meta-analysis of 120 functional neuroimaging

studies. Cereb. Cortex. 19: 2767–2796.

17. Diaz, M.T. & G. McCarthy. 2009. A comparison of

brain activity evoked by single content and function

words: An fMRI investigation of implicit word process-

ing. Brain Res. 1282: 38–49.

18. Whitney, C., W. Huber, J. Klann, et al. 2009. Neural

correlates of narrative shifts during auditory story com-

prehension. Neuroimage 47: 360–366.

19. Mashal, N., M. Faust, T. Hendler & M. Jung-Beeman.

2009. An fMRI study of processing novel metaphoric

sentences. Laterality 14: 30–54.

20. Birn, R.M., L. Kenworthy, L. Case, et al. 2009. Neu-

ral systems supporting lexical search guided by letter

and semantic category cues: a self-paced overt response

fMRI study of verbal fluency. Neuroimage 49: 1099–

1107.

21. Rogalsky, C. & G. Hickok. 2009. Selective attention

to semantic and syntactic features modulates sentence

processing networks in anterior temporal cortex. Cereb.

Cortex 19: 786–796.

22. Visser, M., E. Jefferies & M.A. Lambon Ralph. 2009.

Semantic processing in the anterior temporal lobes: a

meta-analysis of the functional neuroimaging litera-

ture. J. Cogn. Neurosci. doi: 10.1162/jocn.2009.21309.

23. Adank P. & J.T. Devlin. 2009. On-line plasticity in

spoken sentence comprehension: adapting to time-

compressed speech. Neuroimage 49: 1124–1132.

24. Devauchelle, A.D., C. Oppenheim, L. Rizzi, et al. 2009.

Sentence syntax and content in the human temporal

lobe: an fMRI adaptation study in auditory and visual

modalities. J. Cogn. Neurosci. 21: 1000–1112.

25. Hubbard, A.L., S.M. Wilson, D.E. Callan & M. Dapretto.

2009. Giving speech a hand: gesture modulates activ-

ity in auditory cortex during speech perception. Hum.

Brain Mapp. 30: 1028–1037.

26. Snijders, T.M., T. Vosse, G. Kempen, et al. 2009. Re-

trieval and unification of syntactic structure in sentence

comprehension: an FMRI study using word-category

ambiguity. Cereb. Cortex 19: 1493–1503.

27. Kircher, T., B. Straube, D. Leube, et al. Neural inter-

action of speech and gesture: differential activations

84 Ann. N.Y. Acad. Sci. 1191 (2010) 62–88 c© 2010 New York Academy of Sciences.

Page 24: The anatomy of language: a review of 100 fMRI studies ...qibawiki.rsna.org/images/5/5c/DrMohamed-Price_etal_2010-20100420.pdf · The anatomy of language: a review of 100 fMRI studies

Price The anatomy of language

of metaphoric co-verbal gestures. Neuropsychologia 47:

169–179.

28. Holle, H., J. Obleser, S.A. Rueschemeyer & T.C. Gunter.

2009. Integration of iconic gestures and speech in left

superior temporal areas boosts speech comprehension

under adverse listening conditions. Neuroimage 49:

875–884.

29. Robins, D.L., E. Hunyadi & R.T. Schultz. 2009. Supe-

rior temporal activation in response to dynamic audio-

visual emotional cues. Brain Cogn. 69: 269–278.

30. Skipper, J.I., S. Goldin-Meadow, H.C. Nusbaum & S.L.

Small. 2009. Gestures orchestrate brain networks for

language understanding. Curr. Biol. 19: 661–667.

31. Dick, A.S., S. Goldin-Meadow, U. Hasson, et al. 2009.

Co-speech gestures influence neural activity in brain re-

gions associated with processing semantic information.

Hum. Brain Mapp. 30: 3509–3526.

32. Richardson, F.M., M.S. Thomas & C.J. Price. 2009. Neu-

ronal activation for semantically reversible sentences. J.

Cogn. Neurosci. doi: 10.1162/jocn.2009.21277.

33. Carreiras, M., M.L. Seghier, S. Baquero, et al. 2009.

An anatomical signature for literacy. Nature 461: 983–

986.

34. Brownsett, S.L. & R.J. Wise. 2009. The contribution of

the parietal lobes to speaking and writing. Cereb. Cortex.

doi: 10.1093/cercor/bhp120.

35. Ye, Z. & X. Zhou. 2009. Conflict control during sentence

comprehension: fMRI evidence. Neuroimage 48: 280–

290.

36. Bilenko, N.Y., C.M. Grindrod, E.B. Myers & S.E. Blum-

stein. 2009. Neural correlates of semantic competition

during processing of ambiguous words. J. Cogn. Neu-

rosci. 21: 960–975.

37. Willems, R.M., A. Ozyurek & P. Hagoort. 2009. Differ-

ential roles for left inferior frontal and superior tem-

poral cortex in multimodal integration of action and

language. Neuroimage 47: 1992–2004.

38. Desai, R.H., J.R. Binder, L.L. Conant & M.S. Seidenberg.

2010. Activation of sensory-motor areas in sentence

comprehension. Cereb. Cortex. 20: 468–478.

39. Tyler, L.K., M.A. Shafto, B. Randall, et al. 2010. Pre-

serving syntactic processing across the adult life span:

the modulation of the frontotemporal language system

in the context of age-related atrophy. Cereb. Cortex. 20:

352–364.

40. Turner, T.H., J. Fridriksson, J. Baker, et al. 2009. Oblig-

atory Broca’s area modulation associated with passive

speech perception. Neuroreport 20: 492–496.

41. Szycik, G.R., H. Jansma & T.F. Munte. 2009. Audio-

visual integration during speech comprehension: an

fMRI study comparing ROI-based and whole brain

analyses. Hum. Brain Mapp. 30: 1990–1999.

42. Koelsch, S., K. Schulze, D. Sammler, et al. 2009. Func-

tional architecture of verbal and tonal working mem-

ory: an FMRI study. Hum. Brain Mapp. 30: 859–

873.

43. Papoutsi, M., J.A. de Zwart, J.M. Jansma, et al. 2009.

From phonemes to articulatory codes: an fMRI study

of the role of Broca’s area in speech production. Cereb.

Cortex 19: 2156–2165.

44. Nosarti, C., A. Mechelli, D.W. Green & C.J. Price. 2010.

The impact of second language learning on semantic

and nonsemantic first language reading. Cereb. Cortex

20: 315–327.

45. Schafer, R.J. & T. Constable. 2009. Modulation of func-

tional connectivity with the syntactic and semantic de-

mands of a Noun Phrase Formation Task: a possible

role for the Default Network. Neuroimage 46: 882–

890.

46. Aarts, E., A. Roelofs & M. van Turennout. 2009. At-

tentional control of task and response in lateral and

medial frontal cortex: brain activity and reaction time

distributions. Neuropsychologia 47: 2089–2099.

47. Peelle, J.E., V. Troiani & M. Grossman. 2009. Interac-

tion between process and content in semantic memory:

an fMRI study of noun feature knowledge. Neuropsy-

chologia 47: 995–1003.

48. Schmidt, G.L. & C.A. Seger. 2009. Neural correlates of

metaphor processing: the roles of figurativeness, famil-

iarity and difficulty. Brain Cogn. 71: 375–386.

49. Raettig, T., S. Frisch, A.D. Friederici & S.A. Kotz.

2009. Neural correlates of morphosyntactic and verb-

argument structure processing: an EfMRI study.

Cortex. doi: 10.1016/j.cortex.2009.06.003.

50. Makuuchi, M., J. Bahlmann, A. Anwander & A.D.

Friederici. 2009. Segregating the core computational

faculty of human language from working memory. Proc.

Natl. Acad. Sci. USA 106: 8362–8367.

51. Newman, S.D., D. Lee & K.L. Ratliff. 2009. Off-line

sentence processing: what is involved in answering a

comprehension probe? Hum. Brain Mapp. 30: 2499–

2511.

52. Newman, S.D., K. Ratliff, T. Muratore & T. Burns Jr.

2009. The effect of lexical priming on sentence com-

prehension: an fMRI study. Brain Res. 1285: 99–108.

53. Bahlmann, J., R.I. Schubotz, J.L. Mueller, et al. 2009.

Neural circuits of hierarchical visuo-spatial sequence

processing. Brain Res. 1298: 161–170.

54. Tettamanti, M., I. Rotondi, D. Perani, et al. 2009.

Syntax without language: neurobiological evidence for

Ann. N.Y. Acad. Sci. 1191 (2010) 62–88 c© 2010 New York Academy of Sciences. 85

Page 25: The anatomy of language: a review of 100 fMRI studies ...qibawiki.rsna.org/images/5/5c/DrMohamed-Price_etal_2010-20100420.pdf · The anatomy of language: a review of 100 fMRI studies

The anatomy of language Price

cross-domain syntactic computations. Cortex 45: 825–

838.

55. Friederici, A.D., M. Makuuchi & J. Bahlmann. 2009.

The role of the posterior superior temporal cortex in

sentence comprehension. Neuroreport 20: 563–568.

56. Dos Santos Sequeira, S., K. Specht, M. Moosmann, et al.

2009. The effects of background noise on dichotic lis-

tening to consonant-vowel syllables: an fMRI study.

Laterality. doi: 10.1080/13576500903045082.

57. Buchsbaum, B.R. & M. D’Esposito. 2009. Repetition

suppression and reactivation in auditory-verbal short-

term recognition memory. Cereb. Cortex. 19: 1474–

1485.

58. Wilson, S.M., A.L. Isenberg & G. Hickok. 2009. Neu-

ral correlates of word production stages delineated by

parametric modulation of psycholinguistic variables.

Hum. Brain Mapp. 30: 3596–3608.

59. Yoncheva, Y.N., J.D. Zevin, U. Maurer & B.D. McCan-

dliss. 2009. Auditory selective attention to speech mod-

ulates activity in the visual word form area. Cereb. Cor-

tex. doi:10.1093/cercor/bhp129.

60. Devlin, J.T. & J. Aydelott. 2009. Speech perception: mo-

toric contributions versus the motor theory. Curr Biol.

19: R198–R200.

61. Scott, S.K., C. McGettigan & F. Eisner. 2009. A little

more conversation, a little less action: candidate roles

for motor cortex in speech perception. Nature Rev. Neu-

rosci. 10: 295–302.

62. Fischer, J., C. Teufel, M. Drolet, et al. 2009. Orienting

asymmetries and lateralized processing of sounds in

humans. BMC Neurosci. 10: 14.

63. Wong, P.C., J.X. Jin, G.M. Gunasekera, et al. 2009. Aging

and cortical mechanisms of speech perception in noise.

Neuropsychologia 47: 693–703.

64. Harris, K.C., J.R. Dubno, N.I. Keren, et al. 2009. Speech

recognition in younger and older adults: a dependency

on low-level auditory cortex. J. Neurosci. 29: 6078–

6087.

65. Wiethoff, S., D. Wildgruber, W. Grodd & T. Ethofer.

2009. Response and habituation of the amygdala dur-

ing processing of emotional prosody. Neuroreport 20:

1356–1360.

66. Fusar-Poli, P., A. Placentino, F. Carletti, et al. 2009.

Laterality effect on emotional faces processing: ALE

meta-analysis of evidence. Neurosci. Lett. 452: 262–267.

67. Herbert, C., T. Ethofer, S. Anders, et al. 2009. Amygdala

activation during reading of emotional adjectives—an

advantage for pleasant content. Soc. Cogn. Affect. Neu-

rosci. 4: 35–49.

68. Wittfoth, M., C. Schroder, D.M. Schardt, et al. 2010. On

emotional conflict: interference resolution of happy and

angry prosody reveals valence-specific effects. Cereb.

Cortex. 20: 383–392.

69. Kreifelts, B., T. Ethofer, T. Shiozawa, et al. 2009. Cere-

bral representation of non-verbal emotional percep-

tion: fMRI reveals audiovisual integration area between

voice- and face-sensitive regions in the superior tem-

poral sulcus. Neuropsychologia. 47: 3059–3066.

70. Ethofer, T., D. Van De Ville, K. Scherer & P. Vuilleumier.

2009. Decoding of emotional information in voice-

sensitive cortices. Curr. Biol. 19: 1028–1033.

71. Obleser, J. & F. Eisner. 2009. Pre-lexical abstraction of

speech in the auditory cortex. Trends Cogn Sci. 13: 14–

19.

72. Hocking, J., K.L. McMahon & G.I. de Zubicaray. 2009.

Semantic context and visual feature effects in object

naming: an fMRI study using arterial spin labeling. J.

Cogn. Neurosci. 21: 1571–1583.

73. Abel, S., K. Dressel, R. Bitzer, et al. 2009. The separa-

tion of processing stages in a lexical interference fMRI-

paradigm. Neuroimage 44: 1113–1124.

74. Ali, N., D.W. Green, F. Kherif, et al. 2009. Role of the

left caudate in suppressing irrelevant words. J. Cogn.

Neurosci. doi:10.1162/jocn.2009.21352.

75. Peck, K.K., J.F. Galgano, R.C. Branski, et al. 2009. Event-

related functional MRI investigation of vocal pitch vari-

ation. Neuroimage 44: 175–181.

76. de Zubicaray, G. I. & K.L. McMahon. 2009. Audi-

tory context effects in picture naming investigated with

event-related fMRI. Cogn. Affect. Behav. Neurosci. 9:

260–269.

77. Tremblay, P. & V.L. Gracco. 2010. On the selection

of words and oral motor responses: evidence of a

response-independent fronto-parietal network. Cortex.

46: 15–28.

78. Hwang, K., E.D. Palmer, S. Basho, et al. 2009. Category-

specific activations during word generation reflect ex-

periential sensorimotor modalities. Neuroimage 48:

717–725.

79. Whitney, C., S. Weis, T. Krings, et al. 2009. Task-

dependent modulations of prefrontal and hippocam-

pal activity during intrinsic word production. J. Cogn.

Neurosci. 21: 697–712.

80. Davis, M.H., A.M. Di Betta, M.J. Macdonald & M.G.

Gaskell. 2009. Learning and consolidation of novel spo-

ken words. J. Cogn. Neurosci. 21: 803–820.

81. Jeon, H.A., K.M. Lee, Y.B. Kim & Z.H. Cho. 2009.

Neural substrates of semantic relationships: common

and distinct left-frontal activities for generation of syn-

onyms vs. antonyms. Neuroimage 48: 449–457.

86 Ann. N.Y. Acad. Sci. 1191 (2010) 62–88 c© 2010 New York Academy of Sciences.

Page 26: The anatomy of language: a review of 100 fMRI studies ...qibawiki.rsna.org/images/5/5c/DrMohamed-Price_etal_2010-20100420.pdf · The anatomy of language: a review of 100 fMRI studies

Price The anatomy of language

82. Heim, S., S.B. Eickhoff & K. Amunts. 2009. Different

roles of cytoarchitectonic BA 44 and BA 45 in phonolog-

ical and semantic verbal fluency as revealed by dynamic

causal modelling. Neuroimage 48: 616–624.

83. Fridriksson, J., D. Moser, J. Ryalls, et al. 2009. Modu-

lation of frontal lobe speech areas associated with the

production and perception of speech movements. J.

Speech Lang. Hear Res. 52: 812–819.

84. Zheng, Z.Z., K.G. Munhall & I.S. Johnsrude. 2009.

Functional overlap between regions involved in speech

perception and in monitoring one’s own voice

during speech production. J. Cogn. Neurosci. doi:

10.1162/jocn.2009.21324.

85. Heim, S., A.D. Friederici, N.O. Schiller, et al. 2009. The

determiner congruency effect in language production

investigated with functional MRI. Hum. Brain Mapp.

30: 928–940.

86. Heim, S., S.B. Eickhoff, A.D. Friederici & K. Amunts.

2009. Left cytoarchitectonic area 44 supports selec-

tion in the mental lexicon during language production.

Brain Struct. Funct . 213: 441–456.

87. Brown, S., A.R. Laird, P. Q. Pfordresher, et al. 2009.

The somatotopy of speech: phonation and articulation

in the human motor cortex. Brain Cogn. 70: 31–41.

88. Chang, S.E., M.K. Kenney, T.M. Loucks, et al. 2009.

Common neural substrates support speech and non-

speech vocal tract gestures. Neuroimage 47: 314–325.

89. Kircher, T., K. Sass, O. Sachs & S. Krach. 2009. Prim-

ing words with pictures: neural correlates of semantic

associations in a cross-modal priming task using fMRI.

Hum. Brain Mapp. 30: 4116–4128.

90. Melcher, T. & O. Gruber. 2009. Decomposing interfer-

ence during Stroop performance into different conflict

factors: an event-related fMRI study. Cortex 45: 189–

200.

91. Haupt, S., N. Axmacher, M. X. Cohen, et al. 2009.

Activation of the caudal anterior cingulate cor-

tex due to task-related interference in an audi-

tory Stroop paradigm. Hum. Brain Mapp. 30: 3043–

3056.

92. Kotz, S.A., M. Schwartze & M. Schmidt-Kassow. 2009.

Non-motor basal ganglia functions: a review and pro-

posal for a model of sensory predictability in auditory

language perception. Cortex 45: 982–990.

93. Shuster, L.I. 2009. The effect of sublexical and lexical

frequency on speech production: an fMRI investigation.

Brain Lang . 111: 66–72.

94. Moser, D., J. Fridriksson, L. Bonilha, et al. 2009. Neu-

ral recruitment for the production of native and novel

speech sounds. Neuroimage 46: 549–557.

95. Byrd, K.E., L.M. Romito, M. Dzemidzic, et al. 2009.

fMRI study of brain activity elicited by oral parafunc-

tional movements. J. Oral Rehabil. 36: 346–361.

96. Meister, I.G., D. Buelte, M. Staedtgen, et al. 2009. The

dorsal premotor cortex orchestrates concurrent speech

and fingertapping movements. Eur. J. Neurosci. 29:

2074–2082.

97. Peschke, C., W. Ziegler, J. Kappes & A. Baumgaertner.

2009. Auditory-motor integration during fast repeti-

tion: the neuronal correlates of shadowing. Neuroimage

47: 392–402.

98. van de Ven, V., F. Esposito & I.K. Christoffels. 2009.

Neural network of speech monitoring overlaps with

overt speech production and comprehension networks:

a sequential spatial and temporal ICA study. Neuroim-

age 47: 1982–1991.

99. Eickhoff, S.B., S. Heim, K. Zilles & K. Amunts. 2009.

A systems perspective on the effective connectivity of

overt speech production. Philos. Transact. A Math. Phys.

Eng. Sci. 367: 2399–2421.

100. Josse, G., F. Kherif, G. Flandin, et al. 2009. Predicting

language lateralization from gray matter. J. Neurosci.

29: 13516–13523.

101. Morgan, V.L., A. Mishra, A.T. Newton, et al. 2009. In-

tegrating functional and diffusion magnetic resonance

imaging for analysis of structure-function relationship

in the human language network. PLoS One 4: e6660.

102. Seghier, M.L. & C.J. Price. 2009. Reading aloud boosts

connectivity through the putamen. Cereb. Cortex. doi:

10.1093/cercor/bhp123.

103. Levy, J., C. Pernet, S. Treserras, et al. 2009. Testing for the

dual-route cascade reading model in the brain: an fMRI

effective connectivity account of an efficient reading

style. PLoS One 4: e6675.

104. Londei, A., A. D’Ausilio, D. Basso, et al. 2009. Sensory-

motor brain network connectivity for speech compre-

hension. Hum. Brain Mapp. doi: 10.1002/hbm.20888.

105. Heim, S., S.B. Eickhoff, A.K. Ischebeck, et al. 2009.

Effective connectivity of the left BA 44, BA 45, and

inferior temporal gyrus during lexical and phonological

decisions identified with DCM. Hum. Brain Mapp. 30:

392–402.

106. van Atteveldt, N., A. Roebroeck & R. Goebel 2009. Inter-

action of speech and script in human auditory cortex:

Insights from neuro-imaging and effective connectivity.

Hear. Res. 258: 152–164.

107. Xiang, H.D., H.M. Fonteijn, D.G. Norris & P. Hagoort

2009. Topographical functional connectivity pattern in

the perisylvian language networks. Cereb. Cortex. doi:

10.1093/cercor/bhp119.

Ann. N.Y. Acad. Sci. 1191 (2010) 62–88 c© 2010 New York Academy of Sciences. 87

Page 27: The anatomy of language: a review of 100 fMRI studies ...qibawiki.rsna.org/images/5/5c/DrMohamed-Price_etal_2010-20100420.pdf · The anatomy of language: a review of 100 fMRI studies

The anatomy of language Price

108. Liljestrom, M., A. Hulten, L. Parkkonen & R. Salmelin.

2009. Comparing MEG and fMRI views to naming ac-

tions and objects. Hum. Brain Mapp. 30: 1845–1856.

109. Vartiainen, J., T. Parviainen & R. Salmelin. 2009. Spa-

tiotemporal convergence of semantic processing in

reading and speech perception. J. Neurosci. 29: 9271–

9280.

110. Brem, S., P. Halder, K. Bucher, et al. 2009. Tuning of the

visual word processing system: distinct developmental

ERP and fMRI effects. Hum. Brain Mapp. 30: 1833–

1844.

111. Kotz, S.A. 2009. A critical review of ERP and fMRI

evidence on L2 syntactic processing. Brain Lang . 109:

68–74.

112. Pulvermuller, F., Y. Shtyrov & O. Hauk. 2009. Under-

standing in an instant: neurophysiological evidence for

mechanistic language circuits in the brain. Brain Lang .

110: 81–94.

113. D’Ausilio, A., F. Pulvermuller, P. Salmas, et al. 2009. The

motor somatotopy of speech perception. Curr. Biol. 19:

381–385.

114. Ruff, C.C., J. Driver & S. Bestmann. 2009. Combining

TMS and fMRI: from ‘virtual lesions’ to functional-

network accounts of cognition. Cortex 45: 1043–1049.

115. Bestmann, S., C.C. Ruff, F. Blankenburg, et al. 2008.

Mapping causal interregional influences with concur-

rent TMS-fMRI. Exp. Brain Res. 191: 383–402.

116. Goebel, R. & N. van Atteveldt. 2009. Multisensory func-

tional magnetic resonance imaging: a future perspec-

tive. Exp. Brain Res. 198: 153–164.

117. Kiebel, S.J., K. von Kriegstein, J. Daunizeau & K.J. Fris-

ton. 2009. Recognizing sequences of sequences. PLoS

Comput. Biol. 5: e1000464.

118. Dominey, P.F. & T. Inui. 2009 Cortico-striatal function

in sentence comprehension: insights from neurophysi-

ology and modeling. Cortex 45: 1012–1018.

119. Rajeev, D.S., R.D. Raizada, F.M. Tsao, et al. 2010. Quan-

tifying the adequacy of neural representations for a

cross-language phonetic discrimination task: predic-

tion of individual differences. Cereb. Cortex . 20: 1–12.

120. Hernandez, A.E. 2009. Language switching in the bilin-

gual brain: what’s next? Brain Lang . 109: 133–140.

121. Bloch, C., A. Kaiser, E. Kuenzli, et al. The age of sec-

ond language acquisition determines the variability in

activation elicited by narration in three languages in

Broca’s and Wernicke’s area. Neuropsychologia 47: 625–

633.

122. Saur, D., A. Baumgaertner, A. Moehring, et al. 2009.

Word order processing in the bilingual brain. Neuropsy-

chologia 47 158–168.

123. Chee, M.W. 2009. fMR-adaptation and the bilingual

brain. Brain Lang . 109: 75–79.

124. Abutalebi, J., M. Tettamanti & D. Perani. 2009. The

bilingual brain: linguistic and non-linguistic skills.

Brain Lang . 109: 51–54.

125. Sakai, K.L., A. Nauchi, Y. Tatsuno, et al. 2009. Distinct

roles of left inferior frontal regions that explain indi-

vidual differences in second language acquisition. Hum.

Brain Mapp. 30: 2440–2452.

126. Weber, K. & P. Indefrey. 2009. Syntactic priming in

German-English bilinguals during sentence compre-

hension. Neuroimage 46: 1164–1172.

127. Zhang, Y., P.K. Kuhl, T. Imada, et al. 2009. Neural sig-

natures of phonetic learning in adulthood: a magne-

toencephalography study. Neuroimage 46: 226–240.

128. Crinion, J.T., D.W. Green, R. Chung, et al. 2009. Neu-

roanatomical markers of speaking Chinese. Hum. Brain

Mapp. 30: 4108–4115.

129. Grogan, A., D.W. Green, N. Ali, et al. 2009. Structural

correlates of semantic and phonemic fluency ability in

first and second languages. Cereb. Cortex. 19: 2690–

2698.

130. Schnur, T.T., M.F. Schwartz, D.Y. Kimberg, et al.

2009. Localizing interference during naming: conver-

gent neuroimaging and neuropsychological evidence

for the function of Broca’s area. Proc. Natl. Acad. Sci.

USA 106: 322–327.

131. Kherif, F., G. Josse, M.L. Seghier & C.J. Price. 2009.

The main sources of intersubject variability in neuronal

activation for reading aloud. J. Cogn.. Neurosci. 21: 654–

668.

132. Seghier, M.L. & C.J. Price. 2009. Dissociating functional

brain networks by decoding the between-subject vari-

ability. Neuroimage 45: 349–259.

133. Richardson, F.M. & C.J. Price. 2009. Structural MRI

studies of language function in the undamaged brain.

Brain Struct. Funct. 213: 511–523.

134. Floel, A., M.H. de Vries, J. Scholz, et al. 2009. White

matter integrity in the vicinity of Broca’s area predicts

grammar learning success. Neuroimage 47: 1974–1981.

135. Petkov, C.I., N.K. Logothetis & J. Obleser. 2009. Where

are the human speech and voice regions, and do other

animals have anything like them? Neuroscientist 15:

419–429.

88 Ann. N.Y. Acad. Sci. 1191 (2010) 62–88 c© 2010 New York Academy of Sciences.