ARTÍCULO DE INVESTIGACIÓN

ARTÍCULO DE INVESTIGACIÓN

Neural Representation of Articulable and
Inarticulable Novel Sound Contrasts:
The Role of the Dorsal Stream

un acceso abierto

diario

David I. Saltzman and Emily B. miers

University of Connecticut, Storrs, CT

Palabras clave: speech perception, resonancia magnética funcional, multivoxel pattern analysis, speech articulation, dorsal stream,
phonetics

ABSTRACTO

The extent that articulatory information embedded in incoming speech contributes to the
formation of new perceptual categories for speech sounds has been a matter of discourse for
décadas. It has been theorized that the acquisition of new speech sound categories requires
a network of sensory and speech motor cortical areas (the “dorsal stream”) to successfully
integrate auditory and articulatory information. Sin embargo, it is possible that these brain regions
are not sensitive specifically to articulatory information, but instead are sensitive to the abstract
phonological categories being learned. We tested this hypothesis by training participants over
the course of several days on an articulable non-native speech contrast and acoustically
matched inarticulable nonspeech analogues. After reaching comparable levels of proficiency
with the two sets of stimuli, activation was measured in fMRI as participants passively listened
to both sound types. Decoding of category membership for the articulable speech contrast
alone revealed a series of left and right hemisphere regions outside of the dorsal stream that
have previously been implicated in the emergence of non-native speech sound categories,
while no regions could successfully decode the inarticulable nonspeech contrast. A pesar de
activation patterns in the left inferior frontal gyrus, the middle temporal gyrus, y el
supplementary motor area provided better information for decoding articulable (speech)
sounds compared to the inarticulable (sine wave) sounds, the finding that dorsal stream
regions do not emerge as good decoders of the articulable contrast alone suggests that other
factores, including the strength and structure of the emerging speech categories are more likely
drivers of dorsal stream activation for novel sound learning.

INTRODUCCIÓN

Whether an infant learning her first language, or an adult learning his fifth language, in lan-
guage acquisition the learner must learn to perceive as well as produce new speech sounds.
In typical language acquisition, these two processes (perceptual and articulatory learning)
develop in tandem, although not necessarily on the same timeline (Tsao, Liu, & kühl,
2004). Sin embargo, in laboratory conditions, we know that individuals can learn to distinguish
complex auditory speech and nonspeech categories without learning to produce these same
categories. At issue is the degree to which articulatory information associated with speech is
recruited even in the absence of articulatory learning. Eso es, when one learns a new pho-
netic category without learning to produce the same category, are speech motor represen-
tations recruited?

Citación: Saltzman, D. I., & miers, mi. B.
(2020). Neural representation of
articulable and inarticulable novel
sound contrasts: The role of the dorsal
stream. Neurobiology of Language,
1(3), 339–364. https://doi.org/10.1162/
nol_a_00016

DOI:
https://doi.org/10.1162/nol_a_00016

Supporting Information:
https://doi.org/10.1162/nol_a_00016

Recibió: 21 Octubre 2019
Aceptado: 23 Puede 2020

Conflicto de intereses: Los autores tienen
declaró que no hay intereses en competencia
existir.

Autor correspondiente:
David I. Saltzman
david.saltzman@uconn.edu

Editor de manejo:
Patrick Wong

Derechos de autor: © 2020 Massachusetts
Institute of Technology. Publicado
bajo una atribución Creative Commons
4.0 Internacional (CC POR 4.0) licencia.

La prensa del MIT

yo

D
oh
w
norte
oh
a
d
mi
d

F
r
oh
metro
h

t
t

pag

:
/
/

d
i
r
mi
C
t
.

metro

i
t
.

mi
d
tu
norte
oh

/

yo
/

yo

a
r
t
i
C
mi

pag
d

F
/

/

/

/

1
3
3
3
9
1
8
6
7
7
4
6
norte
oh
_
a
_
0
0
0
1
6
pag
d

.

/

yo

F

b
y
gramo
tu
mi
s
t

t

oh
norte
0
8
S
mi
pag
mi
metro
b
mi
r
2
0
2
3

Neural decoding of articulable and inarticulable sounds

Models of speech perception like the motor theory of speech (Liberman, Cooper,
Shankweiler, & Studdert-Kennedy, 1967) and subsequent direct realist approaches (Best,
1995; Fowler, 1986), make explicit predictions about the formative role of articulatory codes
(es decir., the gestures made by the articulators to create a speech sound) in speech perception. En
estos modelos, motor or articulatory representations are the objects of perception, and must be
acquired and then accessed to achieve robust comprehension (see Galantucci, Fowler, &
Turvey, 2006 para una revisión). De este modo, learning to discriminate between a pair of speech sounds
like /b/ and /d/ (which are similar acoustically) requires that the listener access information the
articulatory gestures used to produce those sounds.

Though not explicitly linked, the dual stream model of speech processing (Hickok & Poeppel,
2007) makes biological predictions that also implicate motor systems not in speech sound
percepción, but in speech sound learning. Específicamente, the dorsal stream pathway, consisting of
such brain regions as the inferior frontal gyrus (IFG), the pre-motor cortex, and the temporal-
parietal junction, is claimed to play a critical role in auditory-motor integration and serves as a
“neural mechanism that both codes and maintains instances of speech sounds,” which can then
be later used to aid in production as well as in perception (Hickok & Poeppel, 2007, pag. 399).
According to this view, during the learning of new speech sounds, listeners code the incoming
articulatory gestures in dorsal stream areas in preparation for future speech output (Hickok &
Buchsbaum, 2003). Notablemente, once speech sound representations have been learned, dorsal stream
activation is no longer thought to be required during perception-only tasks.

The role of the dorsal stream during speech perception (for established speech categories and
newly learned ones) has been a matter of significant debate, as has the role of articulatory codes
during speech perception in general (p.ej., Lotto, Hickok, & Holt, 2009; Schomers &
Pulvermüller, 2016; wilson, 2009; Scott, McGettigan, & Eisner, 2009). If articulatory codes
are accessed as a matter of course during perception of speech categories, it should follow that
the encompassed brain regions are sensitive to differences in articulation between speech
sounds during speech perception, especially during speech sound learning (see Meister,
wilson, Deblieck, Wu, & Iacoboni, 2007; wilson & Iacoboni, 2006; wilson, Saygin, Sereno,
& Iacoboni, 2004). While evidence for sensitivity to place/manner of articulation differences is
apparent in the left superior temporal gyrus (STG) using electrocorticography (Mesgarani,
Cheung, Johnson, & Chang, 2014), it is less clear that dorsal stream speech motor regions are
involved in the perception of established, mature speech categories. Using a univariate fMRI
acercarse, Pulvermüller et al. (2006) found differential activation in lip- and tongue-dedicated
regions of the precentral gyrus in response to the syllables /p/ and /t/, which differ in place of
articulation. These results were supported by findings from Correia, Jansma, and Bonte
(2015), who used a multivoxel pattern analysis (MVPA) to identify brain regions that could clas-
sify neural responses to stimuli based on specific articulatory features independent of phonemic
categoría (and subsequent acoustic variation). Activation patterns in pre-motor and motor areas
(among other regions) could distinguish place and manner of articulation contrasts, even though
the classifier was trained on responses to one phonological category (stop consonants) y
tested on responses to another (fricatives). Sin embargo, both Arsenault and Buchsbaum (2015)
and Arsenault and Buchsbaum (2016) failed to find any dorsal stream regions that were sensitive
to different articulatory dimensions of passively heard syllables in both univariate (attempting
to replicate Pulvermüller et al.’s results) and multivariate analyses, calling into question the
sensitivity of the dorsal stream to learned phonological contrasts.

As noted above, there is contradictory evidence that dorsal stream regions are sensitive to
articulatory contrasts during speech perception. Sin embargo, even in the studies that support the

Neurobiology of Language

340

yo

D
oh
w
norte
oh
a
d
mi
d

F
r
oh
metro
h

t
t

pag

:
/
/

d
i
r
mi
C
t
.

metro

i
t
.

mi
d
tu
norte
oh

/

yo
/

yo

a
r
t
i
C
mi

pag
d

F
/

/

/

/

1
3
3
3
9
1
8
6
7
7
4
6
norte
oh
_
a
_
0
0
0
1
6
pag
d

/

.

yo

F

b
y
gramo
tu
mi
s
t

t

oh
norte
0
8
S
mi
pag
mi
metro
b
mi
r
2
0
2
3

Neural decoding of articulable and inarticulable sounds

activation of the dorsal stream during speech processing, an alternative explanation exists: Es
possible that what is being observed is IFG sensitivity to abstract sound categories at large, y
not specifically the articulatory differences between sounds. Eso es, during speech perception,
inferior frontal structures may not be coding articulatory differences between sounds (p.ej., el
difference in lip protrusion in the production of /i/ and /y/), but instead reflect the ultimate
product of those gestures, which are two distinct phonetic categories. IFG sensitivity to speech
categories has been established in a number of studies (hutchison, Blumstein, miers, 2008;
Sotavento, Turkeltaub, Granjero, & Raizada, 2012; miers, 2007; miers, Blumstein, Walsh, &
Eliassen, 2009), but these studies do not disentangle the issue of articulatory codes versus ab-
stract categories as the two are fundamentally confounded in established speech categories. En
each of these previous studies, participants were presented with isolated speech sounds in
their native language. These well-established perceptual categories may activate articulatory
representations of the same sound by association rather than as a means to perceive the sound
sí mismo. Eso es, upon hearing “da,” one may automatically activate gestural codes for producing
“da,” whether or not this articulatory information is necessary for retrieving the speech sound
categoría. Más, as argued in the dual stream model, the dorsal stream may only be recruited
during learning, when participants are learning the links between auditory and motor repre-
sentations of sounds. De este modo, to fully test the predictions of this model, it is necessary to present
listeners with a situation in which a new speech category is acquired alongside an acoustically
emparejado, inarticulable contrast in order to fully establish what dorsal stream activation may
actually reflect.

Liebenthal, Binder, Spitzer, Possing, and Medler (2005) and Liebenthal et al. (2010) ex-
plored this idea by contrasting brain activation from familiar speech sounds (a continuum from
/b/ to /d/) with that of an acoustically unfamiliar matched nonspeech continuum (created by
spectrally inverting the first formant of the speech continuum). Although when participants had
minimal exposure to the speech and nonspeech continua (Liebenthal et al., 2005), no dorsal
stream differences emerged, after participants were trained on the same stimuli for four days,
(Liebenthal et al., 2010), some dorsal stream recruitment was seen. That dorsal stream recruit-
ment arises only during the learning process is consistent with the dual stream model, but par-
adoxically, dorsal stream engagement is not preferentially seen for speech (articulable) estímulos.
En cambio, several significant clusters in the left IFG, the bilateral postcentral gyrus, and the bi-
lateral supplementary motor area display greater activation at posttest for a nonspeech contin-
uum that could not be articulated by a human vocal tract. In a conjunction analysis, en el cual
the neural changes as a result of training were compared across the speech and nonspeech
continua, the left IFG was identified as sensitive specifically to learning in the nonspeech con-
tinuum. This provides further evidence that during learning, IFG activation more likely reflects
the emergent category status of the nonspeech stimuli as listeners begin to form perceptual
categories rather than a reflection of articulatory codes (Myers et al., 2009).

Desai, Liebenthal, Waldron, and Binder (2008), used a similar paradigm to Liebenthal et al.
(2005), but critically substituted sine-wave speech versions of both the speech (/b/-/d/) and non-
speech continua (spectrally rotated /b/-/d/). Adopting the labels from this study, we will refer to
them hereafter as “phonemic” and “nonphonemic,” respectively, to differentiate from the nat-
ural speech versions of these stimuli in previous studies. Sine-wave speech is created by taking
the spectral center and power of each of the formants in a speech sound, and then creating a
sine-wave tone at the same frequency and power (Remez, Frotar, Pisoni, & Carrell, 1981). El
resulting sound contains approximately the same distinctive spectral information as the original
speech sound but is not perceived as speech or speechlike until the listener is informed of the
intended message, after which listeners often report being able to understand the message. Estos

Neurobiology of Language

341

yo

D
oh
w
norte
oh
a
d
mi
d

F
r
oh
metro
h

t
t

pag

:
/
/

d
i
r
mi
C
t
.

metro

i
t
.

mi
d
tu
norte
oh

/

yo
/

yo

a
r
t
i
C
mi

pag
d

F
/

/

/

/

1
3
3
3
9
1
8
6
7
7
4
6
norte
oh
_
a
_
0
0
0
1
6
pag
d

/

.

yo

F

b
y
gramo
tu
mi
s
t

t

oh
norte
0
8
S
mi
pag
mi
metro
b
mi
r
2
0
2
3

Neural decoding of articulable and inarticulable sounds

stimuli allowed the authors to control for familiarity and to manipulate the availability of artic-
ulatory information by revealing to participants midway through the experiment that one of the
continua (the phonemic continuum) was made from familiar speech sounds. Dorsal stream
recruitment was observed at posttest, where greater activation in the left IFG and the left precen-
tral gyrus was seen for the phonemic continuum compared to the nonphonemic continuum. Él
seems possible that after being informed that the phonemic continuum was supposed to reflect
two familiar speech sound categories, listeners began to recover some of the articulatory code
used to create those sounds, and as a result the dorsal stream was recruited during learning. Pero
de nuevo, this possibility does not exclude that neural activation reflects the emergence of two
distinct categories, as perception of the phonemic continuum shifted from a continuous mode
to a categorical mode during posttest.

The Current Study

To investigate the role of the dorsal stream during the learning of new speech sounds, we at-
tempted to equate the degree of proficiency with the two stimulus sets being learned, también
as to manipulate articulatory feasibility. While both variables have been manipulated indepen-
dently in previous studies, to our knowledge no study has controlled for both of these variables
simultaneously. We take a similar approach to Liebenthal et al. (2005, 2010) by introducing a set
of articulable (nonnative speech) versus inarticulable (impoverished sine-wave speech) sounds.
Like Desai et al. (2008), we used sounds that did not initially have established categories for our
listeners, therefore requiring participants to learn both sets of sounds (unlike Liebenthal et al.,
2005, 2010). Sin embargo, in Desai et al., articulatory information is only indirectly implied in their
phonemic continuum, as listeners must recover a phonological category from sine-wave
speech. There is no guarantee that participants are activating articulatory representations of
the source stimuli, and results could reflect listeners treating the sine-wave speech as inarticul-
able sounds but with a defined category structure. Por lo tanto, we chose to use a non-native
speech sound contrast to ensure that articulatory information is theoretically available, mientras
the contrast is still as unfamiliar as the impoverished sine-wave speech contrast. Listeners are
then trained to the same accuracy on both contrasts, after which two competing hypotheses
can be investigated. Primero, following the logic of the dual stream model of speech perception,
dorsal stream regions should only be recruited for auditory-motor integration during the learning
of articulable sounds. Alternativamente, dorsal stream regions (especially the IFG) may play a more
general role in categorical decisions and will be equally recruited during the learning of both
articulable and inarticulable sounds (so long as the two sounds are equally well learned).

The question at hand is especially suited to the use of MVPA (Haxby et al., 2001) porque
these techniques can provide a clue as to the nature of the information stored in different areas
of the cortex. If the patterns of activation are distinctive enough to classify stimulus types, él
implies that there is systematic neural representation for some property of that stimulus.
Más, the differences in activation in response to a contrast like /b/ and /t/ may not be lo-
cated in separate areas of the cortex, but may rather be better characterized by a distributed
pattern of activation (including nonadjacent voxels), the kind of informational code that MVPA
is sensitive to. By using signal decoding techniques, we ask not only whether articulatory ver-
sus auditory regions are recruited, but also whether the information content of those regions
distinguishes between the two categories in each contrast.

We attempted to answer this question by using a multiday training study in which participants
were trained to equal degrees of familiarity on an unfamiliar (but articulable) non-native phonetic
contrast (the /i/ vs. /y/ contrast found in languages like French), as well as acoustically matched

Neurobiology of Language

342

yo

D
oh
w
norte
oh
a
d
mi
d

F
r
oh
metro
h

t
t

pag

:
/
/

d
i
r
mi
C
t
.

metro

i
t
.

mi
d
tu
norte
oh

/

yo
/

yo

a
r
t
i
C
mi

pag
d

F
/

/

/

/

1
3
3
3
9
1
8
6
7
7
4
6
norte
oh
_
a
_
0
0
0
1
6
pag
d

/

.

yo

F

b
y
gramo
tu
mi
s
t

t

oh
norte
0
8
S
mi
pag
mi
metro
b
mi
r
2
0
2
3

Neural decoding of articulable and inarticulable sounds

impoverished sine-wave analogues of the same contrast. After three days of training, Participantes
then engaged in a passive fMRI task in which they heard tokens from the two continua they were
trained on. A series of univariate and multivariate analyses were performed on the fMRI data to
determine the neural architecture involved in the learning of new sounds, and how the articu-
latory feasibility of that sound impacts that architecture.

MATERIALS AND METHOD

Participantes

Thirty-four undergraduate students were recruited from the University of Connecticut. All par-
ticipants indicated that they were right-handed, monolingual native English speakers with nor-
mal hearing and no history of neurological or language disorders. Handedness and hearing
abilities were self-reported by the participant, but not assessed in any other way. Eight partic-
ipants were excluded (because of failure to complete all four sessions, norte = 5, incomplete MRI
datos, norte = 2, withdrawal, norte = 1), leaving data from 26 Participantes (ages = 18–22 years, females =
16, males = 10) for analysis. Written informed consent was obtained from every participant in
accordance with the guidelines of the University of Connecticut ethics committee. Participantes
were paid for their participation.

Estímulos

Speech stimuli

The speech stimuli consisted of tokens taken from a seven-step vowel continuum from /i/ (frente
unrounded vowel, found in English in words like “leaf”) to /y/ (front rounded vowel found in
languages such as French). This contrast was chosen because /y/ is not found in American
English phonology and monolingual English speakers consistently struggle to discriminate these
vowels without training (Exacción & Strange, 2008; Darcy et al., 2012).

Natural speech tokens of /y/ and /i/ were produced by a male native speaker of Parisian
Francés, and the /y/ token was then transformed into synthesized speech using the to
KlattGrid function in Praat (campesinos & Weenink, 2020) to allow for precise control over
acoustic properties. F1, F2, and F4 were fixed at the average values derived from the natural
/y/ production, and F3 was systematically lowered in 7 equal Bark scale steps until the F3
value was equivalent to the naturally produced /i/ token’s F3 value, creating a continuum from
/y/ to /i/. F5 was removed as it is uninformative to the contrast of interest. The synthesized /i/
and /y/ endpoints were submitted for pilot testing to five native speakers of French, who ac-
curately identified the intended vowels and deemed them natural sounding. All tokens had a
duration of 432 ms and were scaled to an intensity of 84 dB. Stimuli were delivered over head-
phones at a volume adjusted by the participant to a loud yet comfortable listening level.

Sine-wave stimuli

Nonspeech acoustic analogues of the speech stimuli were created by transforming the vowel
continuum described in the previous section into sine-wave speech. This was achieved using
a Praat script created by Chris Darwin (available at http://www.lifesci.sussex.ac.uk/ home/
Chris_Darwin/Praatscripts/SWS). The resulting seven tokens were then stripped of the sine-wave
equivalents of F1 and F4, leaving only the critical F2–F3 contrast in place to reduce the likeli-
hood that they would be perceived as speech-like (ver figura 1). The resultant two-formant sine
sweeps are thus acoustically matched to the diagnostic information in the vowel continuum
(the frequency of F3), but critically during initial pilot testing (norte = 11) they were never indicated

Neurobiology of Language

343

yo

D
oh
w
norte
oh
a
d
mi
d

F
r
oh
metro
h

t
t

pag

:
/
/

d
i
r
mi
C
t
.

metro

i
t
.

mi
d
tu
norte
oh

/

yo
/

yo

a
r
t
i
C
mi

pag
d

F
/

/

/

/

1
3
3
3
9
1
8
6
7
7
4
6
norte
oh
_
a
_
0
0
0
1
6
pag
d

/

.

yo

F

b
y
gramo
tu
mi
s
t

t

oh
norte
0
8
S
mi
pag
mi
metro
b
mi
r
2
0
2
3

Neural decoding of articulable and inarticulable sounds

Cifra 1. Spectrograms of the first step of the vowel (izquierda) and sine-wave (bien) continua.

to be perceived as speech (participants frequently described them as “chirps,” “bird sounds,” or
“robot sounds”). The same pilot testing also revealed that a single session was sufficient for par-
ticipants to reach near ceiling accuracy in categorizing sine-wave analogue tokens using the
explicit perceptual fading training task described in the following sections (m = 97.7%, DE =
15%). All tokens had a duration of 432 ms and were scaled to an intensity of 80 dB.

Experiment Schedule

The study took place over four consecutive days (ver figura 2); on the first day, participants went
through the informed consent process and completed the training paradigm for the vowel to-
kens. On the second day, participants returned and repeated training for the vowel tokens. On
the third day, participants again repeated the vowel training followed by the sine-wave training
paradigma. Finalmente, on the fourth day participants completed a short refresher on both stimulus
conjuntos, in which participants were refamiliarized with the correspondence between continuum
sounds and the category they map on to, and then completed the fMRI task. The discrepancy
in the number of training sessions between the vowel stimuli and the sine-wave stimuli was a
result of earlier pilot testing, which indicated that it took participants one session to reach a
threshold of at least 80% accuracy in the most difficult training block on the sine-wave stimuli,
but three days to reach the same criteria for the vowel stimuli. The fMRI task was completed on a
fourth day to allow for consolidation of the recently learned sine-wave stimuli, as sleep consol-
idation has been shown to impact non-native phonetic learning (Earle & miers, 2015).

Behavioral Training Regimen

The behavioral training paradigm consisted of three subtasks that were identical in structure
across both Vowel and Sine training. OpenSesame 3.2.4 (Mathôt, Schreij, & Theeuwes, 2012)

Cifra 2.
(A) Schedule of tasks completed over the course of the four days of training. (B) Schedule of the subtasks that comprised both the
vowel and the sine-wave training sessions. (C) Overview of the stimulus pairs used in the discrimination (above the horizontal continuum line)
and the explicit perceptual fading training (below the horizontal continuum line).

Neurobiology of Language

344

yo

D
oh
w
norte
oh
a
d
mi
d

F
r
oh
metro
h

t
t

pag

:
/
/

d
i
r
mi
C
t
.

metro

i
t
.

mi
d
tu
norte
oh

/

yo
/

yo

a
r
t
i
C
mi

pag
d

F
/

/

/

/

1
3
3
3
9
1
8
6
7
7
4
6
norte
oh
_
a
_
0
0
0
1
6
pag
d

/

.

yo

F

b
y
gramo
tu
mi
s
t

t

oh
norte
0
8
S
mi
pag
mi
metro
b
mi
r
2
0
2
3

Neural decoding of articulable and inarticulable sounds

was used for stimulus presentation, and all responses were made via a Cedrus RB-840 response
box. Participants wore Sony MDR-7506 headphones and were allowed to adjust volume to
their most comfortable listening level. All response options were counterbalanced across
Participantes.

Discrimination

At both the beginning and the end of the behavioral training paradigm, participants completed
an identical AX discrimination task in which they were presented with a pair of tokens from the
appropriate stimulus set that were either identical (p.ej., step 1 vs. step 1) or two steps apart on
the continuum (p.ej., steps 1 vs. step 3; steps 4 vs. step 6) and asked to indicate as quickly and
accurately as possible if the two sounds were the same or different. All pairs were presented
both “forwards” (p.ej., step 1 followed by step 3) or “backwards” (p.ej., step 3 followed by step 1),
with an intertrial interval of 500 EM. There were seven pair types where the two tokens pre-
sented were identical and ten pair types where the stimuli differed. Each pair was presented
11 veces, for a total of 187 ensayos. Accuracy of response and reaction time were both recorded
0
as outcomes of interest, and d
, an unbiased measure of sensitivity, was calculated from the
participants’ responses. Trials in which participants failed to respond before the trial timed out
(6,000 EM) were considered incorrect responses.

Perceptual fading training

Explicit training on the sound categories used a perceptual fading paradigm where participants
were asked to map the two halves of the continuum to a different colored square. On each
successive block, the acoustic difference between the two tokens that participants were asked
to identify was reduced. To familiarize participants with the two categories they would be
hearing, they were first shown a colored square and played 10 repetitions of the continuum
endpoint token that would thereon be arbitrarily associated with that colored square (p.ej., a
red square would appear on the screen and the /i/ endpoint token would play).

In the first training block (labeled as “easy”), participants were asked to identify as quickly
and accurately as possible which colored square corresponded to step 1 and step 7. Sobre el
second training block (labeled as “medium”), participants were asked to identify which col-
ored square corresponded to continuum step 2 and continuum step 6. On the third training
block (labeled as “hard”), participants were asked to identify which colored square corre-
sponded to continuum step 3 and continuum step 5, the tokens that lay on opposing sides
of the acoustic boundary. Notice that step 4, the midpoint of the continuum, was not presented
durante el entrenamiento.

Each block consisted of 60 ensayos (30 repetitions of each token), and participants were re-
quired to achieve an 80% accuracy threshold in order to ascend to the next block. If partic-
ipants did not reach threshold by the end of the block, they would repeat that same block up to
three times before being forced to move on to the subsequent block. Por lo tanto, a given par-
ticipant could experience between one and three repetitions of each block, depending on how
quickly they advanced to more difficult blocks. Participants received explicit feedback during
la tarea. Accuracy of response and reaction time were both recorded as outcomes of interest.

Continuum categorization

Following training, participants engaged in a two-alternative forced choice task in which they
were asked to identify the colored square that corresponded to the sound being presented, con

Neurobiology of Language

345

yo

D
oh
w
norte
oh
a
d
mi
d

F
r
oh
metro
h

t
t

pag

:
/
/

d
i
r
mi
C
t
.

metro

i
t
.

mi
d
tu
norte
oh

/

yo
/

yo

a
r
t
i
C
mi

pag
d

F
/

/

/

/

1
3
3
3
9
1
8
6
7
7
4
6
norte
oh
_
a
_
0
0
0
1
6
pag
d

/

.

yo

F

b
y
gramo
tu
mi
s
t

t

oh
norte
0
8
S
mi
pag
mi
metro
b
mi
r
2
0
2
3

Neural decoding of articulable and inarticulable sounds

no feedback. Había 10 randomly presented repetitions of each token from the full seven-
step continuum, for a total of 70 ensayos. The proportion of step 7 respuestas (which corresponded
to the /y/Vowel and /y/Sine endpoints in the Vowel and Sine conditions respectively) per contin-
uum step were then calculated. Trials in which participants failed to respond were omitted from
análisis.

Imaging Methods

fMRI design and procedure

Before entering the scanner, participants completed a brief task in which they were refamiliar-
ized with the stimuli that would be presented during the fMRI task. Participants were first pre-
sented with a colored square while five repetitions of each of two tokens from that category
were played. Following this, participants engaged in a two-alternative forced choice (2AFC)
task in which they were asked to identify which colored square corresponded to the sound
presentado. Each of the four tokens were repeated 30 veces, and this was repeated for both
the Vowel tokens and the Sine tokens, leading to a total of 240 trials for the 2AFC task.
Explicit feedback was provided during this task.

Participants were instructed to listen attentively and told that on some trials a stimulus
would be presented more quietly than normal. When this happened, they should press any
button on the MRI-safe response devices held in both hands. These catch stimuli consisted
of a randomly chosen token from the eight Vowel and Sine tokens presented at 20 dB lower
intensidad. Participants completed a brief training during the structural scan acquisition to famil-
iarize them with this task. Imaging data from catch trials were modeled in participant-level
analyses but did not contribute to group-level analyses. Participants were on average quite
accurate at detecting catch trials (m = 90%, DE = 30%). Stimuli were delivered by
OpenSesame 3.2.4. The volume was set to a comfortable listening level as determined by each
partícipe.

The fMRI experimental paradigm consisted of 10 carreras, with run order randomized between
Participantes. Trials within each run were presented in a fixed, pseudorandom order determined
using the OptSeq2 tool (https://surfer.nmr.mgh.harvard.edu/optseq/). Steps 1, 3, 5, y 7 de
both the Vowel and Sine continua were each repeated 15 times per run, junto con 15 estímulos
used for the catch task described in the previous paragraph, Resultando en 150 presentations of
each stimulus over the course of the experiment. An auditory stimulus appeared in all but 25
of the silent gaps during scanning, for an average intertrial interval of 2,390 EM. Each run con-
sisted of 161 volumes.

Acquisition parameters

Anatomical and functional MRI data were collected with a 3-T Siemens Prisma scanner
(Erlangen, Alemania). High resolution three dimensional T1-weighted anatomical images were
acquired using a multiecho magnetization prepared rapid gradient echo sequence (MPRAGE;
repetition time [TR] = 2,400 EM, echo time = 2.22 EM, inversion time = 1,000 EM, 0.8 mm3
isotropic voxels, 320 × 320 matrix) and reconstructed into 208 slices. Functional echo planar
images were acquired using an accelerated multiband sequence in ascending, interleaved
orden (52 slices, 2.5 mm thick, 2 mm2 axial in-plane resolution, 110 × 110 matrix, 220 mm3
field of view, flip angle = 62°) and followed a fast, sparse sampling design: Each functional
volume was acquired with a 1,000 ms acquisition time, seguido por 1,000 ms of silence during

Neurobiology of Language

346

yo

D
oh
w
norte
oh
a
d
mi
d

F
r
oh
metro
h

t
t

pag

:
/
/

d
i
r
mi
C
t
.

metro

i
t
.

mi
d
tu
norte
oh

/

yo
/

yo

a
r
t
i
C
mi

pag
d

F
/

/

/

/

1
3
3
3
9
1
8
6
7
7
4
6
norte
oh
_
a
_
0
0
0
1
6
pag
d

/

.

yo

F

b
y
gramo
tu
mi
s
t

t

oh
norte
0
8
S
mi
pag
mi
metro
b
mi
r
2
0
2
3

Neural decoding of articulable and inarticulable sounds

which auditory stimuli were presented (effective TR = 2,000 EM). Stimuli were always presented
during the silent gap.

Univariate fMRI analyses

Data were analyzed using Analysis of Functional NeuroImages (AFNI; Cox, 1996). Preprocesamiento
for the univariate analyses consisted of first transforming images from oblique to cardinal orienta-
ción, followed by motion correction using a six-parameter rigid body transform aligned with each
participant’s anatomical data set, normalization to Talairach space (Talairach & Tournoux, 1988),
and spatial smoothing with a 4 mm Gaussian kernel. Masks were created using each participant’s
anatomical data to eliminate voxels located outside the brain. Individual masks were used to gen-
erate a group mask. Motion and signal fluctuation outliers were removed following standard
procedures.

We generated time series vectors for each of the eight continua tokens ( Vowel step 1,
Vowel step 3, Vowel step 5, Vowel step 7, Sine step 1, Sine step 3, Sine step 5, Sine step 7) como
well as the catch trials and any false alarm trials for each participant in each run. These vectors
contained the onset time of each stimulus and were convolved with a stereotypic gamma hemo-
dynamic function. El 10 condition vectors along with six additional nuisance movement
parameters were submitted to a regression analysis. This analysis generated by-voxel fit coeffi-
cients for each condition for each participant.

For group level analysis, beta coefficients were submitted to a 2 × 3 analysis of variance
(ANOVA) in AFNI (using 3dANOVA3). In the ANOVA, main effects of sound type ( Vowel vs.
Sine), continuum step (1 vs. 3 vs. 5 vs. 7), and their interaction were estimated.

Multivariate fMRI analyses

Data were preprocessed again using AFNI with a modified pipeline compared to the univariate
analiza. Específicamente, multivariate preprocessing consisted of first transforming images from
oblique to cardinal orientation, followed by motion correction using a six-parameter rigid
body transform aligned with each participant’s anatomical data set. No blurring or spatial nor-
malization were performed until after the participant-level analysis to preserve fine-grained
spatial information. The same 10 condition vectors from the univariate analyses along with
six additional nuisance movement parameters were submitted to an individually modulated
regression analysis (using the stim_times_IM flag in AFNI’s 3dDeconvolve), in which each trial
for a condition is given its own amplitude estimation, yielding trial-wise beta coefficients (como
opposed to run-wise beta coefficients, see Rissman, Gazzaley, & D'Esposito, 2004). The ma-
trices from the output of the individually-modulated regression were then submitted to a least
squares sum estimation using AFNI’s 3dLSS, in accordance with best practices for performing
multivariate analyses on fast event-related fMRI designs (Mumford, Tornero, Ashby, & Poldrack,
2012; Mumford, davis, & Poldrack, 2014).

The beta coefficient output from 3dLSS was then used in an MVPA performed in MATLAB
using the Decoding Toolbox (TDT; Hebart, Görgen, & Haynes, 2015). A linear support vector
machine (SVM) classification was used with 3 voxel (6 mm) radius searchlight and a leave-
one-run-out cross validation scheme, in which the SVM was trained on data from nine runs
and tested on a tenth, held-out run. The regularization parameter for the SVM was set at C = 1.
In a searchlight analysis, a roving sphere moves across the whole brain, and in each search-
light sphere an SVM was trained and tested (Kriegeskorte, Goebel, & Bandettini, 2006). El
classification accuracy score was assigned to the central voxel of each searchlight. Este
yielded a participant-level classification accuracy map for each participant in their brain’s

Neurobiology of Language

347

yo

D
oh
w
norte
oh
a
d
mi
d

F
r
oh
metro
h

t
t

pag

:
/
/

d
i
r
mi
C
t
.

metro

i
t
.

mi
d
tu
norte
oh

/

yo
/

yo

a
r
t
i
C
mi

pag
d

F
/

/

/

/

1
3
3
3
9
1
8
6
7
7
4
6
norte
oh
_
a
_
0
0
0
1
6
pag
d

.

/

yo

F

b
y
gramo
tu
mi
s
t

t

oh
norte
0
8
S
mi
pag
mi
metro
b
mi
r
2
0
2
3

Neural decoding of articulable and inarticulable sounds

native space. The following classifications were performed: /i/ versus /y/ for vowel tokens (defined
as step 1 and step 3 vs. step 5 and step 7, hereafter “/i/Vowel vs. /y/Vowel
"), and the analogous
sine version (which we will hereafter refer to as “/i/Sine vs. /y/Sine

").

In preparation for group-level analyses, participant-level classification maps were then
normalized to Talairach space and blurred with a 4 mm Gaussian kernel. Following this,
single-sample t tests against chance levels were performed for the /i/Vowel vs. /y/Vowel and
/i/Sine vs. /y/Sine (using 3dttest++ in AFNI): t test output was masked by a group mask which
only included voxels present in all 26 Participantes. Además, a paired-samples t test was
performed comparing the /i/Vowel vs. /y/Vowel and /i/Sine vs. /y/Sine decodings.

Cluster level correction for multiple comparisons was determined by running 10,000 iter-
ations of Monte Carlo simulations on the group mask. Específicamente, we used the -acf option in
3dFWHMx and 3dClustSim (AFNI) to estimate the spatial smoothness and generate voxel-wise
and cluster-wise inferences. Data were corrected at a cluster level correction of p < 0.05 (voxel level threshold of p < 0.025, 149 contiguous voxels). RESULTS Behavioral Data Analysis and Results Vowel training results All the following behavioral analyses were conducted using the R packages afex and lme4 (Bates, Mächler, Bolker, & Walker, 2015; Singmann, Bolker, Westfall, Aust, & Ben-Shachar, 2020). Model selection was achieved by first creating and running all possible permutations of random effects structures, after which a backwards stepping selection procedure was applied to the models that converged (Matuschek, Kliegl, Vasishth, Baayen, & Bates, 2017). Output from the ANOVA table generated by the afex command “mixed” (run on the appropriate model selected from the earlier described procedure) is reported for all mixed-effects model analyses in this study. Residuals were inspected for each selected model with no deviations from normal encountered. Perceptual fading training. On average, accuracy during training improved over the course of the three days (see Table 1, Figure 3A). Given the adaptive nature of the paradigm, participants completed varying repetitions of each level of difficulty (easy: M = 1.29 blocks, SD = 0.58; medium: M = 1.28 blocks, SD = 0.58; hard: M = 2.12 blocks, SD = 0.90), and as expected there was a strong negative correlation between the number of blocks completed and average accuracy on the explicit perceptual fading training (r(24) = −0.92, p < 0.0001), reflecting the fact that less proficient learners had to complete more blocks to progress through training. Accuracy data during training was submitted to a generalized linear mixed-effects model using the R packages afex and lme4 (Bates, Mächler, Bolker, & Walker, 2015; Singmann, Bolker, Westfall, Aust, & Ben-Shachar, 2020). Model selection was achieved by first creating and running all possible permutations of random effects structures, after which a backwards stepping selection procedure was applied to the models that converged (Matuschek, Kliegl, Vasishth, Baayen, & Bates, 2017). Output from the ANOVA table generated by the afex com- mand “mixed” (run on the appropriate model selected from the earlier described procedure) is reported for all mixed-effects model analyses in this study. To examine the degree to which participants’ accuracy during vowel training increased over time, a generalized linear mixed-effects model was fit The selected model included fixed Neurobiology of Language 348 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . e d u n o / l / l a r t i c e - p d f / / / / 1 3 3 3 9 1 8 6 7 7 4 6 n o _ a _ 0 0 0 1 6 p d / . l f b y g u e s t t o n 0 8 S e p e m b e r 2 0 2 3 Neural decoding of articulable and inarticulable sounds Table 1. Accuracy during explicit perceptual fading training Training Day 1: Vowel Easy (1–7) M = 87.7% (SD = 33%) Medium (2–6) M = 89% (SD = 31%) Hard (3–5) M = 73.9% (SD = 43.9%) Day 2: Vowel M = 96.2% (SD = 19%) M = 93.1% (SD = 25.3%) M = 76.9% (SD = 42.2%) Day 3: Vowel M = 95.7% (SD = 20.3%) M = 93.3% (SD = 25%) M = 79.4% (SD = 40.5%) Day 3: Sine M = 93.7% (SD = 24.3%) M = 94.9% (SD = 22.1%) M = 83.9% (SD = 36.7%) Note: Means (M) and standard deviations (SD) displayed. effects for training difficulty level, training day, and the interaction of training day and level, as well as by-subject random slopes and intercepts for the interaction between training difficulty level and training day. The model revealed a significant main effect of training difficulty ( χ2 = 89.27, p < 0.0001) and a main effect of training day ( χ2 = 7.45, p = 0.02), with higher accu- racy on easier blocks and higher accuracy on successive days of training The interaction be- tween training difficulty and training day did not reach significance ( p = 0.62). l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . e d u n o / l / l a r t i c e - p d f / / / / 1 3 3 3 9 1 8 6 7 7 4 6 n o _ a _ 0 0 0 1 6 p d / . l f b y g u e s t t o n 0 8 S e p e m b e r 2 0 2 3 Figure 3. Behavioral data from the vowel training paradigm. (A) Accuracy on the explicit perceptual fading training task for the vowel stimuli as a function of training level difficulty. Each point indicates an individual participant’s performance, and the shape of the point indicates the –/y/vowel con- number of times that participant repeated that block. (B) Responses to the continuum categorization task as a function of /i/vowel 0 tinuum step and training day. Error bars indicate standard error. (C) d scores from the AX discrimination task for the vowel stimuli across each 0 stimulus as a function of pretest and posttest as well as training day. Error bars indicate standard error. (D) d scores from the AX discrimination task for vowel stimuli with respect to the participant-specific determination of the between-category (BC) and within-category ( WC) tokens as a function of pretest and posttest as well as training day. Error bars indicate standard error. Neurobiology of Language 349 Neural decoding of articulable and inarticulable sounds In general, participants were successful at identifying the end- Vowel categorization posttest. points of the continuum, with at least 90% accuracy on average across each of the three train- ing days (see Figure 3B). Responses from the categorization posttest were transformed into a binary outcome variable and then submitted to a generalized linear mixed-effects model with a logit link function (also instantiated in the R package afex) to assess how categorization of the full continuum changed over the course of the three training days. P values were estimated using a likelihood-ratio test. The model selected included fixed effects for continuum step (centered) and training day, with by-subject random slopes and intercepts for continuum step, training day, and their interaction. Output from the model indicated a main effect of contin- uum step ( χ2 = 44.15, p < 0.0001) and a main effect of training day ( χ2 = 6.08, p = 0.05), indicating that participants’ /y/Vowel responses increased as a result of the F3 manipulation across the continuum as expected, and that participants’ overall proportion of /y/Vowel re- sponses decreased with more experience with the continuum. The interaction was not signif- icant ( p = 0.82), indicating no significant change in the steepness of the categorization function over the three days of training. AX discrimination task. Discrimination data was first separated into pretest (the first discrimina- tion task completed during that day’s training) and posttest (the final discrimination task follow- ing that day’s continuum categorization task), and “different” trials were collapsed such that discrimination token “step 1–step 3” was considered the same as “step 3–step 1.” Following was calculated for each of the resulting five “different” tokens for each subject and this, d training day. For the purposes of analysis, the token “3–5” was considered to be “between- category” (as step 4 of the continuum represents the most ambiguous token in terms of acoustics), and all other tokens were considered “within-category” (Figure 3C). 0 0 A linear mixed-effects model was then fit for participants’ d scores, with fixed effects of token type (between-category vs. within-category), training day, block (pretest vs. posttest) and their interactions, as well as by-subject random intercepts. A main effect of token type 0 was found ( χ2 = 12.30, p < 0.001), in that d scores were higher for the between-category token (step 3 – step 5) than the within-category tokens. In addition, a main effect of training day ( χ2 = 3.09, p = 0.05) arose, indicating that d scores increased over the course of the experiment. All other main effects and interactions were nonsignificant. 0 Using the acoustic boundary to determine the between-category discrimination token ig- nores the fact that participants’ perceptual boundary did not necessarily correspond to the midpoint of the continuum. To remedy this concern, we estimated participant-relative percep- tual boundaries and fit a linear mixed-effects model using the participant-relative determina- tions of the between-category and within-category tokens. The model selected contained fixed effects of token type (between-category vs. within-category), training day, block (pretest vs. posttest) and their interactions, as well as by-subject random intercepts. This model was then compared to the model previously fit using the acoustic boundary to determine whether a token was between-category or within-category. The model that used participant-relative boundaries had superior fit but yielded largely the same findings: a main effect of token type ( χ2 = 38.76, p < 0.0001), a main effect of training day ( χ2 = 5.19, p < 0.01), and a main effect of block ( χ2 = 6.78, p < 0.01). All interactions were nonsignificant. Sine training results Explicit perceptual fading training. Participants again completed varying repetitions of each level of difficulty (easy: M = 1.27 blocks, SD = 0.60; medium: M = 1.35 blocks, SD = 0.56; hard: Neurobiology of Language 350 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . e d u n o / l / l a r t i c e - p d f / / / / 1 3 3 3 9 1 8 6 7 7 4 6 n o _ a _ 0 0 0 1 6 p d . / l f b y g u e s t t o n 0 8 S e p e m b e r 2 0 2 3 Neural decoding of articulable and inarticulable sounds M = 1.69 blocks, SD = 0.84), and there was a similarly strong negative correlation between the number of blocks completed and the average accuracy on the explicit perceptual fading training, r(7) = −0.86, p = 0.003. To examine the efficacy of the single training session on identification of the two Sine tokens, a generalized linear mixed-effects model was performed. The model selected included fixed effects for training difficulty level and by-subject random intercepts (the maximal model for this analysis). The model revealed a significant main effect of training difficulty ( χ2 = 102.98, p < 0.0001), indicating higher accuracy on easier blocks consisting of larger intervals on the continuum. Sine AX discrimination task. Discrimination data for the sine-wave stimuli was analyzed using the same procedure as the vowel stimuli. A linear mixed-effects model was then used for par- ticipants’ d scores, with fixed effects of token type (between-category vs. within-category), block (pretest vs. posttest), and their interaction, as well as by-subject random slopes and intercepts for all fixed effects and interactions. 0 The model revealed that there was no significant difference in d between the between- category and within-category tokens ( p = 0.26). However, there was a main effect of block, 0 scores overall increased from pretest to posttest (F = 9.43, p < 0.01). The interac- in which d tion between token type and block was significant (F = 5.60, p < 0.05), indicating a larger increase in sensitivity at posttest for the between-category token than the within-category token. Recoding the data to use participant-relative determinations of the between-category token did not alter the results (Figure 4D). 0 Comparison of vowel and sine-wave behavioral performance Explicit perceptual fading training. Accuracy on the training task was comparable in sine-wave and vowel tokens, despite participants only having a single day of training (see Table 1, Figure 4A), which was expected from previous pilot testing. Data from the third day of training was analyzed with a generalized linear mixed-effects model with logit link function (the de- pendent variable, accuracy, was coded as binary value for each trial). Fixed effects of training difficulty level, stimulus type ( Vowel vs. Sine), and their interaction, along with by-subject random slopes and intercepts for all fixed effects and their interactions were entered into the model. P values were estimated using a likelihood ratio test. The model output yielded a significant main effect of block ( χ2 = 29.33, p < 0.0001), indicating that participants’ accu- racy decreased as training difficulty increased, regardless of stimulus type, as well as a main effect of stimulus type ( χ2 = 4.39, p = 0.04), in that accuracy was higher for the Sine stimuli compared to the Vowel stimuli. The interaction between block and stimulus type was nonsig- nificant ( p = 0.07). Categorization posttest. Participants’ accuracy at identifying the endpoints of the Sine con- tinuum was not as robust as with the Vowel continuum, but still reached at least 75% ac- curacy at each endpoint token (Figure 4B). Posttest categorization accuracy was analyzed using a generalized linear mixed-effects model where categorization of the Vowel (on the third training day) and Sine continua were compared. Fixed effects of continuum step, con- tinuum type (Sine vs. Vowel), and their interactions, as well as random by-subject slopes and intercepts for all fixed effects and interactions were entered into the model. P values were estimated using a likelihood ratio test. There was a significant effect of continuum step ( χ2 = 42.19, p < 0.0001), but the effect of continuum type was nonsignificant ( p = 0.34). However, there was a significant continuum step by continuum type interaction ( χ2 = 6.17, Neurobiology of Language 351 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . e d u n o / l / l a r t i c e - p d f / / / / 1 3 3 3 9 1 8 6 7 7 4 6 n o _ a _ 0 0 0 1 6 p d / . l f b y g u e s t t o n 0 8 S e p e m b e r 2 0 2 3 Neural decoding of articulable and inarticulable sounds l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . e d u n o / l / l a r t i c e - p d f / Figure 4. Behavioral data from the sine-wave training paradigm. (A) Accuracy on the explicit perceptual fading training task for the sine-wave stimuli as a function of training level difficulty. Each point indicates an individual participant’s performance, and the shape of the point indicates –/y/Sine continuum step. Error bars the number of times that participant repeated that block. (B) Categorization data as a function of the /i/Sine indicate standard error. (C) d scores from the AX discrimination task for the sine-wave stimuli across each stimulus as a function of pretest 0 and posttest. Error bars indicate standard error. (D) d scores from the AX discrimination task for sine-wave stimuli with respect to the participant- relative determination of the between-category (BC) and within-category ( WC) tokens as a function of pretest and posttest. Error bars indicate standard error. 0 / / / 1 3 3 3 9 1 8 6 7 7 4 6 n o _ a _ 0 0 0 1 6 p d . / l f b y g u e s t t o n 0 8 S e p e m b e r 2 0 2 3 p = 0.01). Post hoc pairwise comparisons were performed using the R package lsmeans to investigate the source of the continuum step by continuum type interaction. There were significant differences (after Bonferroni correction) between the categorization of sine-wave and vowel stimuli at step 1 ( p = 0.025), step 5 ( p = 0.012), step 6 ( p = 0.005), and step 7 ( p = 0.004), which reflects the lesser degree of certainty at the endpoints for the sine-wave continuum. Another metric of differences in the mode of categorization between the two stimulus sets (i.e., categorical perception vs. more continuous perception) is the slope of the categorization function. To this end, psychometric functions were fit to participants’ continuum categoriza- tion responses from the third day of training for both the sine-wave stimuli and the vowel stim- uli using the package quickpsy (Linares & López-Moliner, 2016) in R. The slope of the psychometric function was used as a tool to quantify how strongly categorical an individual’s perception is, similar to the “categorical index” approach in Liebenthal et al. (2005, 2010), with the assumption that a steeper slope is among the behavioral hallmarks of more rigidly categorical perception. A linear mixed-effects model was then fit to the slopes of the psycho- metric functions, with a fixed effect of stimulus type ( Vowel vs. Sine-wave) and random by- subject intercepts. There was no significant difference in slope values between stimulus sets (F = 0.55, p = 0.47). Neurobiology of Language 352 Neural decoding of articulable and inarticulable sounds Finally, to achieve converging evidence regarding the mode of categorization between the vowel and sine-wave stimuli, we also analyzed participants’ reaction times in the phonetic categorization task, as categorical perception is usually accompanied by slower responses for more ambiguous tokens in the center of a continuum than the clearer endpoint tokens (Pisoni & Tash, 1974). We again used responses only from the third day of training. To account for overall slower responses to the sine-wave stimuli compared to the vowel stimuli (sine- wave: M = 638.83 ms, SD = 577.35 ms, vowel: M = 473.64 ms, SD = 464.99 ms), we con- verted reaction times to z-scores. A linear mixed-effects model was fit to the z-scored reaction times, with main effects of stimulus type ( Vowel vs. Sine-wave), continuum step (squared, to better reflect the quadratic shape of participants’ responses), and their interaction. The model selected included by-subject random slopes and intercepts for stimulus type and continuum step. There was no significant main effect of stimulus type ( p = 0.93) or continuum step ( p = 0.16), and critically no interaction between stimulus type and continuum step ( p = 0.85), in- dicating that the pattern of responses for each step of the continuum, regardless of stimulus set, was not significantly different (see Figure 5). AX discrimination task. To evaluate the behavioral equivalence across stimulus sets, the AX discrimination data from the third day of training for the vowel stimuli was combined with the data from the sine-wave stimuli and analyzed using a linear mixed-effects model on the 0 d scores. The model selected included main effects of token type (between-category vs. within- category), block (pretest vs. posttest), stimulus set ( Vowel vs. Sine), and their interactions, as well as by-subject random slopes and intercepts for each main effect. A main effect of token type was found (F = 4.53, p < 0.05), indicating higher overall sen- sitivity for between-category tokens than within-category tokens, as well as a main effect of 0 block (F = 6.58, p < 0.05), indicating higher d scores at posttest than pretest, and a main effect of stimulus set (F = 21.68, p < 0.0001), with participants showing better overall sensitivity to sine-wave stimuli. In addition, a significant interaction of token type and stimulus set was found (F = 12.06, p < 0.001), as well as a significant interaction between block and stimulus set (F = 9.72, p < 0.01). Post hoc pairwise comparisons on both interactions were again performed using the R package lsmeans with Bonferroni correction. The interaction between token type and stimulus set was driven by a significant difference between within-category and between-category Figure 5. Reaction time (z-scored) in the continuum categorization task on the third day of training as a function of continuum step and stimulus set. Error bars indicate standard error. Neurobiology of Language 353 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . e d u n o / l / l a r t i c e - p d f / / / / 1 3 3 3 9 1 8 6 7 7 4 6 n o _ a _ 0 0 0 1 6 p d . / l f b y g u e s t t o n 0 8 S e p e m b e r 2 0 2 3 Neural decoding of articulable and inarticulable sounds Figure 6. Regions that showed greater activation for sine-wave tokens than vowel tokens in the ANOVA. Clusters corrected at the voxel level of p < 0.025, with 149 contiguous voxels and a cor- rected threshold of p < 0.05. tokens in the vowel data ( p = 0.001), but not the sine-wave data ( p = 0.34), while the inter- action between block and stimulus set reflected an overall increase in d from pretest to posttest in the sine-wave condition ( p = 0.0004) but no difference in the vowel condition ( p = 0.51). 0 fMRI results Imaging data from Sine and Vowel tokens were pooled in two analyses; a univariate approach to look for differences in activation magnitude, and a multivariate approach to investigate dif- ferences in activation patterns. Univariate analysis. Contrasting functional activation for vowel tokens compared to sine-wave tokens (Figure 6, Table 2) revealed three prominent clusters, all showing more activation for sine-wave tokens compared to vowel tokens. Significant activation differences were found in a large right hemisphere cluster in the STG extending into the Rolandic operculum, the insula, the temporal pole, and Heschl’s gyrus; an analogous left hemisphere cluster in the STG ex- tending into Heschl’s gyrus, the Rolandic operculum, and the temporal pole and a smaller cluster in the left cerebellum. To investigate how individual differences in proficiency with the two continua may have impacted neural results, a linear mixed-effects model (with fixed effects of stimulus type and continuum step, and by-subject random intercepts) was performed on the functional data using AFNI’s 3dLME. Accuracy on the hardest level of difficulty for the explicit perceptual fading training for both the vowel tokens and sine-wave tokens on the final day of training was selected as a covariate. After controlling for participants’ proficiency with the two Table 2. Results of univariate ANOVA comparing activation to Sine versus Vowel stimuli Area Sine > Vowel

L STG, L RO

R STG, R RO

L Cere (Crus 2), L Cere (Crus 1)

Cluster size ( Voxels)

1,807

1,777

150

X

−45

47

−15

y

−11

−9

−69

z

Maximum t value

6

6

−34

7.85

9.55

3.90

Nota. L = left; R = right; STG = superior temporal gyrus; RO = Rolandic operculum; Cere = cerebellum. Clusters corrected at the voxel level of p < 0.025, with 149 contiguous voxels and a corrected threshold of p < 0.05. Neurobiology of Language 354 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . e d u n o / l / l a r t i c e - p d f / / / / 1 3 3 3 9 1 8 6 7 7 4 6 n o _ a _ 0 0 0 1 6 p d . / l f b y g u e s t t o n 0 8 S e p e m b e r 2 0 2 3 Neural decoding of articulable and inarticulable sounds continua, similar bilateral STG clusters emerged in which sine-wave tokens elicited greater activation than vowel tokens. Multivariate analysis. Participant-level classification accuracy maps from the searchlight MVPA analysis were submitted to single sample t tests against chance level (50%) to determine cortical regions that were able to discriminate category differences (e.g., all “A” tokens vs. all “B” tokens). For the decoding of the two vowel categories (/i/Vowel vs. /y/Vowel), three significant clusters were found: (1) a cluster in the right middle temporal gyrus (MTG) extending into the right STG, (2) a cluster in the right IFG (specifically in the pars triangularis and the pars oper- cularis), and (3) a cluster in left MTG (see Table 3, Figure 7A). The same analysis for the sine- wave speech tokens, decoding /i/Sine versus /y/Sine did not yield any regions that met the thresh- old for significance, but voxel-level decoding results from the searchlight analysis contributed to the Vowel decoding−Sine decoding comparison below. Next, we contrasted the sine-wave and vowel decoding maps to identify regions that dif- fered significantly in decoding performance for the two stimulus sets. Participant-level classi- fication accuracy maps from the sine-wave and vowel decoding analyses were submitted to a paired-samples t test to identify regions that showed significantly different decoding perfor- mance for Vowel category differences than for Sine category differences. All such clusters showed superior decoding of Vowel contrasts and included a cluster that extended from the l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . e d u n o / l / l a r t i c e - p d f / Table 3. Results of t tests performed on MVPA classification data Area /i/vowel vs. /y/vowel decoding Cluster size (voxels) x R MTG, R STG R IFG L MTG /i/Sine vs. /y/Sine decoding No significant clusters Vowel decoding−Sine decoding L Thalamus L Insula, L TP, L RO, L IFG L MTG R SMA, R MCC, L SMA R Cere ( VI), R Cere (IV-V) L Cere (IV-V), Cere Verm (4/5) R HC, R Put L Cere ( VI), L Cere (Crus 1) 200 181 159 – 422 404 288 227 191 186 171 162 y −7 19 −43 47 49 −53 – – −21 −33 −47 9 15 −3 27 −23 −17 −5 −59 −17 −53 −57 −29 −55 / / / 1 3 3 3 9 1 8 6 7 7 4 6 n o _ a _ 0 0 0 1 6 p d / . l f b y g u e s t t o n 0 8 S e p e m b e r 2 0 2 3 z −12 30 2 – 6 20 6 44 −22 −8 0 −26 Peak accuracy 51.64% 51.66% 51.47% – Maximum t value 3.04 2.94 2.87 2.93 2.97 3.18 3.06 2.93 Note. L = left; R = right; MTG = middle temporal gyrus; STG = superior temporal gyrus; IFG = inferior frontal gyrus; TP = temporal pole; RO = Rolandic operculum; SMA = supplementary motor area; MCC = middle cingulate cortex; Cere = cerebellum; HC = hippocampus; Put = putamen; Cere Verm = cerebral vermis. Clusters corrected at the voxel level of p < 0.025, with 149 contiguous voxels yielding a corrected threshold of p < 0.05. Neurobiology of Language 355 Neural decoding of articulable and inarticulable sounds l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . e d u n o / l / l a r t i c e - p d f / / / / 1 3 3 3 9 1 8 6 7 7 4 6 n o _ a _ 0 0 0 1 6 p d / . l f b y g u e s t t o n 0 8 S e p e m b e r 2 0 2 3 Figure 7. Results from the multivoxel pattern analyses. (A) Regions that showed significant classi- fication of the two Vowel categories. Clusters corrected at the voxel level of p < 0.025, with 149 contiguous voxels and a corrected threshold of p < 0.05. (B) Output from the paired t test showing regions that classified Vowel categories significantly more accurately than Sine stimuli. The pop-out image displays cerebellar clusters and a thalamus cluster not rendered in surface image. left insula into the left temporal pole and the left IFG, as well as another cluster in the left MTG (Table 3, Figure 7B). DISCUSSION Behavioral Results The intent in this study was to train participants to equal proficiency on two novel contrasts, a non-native vowel contrast and an acoustically matched sine-wave contrast. There are many metrics that can be considered when evaluating the degree to which we were successful in doing so; overall accuracy during training reflects the degree to which contrasts were learn- able in the moment, the slope of the categorization curve reflects the precision with which participants separated the categories, and differences in discrimination performance (in partic- ular spikes in discriminability of tokens that cross the category boundary) reflect the emerging “categorical” status of the sound category. It is notable that these metrics do not necessarily pattern together, even in well-developed native language categories. For instance, several studies have found no relationship between the categorization function and discrimination performance (Gerrits & Schouten, 2004; Hary & Massaro, 1982; van Hessen & Schouten, 1999; Schouten, Gerrits, & van Hessen, 2003). On balance, participants showed successful learning of both contrasts: Participants’ ability to both identify and discriminate between the two categories in each continuum increased Neurobiology of Language 356 Neural decoding of articulable and inarticulable sounds with training, and ultimately, performance on the training task, the slope of the category boundary, and the shape of the reaction time function were not significantly different between the two continua. The continua differed, however, in the degree of “acquired distinctiveness” in the discrimination data, with the vowel stimuli showing marked increases in sensitivity to between-category tokens, while sine-wave stimuli showed much higher overall discriminabil- ity, but no obvious peak for between-category tokens. This pattern may reflect qualitative dif- ferences in the way that speech versus nonspeech categories emerge, with speech sound categories more likely to have perceptual warping within the category, a classic hallmark of native language speech categories. Below we consider our results in the context of other lit- erature on speech and nonspeech category learning. The vowel contrast in particular showed a behavioral pattern that qualitatively resembles nativelike perception (i.e., an increased sensitivity across category boundaries, a decreased sensitivity within category boundaries, a sharp categorization curve, and an increased reaction time to near-boundary decisions (Liberman, Harris, Hoffman, & Griffith, 1957; Pisoni & Tash 0 1974). On the last day of training, d values from the AX discrimination task for within-category 0 tokens, (e.g., 1–3 or 5–7) were much lower than on the first day of training, whereas d scores increased for the tokens most often containing participants’ perceptual boundary (2–4 and 3–5). These results are congruent with previous studies of native English speakers’ perception of the front-rounded vowel /y/, which shows that listeners will initially assimilate /y/ to the American English vowel /i/. Nonetheless, this contrast is ultimately learnable to a high level of proficiency (Levy & Strange, 2008; Levy, 2009). Sine-wave versions of speech stimuli are not automatically mapped to speech analogues. In our study, we removed F1 and F4 to prevent listeners from reconstructing the formants for the /i/ and /y/ vowels and thus retrieving those speech categories. Studies of the perception of sine- wave speech continua have found that participants’ responses in identification tasks were ei- ther random or largely continuous (Grunke & Pisoni, 1982; Johnson & Ralston, 1994, exp. 1; Desai et al., 2008), though a sine-wave speech version of a vowel continuum was perceived categorically regardless of whether the listeners interpreted them as speech or nonspeech (Johnson & Ralston, 1994, exp. 2). As discussed above, the sine-wave behavior is betrayed by one finding that suggests that the sine-wave category was more poorly learned (or less fully elaborated) than the vowel stimuli, namely a lack of an obvious peak in the discrimination function around the category boundary. However, the significant interaction present in the sine-wave AX discrimination analysis, wherein there was a greater increase in sensitivity for the between-category token than the within-category tokens at posttest, does point to an emerging pattern for acquired distinctiveness. Without additional exposure, this point will re- main speculative as there is a lack of converging research on discrimination performance for sine-wave speech after multiple days of training. fMRI Results In general, greater activation for the sine-wave tokens in the univariate analysis points to the possibility that they were being processed less efficiently by a general auditory, and not a speech-specific, mechanism. A broad scale contrast of areas where sine-wave tokens elicited greater activation than vowel tokens revealed bilateral clusters in STG, similar to the findings in Liebenthal et al. (2010), where nonspeech stimuli showed greater activation than speech stimuli at posttest in STG. This could be due to the fact that the sine-wave tokens were still more novel and participants had less of an opportunity to habituate to them. However, our results replicate Liebenthal et al. (2010), where participants were trained over the course of Neurobiology of Language 357 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . e d u n o / l / l a r t i c e - p d f / / / / 1 3 3 3 9 1 8 6 7 7 4 6 n o _ a _ 0 0 0 1 6 p d / . l f b y g u e s t t o n 0 8 S e p e m b e r 2 0 2 3 Neural decoding of articulable and inarticulable sounds four sessions on their nonspeech stimuli, suggesting that mere lack of exposure is not the only driving factor behind the sine-wave vs. vowel activation differences. Given the broad scale of dimensions that vary between the two stimulus sets (e.g., articulatory information or acoustic complexity), it is difficult to draw conclusions about the driving force of these neural differ- ences using only univariate methods. Regions identified in vowel decoding analysis are outside of dorsal stream As discussed earlier, the dual stream model posits that the dorsal stream is recruited during the learning of speech sounds, where it serves as an auditory-motor integration mechanism. Evidence for this role has been difficult to ascertain given that many previous studies have used stimuli where only the inarticulable stimuli were unfamiliar (e.g., Liebenthal et al. 2005, 2010). To more conclusively answer this question, we trained participants on an unfamiliar vowel con- trast, and used multivariate decoding analyses to investigate which brain regions could represent the differences between these sounds. Any region that can decode these newly learned vowel categories could do so by virtue of access to (1) articulatory differences (lip rounding differences that listeners infer or extract from the auditory input), (2) abstract phonological categories (/i/ vs. /y/), or (3) the acoustic differences between the two sounds (F3 is higher in /y/). We found that the right MTG and the right STG were sensitive to vowel category information. Crucially these areas are not included in the dorsal stream or implicated in speech-motor pro- cessing but have been identified in previous imaging studies of non-native phonetic learning (Callan et al., 2003; Wong, Perrachione, & Parrish, 2007). Additionally, another significant clus- ter was found in the right IFG, which along with the right MTG, has been implicated as playing a role in perceptual learning for speech, as Myers and Mesite (2014) found it to be sensitive to changes in representation of an /s/-/∫/ contrast in response to a talker-specific artificial accent. The right IFG has also been shown to be sensitive to nonlinguistic acoustic processing, like pitch (Wang, Sereno, Jongman, & Hirsch, 2003), as well as to play a role in non-native phonetic learn- ing (Myers & Swan, 2012). Thus, we take it as more likely that decoding sensitivity in these right hemisphere regions is based upon differential representations of the abstract categories of /i/Vowel and /y/Vowel and not the differences in articulation between them (or simply the acoustic differ- ences). The only left hemisphere region that successfully decoded the two vowel categories was the left MTG, and the cluster overlaps strongly with the phonological network described as part of the ventral stream (Hickok & Poeppel, 2007). As such, our preferred interpretation of this pat- tern is that the left MTG likely represents phonological category differences between /i/ and /y/. The failure to find evidence for dorsal stream representation of different speech sounds during phonetic learning, when it is predicted by the dual stream model, suggests that articulatory in- formation is either coded in areas not involved in speech-motor gestures (e.g., the right MTG and the STG) along with phonological representations, or represented at a level that our multivariate analysis is not sensitive to, or else simply not recruited during this task. Dorsal stream identified only when vowel decoding is contrasted against sine-wave decoding It is possible the sensitivity to newly learned vowel categories reflects acquisition of any new auditory categories, and is not speech-specific. In order to differentiate these possibilities, we compared classification results from the /i/Vowel vs. /y/Vowel and the /i/Sine vs. /y/Sine decoding maps. The resulting statistical map indicates regions that better decoded the two different vowel categories than the two sine-wave categories, or put differently, regions that are sensitive specifically to auditory contrasts that differ in articulation. The critical comparison of interest is whether dorsal stream regions show discriminable patterns of activation only during the Neurobiology of Language 358 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . e d u n o / l / l a r t i c e - p d f / / / / 1 3 3 3 9 1 8 6 7 7 4 6 n o _ a _ 0 0 0 1 6 p d . / l f b y g u e s t t o n 0 8 S e p e m b e r 2 0 2 3 Neural decoding of articulable and inarticulable sounds learning of articulable sounds ( Vowel) but not nonarticulable sounds (Sine), as would be pre- dicted by the conception of the dorsal stream as an auditory-motor integration mechanism. Several dorsal stream regions emerged from this analysis; the left IFG and the left insula, the bilateral supplementary motor areas, and a left MTG extending into posterior STG (see Figure 7B) all showed superior decoding of vowel tokens compared to the sine-wave analogues. These results are consistent with several transcranial magnetic stimulation (TMS) studies of the role of motor regions in speech perception, in which “artificial lesions” of speech motor regions (i.e., primary and pre-motor cortices) lead to changes in discriminating or identifying native language speech sounds, indicating a causal role of the articulatory regions during speech perception (Rogers, Möttönen, Boyles, & Watkins, 2014; Möttönen & Watkins, 2009; Krieger-Redwood, Gaskell, Lindsay, & Jefferies, 2013; Sato, Tremblay, & Gracco, 2009; Meister et al. 2007; Smalle, Rogers, & Möttönen, 2015). However, these studies have not necessarily addressed the dorsal stream as circuit but have instead focused on its constit- uent regions one at a time. For instance, Krieger-Redwood et al. (2013) showed that stimula- tion of dorsal stream regions affected phonetic decisions but not performance, on a more naturalistic task where listeners had to match a word to a sample picture. Murakami, Kell, Restle, Ugawa, and Ziemann (2015) applied TMS simultaneously to the left pre-motor cortex and the left IFG (two of the three regions of the dorsal stream) and found decrements in pho- nological processing only in noisy conditions, which suggests that the dorsal stream may only be brought online during perception when stimulus identity is unclear. In our study, it is pos- sible that the dorsal stream is recruited when the newly learned vowel /y/ must be distin- guished from similar English vowels already in the repertoire. By contrast, the sine-wave stimuli most likely do not assimilate to any existing auditory categories, and thus their identity is more easily resolved. The results from the Vowel decoding−Sine-wave decoding analysis do provide support for the dual stream model’s conception of the dorsal stream as an auditory-motor integration mechanism, but we hesitate to conclusively accept this interpretation. As we will discuss in the next section, no brain regions significantly decoded sine-wave stimuli, which means that the Vowel decoding−Sine-wave decoding analysis is displaying areas that have higher accuracy in the vowel decoding, but nevertheless do not reach significance in the vowel decoding alone (there was essentially no overlap between the map generated by the Vowel decoding−Sine-wave decoding analysis and the map generated by the vowel decoding alone beyond the left MTG). To that end, we believe the results from the vowel decoding, which do not involve any dorsal stream regions, sufficiently address the original motivation of this study. Based upon these results, future research should step back and more thoroughly examine if the dorsal stream is truly involved in phonetic learning at all. Another potential issue with the Vowel decoding−Sine-wave decoding lies in the present study’s ability to completely control the equivalence of category structure across the vowel and sine-wave stimuli. As our motivating goal was to investigate whether dorsal stream recruit- ment in previous studies reflected auditory-motor integration or simply the formation of new auditory categories, it was necessary to have two sets of stimuli that differed only on the dimension of articulatory feasibility. As discussed above, although our stimuli were equated in terms of participants’ success on the training task, that equivalent accuracy may have led to qualitative differences in the shape of the emerging categories. Specifically, by one metric (discrimination) our sine-wave stimuli differed from our vowel stimuli in their degree of categoricity. Therefore, it is possible that the regions found in the Vowel decoding−Sine-wave Neurobiology of Language 359 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . e d u n o / l / l a r t i c e - p d f / / / / 1 3 3 3 9 1 8 6 7 7 4 6 n o _ a _ 0 0 0 1 6 p d . / l f b y g u e s t t o n 0 8 S e p e m b e r 2 0 2 3 Neural decoding of articulable and inarticulable sounds decoding could reflect differences in the emergence of categorical perception and not just articulatory information. Future research will need to more strongly control for category struc- ture to more precisely equivalate between speech and nonspeech sounds. Gradient neural representations of sine-wave stimuli may have impeded decoding The sine-wave decoding analysis did not uncover any brain regions that reliably represented the two sine-wave categories. This is a bit of a puzzle given strong behavioral evidence that participants were able to distinguish the categories. One explanation for the failure to find neural sensitivity to sine-wave tokens appeals to the degree to which these stimuli showed “acquired equivalence” within the sound category. In speech processing, acoustic cues to speech sounds are initially encoded continuously at the neural level (Blumstein, Myers, & Rissman, 2005; Frye et al., 2007; Toscano, McMurray, Dennhardt, & Luck, 2010), then cate- gorical structure is imposed on them quickly and neural responses begin to reflect phonolog- ical category membership early in processing (Toscano, Anderson, Fabiani, Gratton, & Garnsey, 2018). We speculate that by contrast, nonspeech categories such as the sine-wave speech tokens used in the current study may continue to be represented in a graded fashion at the neural level. Because our analysis favored grouping of two more perceptually distinct sine- wave stimuli to reflect a single category (e.g., steps 1 and 3 for /i/Sine), this could have created a more difficult classification problem for sine-wave tokens. Submitting just the endpoints of the sine-wave continuum (step 1 and step 7) to the same decoding analysis yielded a significant cluster in the right MFG. Notably, this analysis suffers from less power than that reported in the main body of the text, using only half of the tokens in the decoding analysis. This finding con- trasts with a previous multivariate fMRI study of trained nonspeech novel auditory categories that found that bilateral primary auditory cortex activation could successfully classify the two categories (Ley et al., 2012). On the topic of classifier accuracy, observed accuracy at the group levels was numerically low, but statistically reliable. For the goals of this experiment, maximizing accuracy of the classifier is not of primary importance; significantly above-chance findings indicate that there are structured neural patterns in response to different stimuli, and using accuracy levels as a measure of the size of this effect is a faulty assumption. For a detailed explanation of the dif- ferences between MVPA for interpretation versus prediction, see Hebart and Baker (2018). Subcortical regions implicated only in learning of speech sounds Speech sound processing, especially during learning and adapting to speech sound variants, may not be limited to cortical regions. In the present data, better decoding for the vowel con- trast compared to the sine-wave contrast was found in the left thalamus. Previous work has observed that the thalamus is sensitive to human speech sounds: Dehaene-Lambertz et al. (2005) found that the thalamus was generally more active for speech rather than sine-wave speech analogues in humans, while Kraus et al. (1994) found that the guinea pig thalamus is sensitive to complex spectral differences between human speech sounds even when animals were not exposed to any kind of training on these sounds. However, more relevant to this goal of this study is the relationship between the thalamus and articulatory information; neuropsy- chological investigations have found that damage to the thalamus often yields difficulties with the articulation of speech sounds (Jonas, 1982; Wallesch et al., 1983), and our results suggest that during the learning of new speech sounds, the thalamus may be representing the articu- latory codes that will later be used for production, similar to the role of the dorsal stream in the dual stream model. Neurobiology of Language 360 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . e d u n o / l / l a r t i c e - p d f / / / / 1 3 3 3 9 1 8 6 7 7 4 6 n o _ a _ 0 0 0 1 6 p d . / l f b y g u e s t t o n 0 8 S e p e m b e r 2 0 2 3 Neural decoding of articulable and inarticulable sounds More broadly, there has been increasing interest in the contribution of subcortical brain structures in experience-dependent plasticity in the auditory system (Chandrasekaran, Skoe, & Kraus, 2014), and it is suggested that the cerebellum is involved in adaptation to alterations in speech through its functional connectivity to cortical language areas (Guediche, Holt, Laurent, Lim, & Fiez, 2015). In the current study, the significant right cerebellar clusters may reflect the engagement of this language plasticity network (similar to Guediche et al., 2015) as a result of the nascent formation of a speech sound category. Language related cer- ebellar function is heavily right lateralized, while the left cerebellum, especially lobule VI and Crus 1, are implicated as having a role in executive functioning. Left cerebellar engagement in this study could reflect a decision-making process about category membership for the vowel contrast (for a review, see Stoodley & Schmahmann, 2009). Conclusion In sum, the MVPA results from the present experiment provide support for the interpretation that dorsal stream regions are recruited during the learning of articulable sounds only, a ques- tion that has been obscured in previous studies due to stimuli not being equated on profi- ciency or articulatory feasibility. We did find evidence for dorsal stream recruitment when looking for regions that could preferentially discriminate category membership for articulable contrasts (our vowel contrast) rather than inarticulable contrasts (our sine-wave contrast); however, differences in category structure between the articulable and inarticulable contrasts (in that the former was perceived more categorically by some metrics than the latter) also allow for the possibility that this recruitment reflects sensitivity to category structure and not articulatory feasibility. However, these regions did not appear when examining decoding of the two articulable vowel sounds alone. Instead, a series of regions outside of the dorsal stream previously implicated in non-native phonetic learning could successfully classify the two vowel categories. Future work will need to address the interplay between recruitment of the dorsal stream in the initial stages of phonetic learning and the contribution of right hemi- sphere cortical regions. l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . e d u n o / l / l a r t i c e - p d f / / / / 1 3 3 3 9 1 8 6 7 7 4 6 n o _ a _ 0 0 0 1 6 p d / . l FUNDING INFORMATION Emily B. Myers, National Institute on Deafness and Other Communication Disorders (http://dx. doi.org/10.13039/100000055), Award ID: R01DC013064. James Magnuson, National Science Foundation (http://dx.doi.org/10.13039/100000001), Award ID: 1735225. AUTHOR CONTRIBUTIONS Emily Myers proposed the original motivation and design of the experiment. Emily Myers and David Saltzman created the materials and final design. David Saltzman collected the data and performed all analyses. David Saltzman wrote the initial draft of the manuscript, with additions and revisions by Emily Myers. David Saltzman responded to reviewers’ comments with addi- tions from Emily Myers. f b y g u e s t t o n 0 8 S e p e m b e r 2 0 2 3 REFERENCES Arsenault, J. S., & Buchsbaum, B. R. (2015). Distributed neural rep- resentations of phonological features during speech perception. Journal of Neuroscience, 35(2), 634–642. https://doi.org/ 10.1523/JNEUROSCI.2454-14.2015 Arsenault, J. S., & Buchsbaum, B. R. (2016). No evidence of soma- totopic place of articulation feature mapping in motor cortex during passive speech perception. Psychonomic Bulletin & Review, 23(4), 1231–1240. https://doi.org/10.3758/s13423-015- 0988-z Bates, D, Mächler, M, Bolker, B, & Walker, S. (2015). Fitting linear mixed-effects models using lme4. Journal of Statistical Software, 67(1), 1–48. https://doi.org/10.18637/jss.v067.i01 Neurobiology of Language 361 Neural decoding of articulable and inarticulable sounds Best, C. T. (1995). A direct realist view of cross-language speech perception. Speech Perception and Linguistic Experience, 171–206. Blumstein, S. E., Myers, E. B., & Rissman, J. (2005). The perception of voice onset time: An fMRI investigation of phonetic category structure. Journal of Cognitive Neuroscience, 17(9), 1353–1366. https://doi.org/10.1162/0898929054985473 Boersma, P., & Weenink, D. (2020). Praat: Doing phonetics by computer [Computer program]. Version 6.1.16, retrieved 6 June 2020 from http://www.praat.org/ Callan, D. E., Tajima, K., Callan, A. M., Kubo, R., Masaki, S., & Akahane-Yamada, R. (2003). Learning-induced neural plasticity associated with improved identification performance after train- ing of a difficult second-language phonetic contrast. NeuroImage, 19(1), 113–124. https://doi.org/10.1016/S1053- 8119(03)00020-X Chandrasekaran, B., Skoe, E., & Kraus, N. (2014). An integrative model of subcortical auditory plasticity. Brain Topography, 27(4), 539–552. https://doi.org/10.1007/s10548-013-0323-9 Correia, J. M., Jansma, B. M. B., & Bonte, M. (2015). Decoding ar- ticulatory features from fMRI responses in dorsal speech regions. Journal of Neuroscience, 35(45), 15015–15025. https://doi.org/ 10.1523/JNEUROSCI.0977-15.2015 Cox, R. W. (1996). AFNI: Software for analysis and visualization of functional magnetic resonance neuroimages. Computers and Biomedical Research, an International Journal, 29(3), 162–173. https://doi.org/10.1006/cbmr.1996.0014 Darcy, I., Dekydtspotter, L., Sprouse, R. A., Glover, J., Kaden, C., McGuire, M., & Scott, J. H. (2012). Direct mapping of acoustics to phonology: On the lexical encoding of front rounded vowels in L1 English–L2 French acquisition. Second Language Research, 28(1), 5–40. https://doi.org/10.1177/0267658311423455 Dehaene-Lambertz, G., Pallier, C., Serniclaes, W., Sprenger- Charolles, L., Jobert, A., & Dehaene, S. (2005). Neural correlates of switching from auditory to speech perception. NeuroImage, 24(1), 21–33. https://doi.org/10.1016/j.neuroimage.2004.09.039 Desai, R., Liebenthal, E., Waldron, E., & Binder, J. R. (2008). Left posterior temporal regions are sensitive to auditory categoriza- tion. Journal of Cognitive Neuroscience, 20(7), 1174–1188. https://doi.org/10.1162/jocn.2008.20081 Earle, F. S., & Myers, E. B. (2015). Sleep and native language inter- ference affect non-native speech sound learning. Journal of Experimental Psychology. Human Perception and Performance, 41(6), 1680–1695. https://doi.org/10.1037/xhp0000113 Fowler, C. A. (1986). An event approach to the study of speech per- ception from a direct–realist perspective. Journal of Phonetics, 14(1), 3–28. https://doi.org/10.1016/S0095-4470(19)30607-2 Frye, R. E., Fisher, J. M., Coty, A., Zarella, M., Liederman, J., & Halgren, E. (2007). Linear coding of voice onset time. Journal of Cognitive Neuroscience, 19(9), 1476–1487. https://doi.org/ 10.1162/jocn.2007.19.9.1476 Galantucci, B., Fowler, C. A., & Turvey, M. T. (2006). The motor theory of speech perception reviewed. Psychonomic Bulletin & Review, 13(3), 361–377. https://doi.org/10.3758/BF03193857 Gerrits, E., & Schouten, M. E. H. (2004). Categorical perception de- pends on the discrimination task. Perception & Psychophysics, 66(3), 363–376. https://doi.org/10.3758/BF03194885 Grunke, M. E., & Pisoni, D. B. (1982). Some experiments on per- ceptual learning of mirror-image acoustic patterns. Perception & Psychophysics, 31(3), 210–218. https://doi.org/10.3758/ BF03202525 Guediche, S., Holt, L. L., Laurent, P., Lim, S.-J., & Fiez, J. A. (2015). Evidence for cerebellar contributions to adaptive plasticity in speech perception. Cerebral Cortex, 25(7), 1867–1877. https:// doi.org/10.1093/cercor/bht428 Hary, J. M., & Massaro, D. W. (1982). Categorical results do not imply categorical perception. Perception & Psychophysics, 32(5), 409–418. https://doi.org/10.3758/BF03202770 Haxby, J. V., Gobbini, M. I., Furey, M. L., Ishai, A., Schouten, J. L., & Pietrini, P. (2001). Distributed and overlapping representations of faces and objects in ventral temporal cortex. Science, 293(5539), 2425–2430. https://doi.org/10.1126/science.1063736 Hebart, M. N., & Baker, C. I. (2018). Deconstructing multivariate decoding for the study of brain function. NeuroImage, 180, 4–18. https://doi.org/10.1016/j.neuroimage.2017.08.005 Hebart, M. N., Görgen, K., & Haynes, J.-D. (2015). The Decoding Toolbox (TDT): A versatile software package for multivariate anal- yses of functional imaging data. Frontiers in Neuroinformatics, 8. https://doi.org/10.3389/fninf.2014.00088 Hickok, G., & Buchsbaum, B. (2003). Temporal lobe speech percep- tion systems are part of the verbal working memory circuit: Evidence from two recent fMRI studies. Behavioral and Brain S c i e n c e s , 2 6 ( 6 ) , 7 4 0 – 7 4 1 . h t t p s : / / d o i . o r g / 1 0 . 1 0 1 7 / S0140525X03340166 Hickok, G., & Poeppel, D. (2007). The cortical organization of speech processing. Nature Reviews Neuroscience, 8(5), 393–402. https://doi.org/10.1038/nrn2113 Hutchison, E. R., Blumstein, S. E., & Myers, E. B. (2008). An event- related fMRI investigation of voice-onset time discrimination. NeuroImage, 40(1), 342–352. https://doi.org/10.1016/j. neuroimage.2007.10.064 Johnson, K., & Ralston, J. V. (1994). Automaticity in speech percep- tion: Some speech/nonspeech comparisons. Phonetica, 51(4), 195–209. https://doi.org/10.1159/000261975 Jonas, S. (1982). The thalamus and aphasia, including transcortical aphasia: A review. Journal of Communication Disorders, 15(1), 31–41. https://doi.org/10.1016/0021-9924(82)90042-9 Kraus, N., Mcgee, T., Carrell, T., King, C., Littman, T., & Nicol, T. (1994). Discrimination of speech-like contrasts in the auditory thalamus and cortex. The Journal of the Acoustical Society of America, 96, 2758. https://doi.org/10.1121/1.411282 Krieger-Redwood, K., Gaskell, M. G., Lindsay, S., & Jefferies, E. (2013). The selective role of premotor cortex in speech percep- tion: A contribution to phoneme judgements but not speech comprehension. Journal of Cognitive Neuroscience, 25(12), 2179–2188. https://doi.org/10.1162/jocn_a_00463 Kriegeskorte, N., Goebel, R., & Bandettini, P. (2006). Information- based functional brain mapping. Proceedings of the National Academy of Sciences, 103(10), 3863–3868. https://doi.org/ 10.1073/pnas.0600244103 Lee, Y.-S., Turkeltaub, P., Granger, R., & Raizada, R. D. S. (2012). Categorical speech processing in Broca’s area: An fMRI study Journ al o f using m ultiva riate pa ttern-based analysis. Neuroscience, 32(11), 3942–3948. https://doi.org/10.1523/ JNEUROSCI.3814-11.2012 Levy, E. S. (2009). Language experience and consonantal context effects on perceptual assimilation of French vowels by American-English learners of French. The Journal of the Acoustical Society of America, 125(2), 1138–1152. https://doi. org/10.1121/1.3050256 Levy, E. S., & Strange, W. (2008). Perception of French vowels by American English adults with and without French language expe- rience. Journal of Phonetics, 36(1), 141–157. https://doi.org/ 10.1016/J.WOCN.2007.03.001 Ley, A., Vroomen, J., Hausfeld, L., Valente, G., De Weerd, P., & Formisano, E. (2012). Learning of new sound categories shapes Neurobiology of Language 362 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . e d u n o / l / l a r t i c e - p d f / / / / 1 3 3 3 9 1 8 6 7 7 4 6 n o _ a _ 0 0 0 1 6 p d / . l f b y g u e s t t o n 0 8 S e p e m b e r 2 0 2 3 Neural decoding of articulable and inarticulable sounds neural response patterns in human auditory cortex. Journal of Neuroscience, 32(38), 13273–13280. https://doi.org/10.1523/ JNEUROSCI.0584-12.2012 Liberman, A. M., Cooper, F. S., Shankweiler, D. P., & Studdert- K e n n e d y , M . ( 1 9 6 7 ) . P e r c e p t i o n o f t h e s p e e c h c o d e . Psychological Review, 74(6), 431–461. https://doi.org/10.1037/ h0020279 Liberman, A. M., Harris, K. S., Hoffman, H. S., & Griffith, B. C. (1957). The discrimination of speech sounds within and across phoneme boundaries. Journal of Experimental Psychology, 54(5), 358–368. https://doi.org/10.1037/h0044417 Liebenthal, E., Binder, J. R., Spitzer, S. M., Possing, E. T., & Medler, D. A. (2005). Neural substrates of phonemic perception. Cerebral Cortex, 15(10), 1621–1631. https://doi.org/10.1093/cercor/ bhi040 Liebenthal, E., Desai, R., Ellingson, M. M., Ramachandran, B., Desai, A., & Binder, J. R. (2010). Specialization along the left su- perior temporal sulcus for auditory categorization. Cerebral Cortex, 20(12), 2958–2970. https://doi.org/10.1093/cercor/ bhq045 Linares, D., & López-Moliner, J. (2016). quickpsy: An R Package to fit psychometric functions for multiple groups. http://diposit.ub. edu/dspace/handle/2445/116040 Lotto, A. J., Hickok, G. S., & Holt, L. L. (2009). Reflections on mirror neurons and speech perception. Trends in Cognitive Sciences, 13(3), 110–114. https://doi.org/10.1016/j.tics.2008.11.008 Mathôt, S., Schreij, D., & Theeuwes, J. (2012). OpenSesame: An open-source, graphical experiment builder for the social sci- ences. Behavior Research Methods, 44(2), 314–324. https://doi. org/10.3758/s13428-011-0168-7 Matuschek, H., Kliegl, R., Vasishth, S., Baayen, H., & Bates, D. (2017). Balancing type I error and power in linear mixed models. Journal of Memory and Language, 94, 305–315. https://doi.org/ 10.1016/j.jml.2017.01.001 Meister, I. G., Wilson, S. M., Deblieck, C., Wu, A. D., & Iacoboni, M. (2007). The essential role of premotor cortex in speech per- ception. Current Biology, 17(19), 1692–1696. https://doi.org/ 10.1016/J.CUB.2007.08.064 Mesgarani, N., Cheung, C., Johnson, K., & Chang, E. F. (2014). Phonetic feature encoding in human superior temporal gyrus. Science, 343(6174), 1006–1010. https://doi.org/10.1126/ science.1245994 Möttönen, R., & Watkins, K. E. (2009). Motor representations of ar- ticulators contribute to categorical perception of speech sounds. Journal of Neuroscience, 29(31), 9819–9825. https://doi.org/ 10.1523/JNEUROSCI.6018-08.2009 Mumford, J. A., Davis, T., & Poldrack, R. A. (2014). The impact of study design on pattern estimation for single-trial multivariate pattern analysis. NeuroImage, 103, 130–138. https://doi.org/ 10.1016/j.neuroimage.2014.09.026 Mumford, J. A., Turner, B. O., Ashby, F. G., & Poldrack, R. A. (2012). Deconvolving BOLD activation in event-related designs for multivoxel pattern classification analyses. NeuroImage, 59(3), 2636–2643. https://doi.org/10.1016/j.neuroimage.2011.08.076 Murakami, T., Kell, C. A., Restle, J., Ugawa, Y., & Ziemann, X. U. (2015). Left dorsal speech stream components and their con- tribution to phonological processing. Journal of Neuroscience, 35(4), 1411–1422. https://doi.org/10.1523/ JNEUROSCI.0246- 14.2015 Myers, E. B. (2007). Dissociable effects of phonetic competition and category typicality in a phonetic categorization task: An fMRI investigation. Neuropsychologia, 45(7), 1463–1473. https://doi.org/10.1016/j.neuropsychologia.2006.11.005 Myers, E. B., Blumstein, S. E., Walsh, E., & Eliassen, J. (2009). Inferior frontal regions underlie the perception of phonetic cate- gory invariance. Psychological Science, 20(7), 895–903. https:// doi.org/10.1111/j.1467-9280.2009.02380.x Myers, E. B., & Mesite, L. M. (2014). Neural systems underlying perceptual adjustment to non-standard speech tokens. Journal of Memory and Language, 76, 80–93. https://doi.org/10.1016/j. jml.2014.06.007 Myers, E. B., & Swan, K. (2012). Effects of category learning on neu- ral sensitivity to non-native phonetic categories. Journal of Cognitive Neuroscience, 24(8), 1695–1708. https://doi.org/ 10.1162/jocn_a_00243 Pisoni, D. B., & Tash, J. (1974). Reaction times to comparisons within and across phonetic categories. Percep tion & Psychophysics, 15(2), 285–290. https://doi.org/10.3758/ BF03213946 Pulvermüller, F., Huss, M., Kherif, F., Martin, F. M. del P., Hauk, O., & Shtyrov, Y. (2006). Motor cortex maps articulatory features of speech sounds. Proceedings of the National Academy of Sciences, 103(20), 7865–7870. https://doi.org/10.1073/ pnas.0509989103 Remez, R. E., Rubin, P. E., Pisoni, D. B., & Carrell, T. D. (1981). Speech perception without traditional speech cues. Science, 212(4497), 947–949. https://doi.org/10.1126/science.7233191 Rissman, J., Gazzaley, A., & D’Esposito, M. (2004). Measuring functional connectivity during distinct stages of a cognitive task. NeuroImage, 23(2), 752–763. https://doi.org/10.1016/j. neuroimage.2004.06.035 Rogers, J. C., Möttönen, R., Boyles, R., & Watkins, K. E. (2014). Discrimination of speech and non-speech sounds following theta- burst stimulation of the motor cortex. Frontiers in Psychology, 5. https://doi.org/10.3389/fpsyg.2014.00754 Sato, M., Tremblay, P., & Gracco, V. L. (2009). A mediating role of the premotor cortex in phoneme segmentation. Brain and Language, 111(1), 1–7. https://doi.org/10.1016/j.bandl. 2009.03.002 Schomers, M. R., & Pulvermüller, F. (2016). Is the sensorimotor cor- tex relevant for speech perception and understanding? An inte- grative review. Frontiers in Human Neuroscience, 10, 435–435. https://doi.org/10.3389/fnhum.2016.00435 Schouten, B., Gerrits, E., & van Hessen, A. (2003). The end of cat- egorical perception as we know it. Speech Communication, 41(1), 71–80. https://doi.org/10.1016/S0167-6393(02)00094-8 Scott, S. K., McGettigan, C., & Eisner, F. (2009). A little more con- versation, a little less action—Candidate roles for the motor cor- tex in speech perception. Nature Reviews Neuroscience, 10(4), 295–302. https://doi.org/10.1038/nrn2603 Singmann, H., Bolker, B., Westfall, J., Aust, F., & Ben-Shachar, M. S. (2020). afex: Analysis of factorial experiments. R package version 0.27-2. https://CRAN.R-project.org/package=afex Smalle, E. H. M., Rogers, J., & Möttönen, R. (2015). Dissociating contributions of the motor cortex to speech perception and re- sponse bias by using transcranial magnetic stimulation. Cerebral Cortex, 25(10), 3690–3698. https://doi.org/10.1093/cer- cor/bhu218 Stoodley, C. J., & Schmahmann, J. D. (2009). Functional topogra- phy in the human cerebellum: A meta-analysis of neuroimaging studies. NeuroImage, 44(2), 489–501. https://doi.org/10.1016/j. neuroimage.2008.08.039 Talairach, J., & Tournoux, P. (1988). Co-planar stereotaxic atlas of the human brain. Stuttgart: George Thieme. Toscano, J. C., Anderson, N. D., Fabiani, M., Gratton, G., & Garnsey, S. M. (2018). The time-course of cortical responses to Neurobiology of Language 363 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . e d u n o / l / l a r t i c e - p d f / / / / 1 3 3 3 9 1 8 6 7 7 4 6 n o _ a _ 0 0 0 1 6 p d / . l f b y g u e s t t o n 0 8 S e p e m b e r 2 0 2 3 Neural decoding of articulable and inarticulable sounds speech revealed by fast optical imaging. Brain and Language, 184, 32–42. https://doi.org/10.1016/j.bandl.2018.06.006 Toscano, J. C., McMurray, B., Dennhardt, J., & Luck, S. J. (2010). Continuous perception and graded categorization: Electrophys- iological evidence for a linear relationship between the acoustic signal and perceptual encoding of speech. Psychological Science, 21(10), 1532–1540. https://doi.org/10.1177/0956797610384142 Tsao, F.-M., Liu, H.-M., & Kuhl, P. K. (2004). Speech perception in infancy predicts language development in the second year of life: A longitudinal study. Child Development, 75(4), 1067–1084. https://doi.org/10.1111/j.1467-8624.2004.00726.x van Hessen, A. J., & Schouten, M. E. H. (1999). Categorical percep- tion as a function of stimulus quality. Phonetica, 56(1–2), 56–72. https://doi.org/10.1159/000028441 Wallesch, C.-W., Kornhuber, H. H., Brunner, R. J., Kunz, T., Hollerbach, B., & Suger, G. (1983). Lesions of the basal ganglia, thalamus, and deep white matter: Differential effects on language functions. Brain and Language, 20(2), 286–304. https://doi.org/ 10.1016/0093-934X(83)90046-9 Wang, Y., Sereno, J. A., Jongman, A., & Hirsch, J. (2003). fMRI ev- idence for cortical modification during learning of Mandarin lex- ical tone. Journal of Cognitive Neuroscience, 15(7), 1019–1027. https://doi.org/10.1162/089892903770007407 Wilson, S. M. (2009). Speech perception when the motor system is compromised. Trends in Cognitive Sciences, 13(8), 329–331. https://doi.org/10.1016/j.tics.2009.06.001 Wilson, S. M., & Iacoboni, M. (2006). Neural responses to non- native phonemes varying in producibility: Evidence for the sensorimotor nature of speech perception. NeuroImage, 33(1), 316–325. https://doi.org/10.1016/J.NEUROIMAGE.2006.05.032 Wilson, S. M., Saygin, A. P., Sereno, M. I., & Iacoboni, M. (2004). Listening to speech activates motor areas involved in speech production. Nature Neuroscience, 7, 701–702. https://doi.org/ 10.1038/nn1263 Wong, P. C. M., Perrachione, T. K., & Parrish, T. B. (2007). Neural characteristics of successful and less successful speech and word learning in adults. Human Brain Mapping, 28(10), 995–1006. https://doi.org/10.1002/hbm.20330 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . e d u n o / l / l a r t i c e - p d f / / / / 1 3 3 3 9 1 8 6 7 7 4 6 n o _ a _ 0 0 0 1 6 p d / . l f b y g u e s t t o n 0 8 S e p e m b e r 2 0 2 3 Neurobiology of Language 364ARTÍCULO DE INVESTIGACIÓN imagen
ARTÍCULO DE INVESTIGACIÓN imagen
ARTÍCULO DE INVESTIGACIÓN imagen
ARTÍCULO DE INVESTIGACIÓN imagen
ARTÍCULO DE INVESTIGACIÓN imagen
ARTÍCULO DE INVESTIGACIÓN imagen

Descargar PDF