Early Electrophysiological Markers of
Navigational Affordances in Scenes
Assaf Harel1, Jeffery D. Nador1, Michael F. Bonner2, and Russell A. Epstein3
Abstracto
■ Scene perception and spatial navigation are interdependent
cognitive functions, and there is increasing evidence that corti-
cal areas that process perceptual scene properties also carry
information about the potential for navigation in the environ-
mento (navigational affordances). Sin embargo, the temporal stages
by which visual information is transformed into navigationally
relevant information are not yet known. Nosotros planteamos la hipótesis de que
navigational affordances are encoded during perceptual pro-
cessing and therefore should modulate early visually evoked
ERPs, especially the scene-selective P2 component. To test this
idea, we recorded ERPs from participants while they passively
viewed computer-generated room scenes matched in visual
complejidad. By simply changing the number of doors (0 doors,
1 door, 2 doors, 3 doors), we were able to systematically vary
the number of pathways that afford movement in the local
ambiente, while keeping the overall size and shape of the
environment constant. We found that rooms with 0 doors
evoked a higher P2 response than rooms with three doors, estafa-
sistent with prior research reporting higher P2 amplitude to
closed relative to open scenes. Además, we found P2 ampli-
tude scaled linearly with the number of doors in the scenes.
Navigability effects on the ERP waveform were also observed
in a multivariate analysis, which showed significant decoding
of the number of doors and their location at earlier time win-
dows. Juntos, our results suggest that navigational affor-
dances are represented in the early stages of scene perception.
This complements research showing that the occipital place
area automatically encodes the structure of navigable space
and strengthens the link between scene perception and
navigation. ■
INTRODUCCIÓN
How do we find our way in the environment? What allows
us to successfully move about in the world without getting
lost? Navigation, the act of finding one’s way to a given
destination, is a multisensory process requiring the inte-
gration of multiple types of sensory information (visual,
vestibular, proprioceptive) about the environment,
including information about direction, distancia, and loca-
ción (Ekstrom, Spiers, Bohbot, & Rosenbaum, 2018). En
humanos, the most prominent sensory modality guiding
navigation in the environment is vision (Ekstrom, 2015).
The particular advantage that vision confers to navigation
is that it allows observers to recognize their surroundings
remotely, even before they embark on any movement in
their environment. A este respecto, visual scene perception,
eso es, the recognition of one’s surroundings, can be con-
sidered as an essential part of navigation, serving as a first
vital stage in a cascade of processes, which ultimately
culminate in successful navigation ( Julian, Keinath,
Marchette, & Epstein, 2018). Under this conceptualiza-
ción, scene perception serves as the gate to navigation,
demonstrating more broadly how visual processing
(es decir., scene recognition) is critical for and cannot be
1Wright State University, Dayton, OH, 2Johns Hopkins Univer-
sity, baltimore, Maryland, 3Universidad de Pennsylvania, Filadelfia
© 2022 Instituto de Tecnología de Massachusetts
dissociated from action, planificación, and memory systems
(es decir., navigation).
Support for the link between scene recognition and spa-
tial navigation1 comes from neuroimaging. Several studies
using fMRI reported that scene-selective regions sensitive
to visual scene properties, such as category ( Walther,
Chai, Caddigan, Arroyo, & Fei-Fei, 2011; Walther, Caddigan,
Fei-Fei, & Arroyo, 2009) and spatial expanse (Harel, Kravitz,
& Panadero, 2013; Kravitz, Peng, & Panadero, 2011), also carry
pertinent information for navigation (Persichetti & Dilks,
2018; Bonner & Epstein, 2017; Kamps, Julian, Kubilius,
Kanwisher, & Dilks, 2016). One region in particular that
has been suggested to be involved in representing visually
guided navigation information is the occipital place area
(OPA: Dilks, Julian, Paunov, & Kanwisher, 2013; Hasson,
Harel, Exacción, & Malach, 2003; Nakamura et al., 2000).
OPA activity is sensitive to various forms of navigation-
relevant information, including egocentric position and
heading (es decir., “viewpoint”: Epstein, Higgins, Jablonski, &
Feiler, 2007; Epstein, Higgins, & Thompson-Schill, 2005)
“sense” (left–right: Dilks, Julian, Kubilius, Spelke, &
Kanwisher, 2011), egocentric distance (proximal-distal:
Persichetti & Dilks, 2016), first-person perspective motion
through scenes (Kamps, Lall, & Dilks, 2016), and number
of local elements in a scene, which can be used for obstacle
avoidance (Kamps, Julian, et al., 2016). En tono rimbombante, OPA
response patterns have also been reported to specifically
Revista de neurociencia cognitiva 34:3, páginas. 397–410
https://doi.org/10.1162/jocn_a_01810
yo
D
oh
w
norte
oh
a
d
mi
d
F
r
oh
metro
h
t
t
pag
:
/
/
d
i
r
mi
C
t
.
metro
i
t
.
mi
d
tu
/
j
/
oh
C
norte
a
r
t
i
C
mi
–
pag
d
yo
F
/
/
/
/
3
4
3
3
9
7
1
9
8
5
0
1
9
/
j
oh
C
norte
_
a
_
0
1
8
1
0
pag
d
.
F
b
y
gramo
tu
mi
s
t
t
oh
norte
0
7
S
mi
pag
mi
metro
b
mi
r
2
0
2
3
index the spatial structure of navigational affordances (es decir.,
potential paths for movement in a scene), as operational-
ized by the number and locations of paths in the environ-
mento (Bonner & Epstein, 2017).
Although there is clear evidence that scene-selective cor-
tical regions carry navigation-related information, it is still
not clear how incoming visual information is transformed
into navigation-relevant information that could potentially
be used to guide movement in space. A recent TMS study
demonstrated that OPA plays a causal role in transforming
perceptual inputs into spatial memories linked to environ-
mental boundaries ( Julian, ryan, hamilton, & Epstein,
2016), but this study used a continuous theta-burst proto-
col that lacked the temporal resolution to probe the tem-
poral sequence of these processes. Information about
timing is essential for determining the extent to which
the extraction of navigation-related information from a
scene is indeed a visually driven perceptual process.
One alternative, por ejemplo, is that OPA sensitivity to nav-
igational affordances might reflect recurrent feedback
from posterior parietal cortex (Kravitz, Saleem, Panadero, &
Mishkin, 2011) rather than early sensory processing.2
Determining the nature of scene affordances processing
thus requires a method with high temporal resolution,
such as magnetoencephalography (MEG) and EEG,
which can establish the temporal dynamics of scene per-
ception for navigation. Yet another advantage of EEG is
that it can establish the processes involved in the extrac-
tion of navigationally relevant information: Various ERPs
have been shown to index multiple cognitive processes
(p.ej., attentional allocation, semantic categorization,
memory encoding; para una revisión, see the work of Luck,
2014) and thus be used to determine the mechanisms
underlying a certain task or experimental manipulation.
M/EEG studies have recently started to uncover the
time course of scene perception, by examining how differ-
ent scene properties across a variety of complexity levels
get processed over time. At a categorical level, the process-
ing of scenes can be distinguished from the processing of
other complex visual categories by 220 msec poststimulus
onset (Harel, Groen, Kravitz, Deouell, & Panadero, 2016; Sato
et al., 1999). Específicamente, the amplitude of the posterior
P2 ERP component (peaking around 220 msec after stim-
ulus onset) is higher in response to scenes than to faces
and common objects. The P2 component has thus been
suggested to index scene-selective processing (analogous
to the face-selective N170 ERP component), particularly
the processing of high-level global scene information
(Harel et al., 2016, 2020; Hansen, Noesen, Nador, & Harel,
2018). En efecto, P2 amplitude not only distinguishes
between scenes and other visual categories but is also sen-
sitive to various global scene properties (GSPs), como
spatial expanse and naturalness: It is higher to closed than
open scenes, can distinguish natural from man-made
escenas (Harel et al., 2016, 2020; Hansen et al., 2018),
and notably, it is not modulated by local texture informa-
ción, in contrast to earlier visually evoked components
(Harel et al., 2020). P2 amplitude is also diagnostic of
scene naturalness and spatial expanse at the level of
individual scene images, with its response variance to indi-
vidual scenes significantly explained by both summary
image statistics (approximating naturalness and spatial
expanse) and subjective behavioral ratings (Harel et al.,
2016; see also the work of Cichy, Khosla, Pantazis, & Oliva,
2017). Studies using single-trial decoding approaches
revealed that scene naturalness and spatial expanse (como
well as basic-level scene category) can also be decoded
from the neural signals even earlier, within the first
100 msec of processing (Henriksson, Mur, & Kriegeskorte,
2019; Lowe, Rajsic, Ferber, & Walther, 2018; see also the
works of Groen, Ghebreab, Lamme, & Scholte, 2016;
Groen, Ghebreab, Prins, Lamme, & Scholte, 2013).3 El
latency of both time windows suggests that low- también
as high-level diagnostic scene information is extracted at
the early perceptual stages of processing, supporting rapid
pre-attentive scene categorization (Hansen et al., 2018;
Groen et al., 2016; verde & Oliva, 2009; Rousselet,
Joubert, & Fabre-Thorpe, 2005).
Although the above M/ EEG studies cannot directly
establish OPA as the generator of the spatial expanse sig-
nals indexed by the P2 component, they are nevertheless
invaluable in providing a framework for thinking about the
visual system’s time course for extracting information
about scene navigability. If the processing of ecological,
navigability information occurs at the perceptual stages
of processing, this should be reflected in a modulation
of visually evoked activity during the first 250 mseg
(primarily the scene-selective P2 component) by naviga-
tionally relevant information, suggesting that scene per-
ception and navigation are indeed intrinsically linked.
Alternativamente, the extraction of navigability information
from the scene might reflect postperceptual processes
(p.ej., stimulus evaluation, Toma de decisiones, action plan-
y), which would result in a late effect of navigability,
and hence not impact the P2. One indication that the more
probable alternative of the two is former rather than the
latter is the ubiquity of spatial expanse—the extent to
which a scene depicts an enclosed or an open space—in
modulating early scene-evoked neural responses. El
effect of spatial expanse can be observed during early
stages of visual processing (Henriksson et al., 2019;
Hansen et al., 2018), across a variety of stimulus sets and
physical image properties (p.ej., with both line drawings
and photographs; grayscale as well as color images; con
both artificial and naturalistic scene images: Harel et al.,
2016, 2020; Hansen et al., 2018; Lowe et al., 2018; Cichy
et al., 2017) and across various task contexts (Hansen
et al., 2018; Lowe et al., 2018), indicating its centrality as
a source of information for scene perception. One reason
why spatial expanse may prove to be important for scene
perception is because of its potential link with navigability.
The spatial expanse of a scene conveys information not
only about the structure and geometry of the scene but
also about its function, a saber, the conceivable
398
Revista de neurociencia cognitiva
Volumen 34, Número 3
yo
D
oh
w
norte
oh
a
d
mi
d
F
r
oh
metro
h
t
t
pag
:
/
/
d
i
r
mi
C
t
.
metro
i
t
.
mi
d
tu
/
j
/
oh
C
norte
a
r
t
i
C
mi
–
pag
d
yo
F
/
/
/
/
3
4
3
3
9
7
1
9
8
5
0
1
9
/
j
oh
C
norte
_
a
_
0
1
8
1
0
pag
d
.
F
b
y
gramo
tu
mi
s
t
t
oh
norte
0
7
S
mi
pag
mi
metro
b
mi
r
2
0
2
3
possibilities for movement in space. And because the spa-
tial expanse of a scene, or rather its openness, is perceived
by humans as a continuous dimension (zhang, Houpt, &
Harel, 2019), closed and open scenes can thus be thought
of as two ends on an ease-of-navigation (navigability)
continuum, with closed environments posing more con-
straints on navigation than open ones. De este modo, it may be
argued that the degree of openness or enclosure also con-
fer constraints on navigability. It therefore stands to reason
that early neural responses to spatial expanse index infor-
mation about scene structure not only with regard to its
openness, but also as it defines navigable space. En efecto,
ERP amplitudes produced by spatial expanse variation
are very slightly, en todo caso, modulated by task demands
(Hansen et al., 2018), in line with the notion that naviga-
tional affordances are extracted automatically (Bonner &
Epstein, 2017).
The present work sought to establish the time course of
the extraction of navigational affordances and, específicamente,
to explicitly test the idea that early scene-evoked activity
represents the potential for navigation in a scene.
Although previous studies have uncovered the temporal
dynamics of spatial layout processing, they did not explic-
itly establish the relation between these neural signatures
and the extraction of navigational affordances. We argued
that if scene perception involves the extraction of func-
tional information for navigation, then the structure of
navigable space (es decir., navigational affordances) should be
encoded relatively early in processing and thus manifest in
early visually evoked ERPs. We hypothesized that the nav-
igational affordances of the visual environment would be
resolved at the P2 time window and, específicamente, that the
amplitude of the P2 component would capture the ease of
navigation in the environment given that the P2 distin-
guishes between open and closed scenes, (which can be
thought of as offering more or less navigability, respetar-
activamente). To test this hypothesis, we recorded ERPs from
participants while they passively viewed computer-
generated room scenes matched in visual complexity
(used in a previous fMRI study of navigability, see the work
of Bonner & Epstein, 2017). The rooms varied in the
number of pathways that afford movement in the local
ambiente: By simply changing the number of doors
(0 doors, 1 door, 2 doors, 3 doors) in the room, nosotros
were able to systematically control the number of move-
ment paths in the scene, while keeping the overall size
and shape of the environment constant. If encoding
navigational affordances engages stimulus-driven percep-
tual processes, then sensitivity to number of movement
paths in the environment should emerge within the first
250 msec of processing. Específicamente, neural activity during
the P2 time window should be modulated by the number
of doors in the scenes, with increasing P2 amplitude as a
function of constraints on navigability (es decir., decreasing
number of doors). Alternativamente, if manipulating of num-
ber of doors in the scene does not result in early pertur-
bations of neural activity and is observed later, then that
would lend support to the idea that encoding navigation
information reflects postperceptual processing rather
than stimulus-driven visual processing. Complementing
the hypothesis-driven univariate ERP analysis of P2 ampli-
tude, we also conducted a more data-driven, multivariate
analysis to assess the decoding of navigability information
con el tiempo, how early it emerges, and to what extent it rep-
resents the extraction of information about the local posi-
tion of navigability cues, or whether that information is
extracted in a position-invariant fashion.
MÉTODOS
Participantes
Thirty-six Wright State University students (21 women;
mean age: 20 años) participated in the study. Participantes
signed an informed written consent form according to the
guidelines of the institutional review board of Wright State
University and were compensated monetarily or with
course credit. All participants had normal or corrected-
to-normal visual acuity and no history of neurological dis-
ease, and all but one were right-handed. Six participants
were excluded from final analyses because of excessive
EEG artifacts.
Estímulos
The stimuli comprised 144 computer-generated images of
simple rectangular rooms, used in a previous neuroimag-
ing study (Bonner & Epstein, 2017). Every room scene
contained either a door or a painting on each of its three
visible walls, yielding eight navigability conditions accord-
ing to the number of door elements present combined
with the walls on which they appeared: 0 doors; 1 door
izquierda, bien, or center; 2 doors left–center, left–right, o
center–right; y 3 doors (ver figura 1 for examples of
estímulos). For each navigability condition, 18 unique rooms
were created by applying different textures to the walls,
making a total of 144 unique exemplars. The stimuli were
presented using Presentation software (Neurobehavioral
Sistemas, Cª, www.neurobs.com). Images were displayed
in 8-bit color at the center of a Dell LCD monitor (1920 ×
1080 píxeles) at a viewing distance of about 150 cm, sub-
tending 12 × 14 degrees of visual angle.
Experimental Design and Procedure
Participants viewed the 144 individual scene stimuli 8
veces (eight blocks), each block containing all 144 estímulos
(a total of 1152 ensayos). Scene stimuli were pseudoran-
domized within individual blocks and across the eight
blocks to prevent direct repetition of any stimulus, naviga-
bility condition, or texture. Stimuli were presented for
500 msec with a randomly jittered interstimulus interval
que van desde 1000 a 2000 mseg. Los participantes actuaron.
a fixation cross task, in which they were required to
Harel et al.
399
yo
D
oh
w
norte
oh
a
d
mi
d
F
r
oh
metro
h
t
t
pag
:
/
/
d
i
r
mi
C
t
.
metro
i
t
.
mi
d
tu
/
j
/
oh
C
norte
a
r
t
i
C
mi
–
pag
d
yo
F
/
/
/
/
3
4
3
3
9
7
1
9
8
5
0
1
9
/
j
oh
C
norte
_
a
_
0
1
8
1
0
pag
d
.
F
b
y
gramo
tu
mi
s
t
t
oh
norte
0
7
S
mi
pag
mi
metro
b
mi
r
2
0
2
3
Cifra 1. Examples of room
images used in the experiment.
The rooms varied in the ease
of navigability they afford,
operationalized by the number
of doors in the scene. Por
changing the number of doors
in the room (0 door, 1 door,
2 doors, 3 doors), we were able
to systematically control the
number of movement paths
in the scene, while keeping
the overall size and shape
of the environment constant.
Depicted here are eight scenes
(out of a total 144 escenas)
spanning all four navigability
condiciones. Note that the 1-door
and 2-doors conditions included
three variations in door
ubicación. The rooms also varied
in their spatial layout and
texture, yielding 18 distinto
room variants.
report whether the horizontal or vertical bar of the central
fixation cross (1-degree visual angle) lengthened on each
trial by a factor of 25%. Changes in the fixation cross were
randomized across trials and, hence, were independent
from the actual content of the underlying image, essen-
tially requiring the participants to pay very little, if any,
attention to the background images while completing
this task. This same task has been employed in previous
EEG studies of scene processing using naturalistic real-
world stimuli (Hansen et al., 2018; Harel et al., 2016), como
well as computer-generated room-like stimuli (Harel
et al., 2020). We verified that the task conditions were
indeed independent from the stimulus conditions by
conducting a Task × Condition ANOVA on participants’
accuracy scores (see Results section).
EEG Recording
Analog EEG signals were recorded using 64 Ag-AgCl pin-
type active electrodes (Biosemi ActiveTwo) mounted on
an elastic cap (Electro-Cap International, Cª) according
to the extended 10–20 system. EOGs were recorded from
two additional pairs of pupil-aligned electrodes: One pair
was placed on the skin over the right and left temporal
zygomatic bones; the other was placed over the nasal
zygomatic and frontal bones. Analog EEG data from all
electrodes were referenced to the common mode signal
electrode placed between electrodes PO3 and PO4.
Impedance of all channels was measured before the start
of each recording session, to ensure that all fell below
50 KOhms, and data for each electrode were inspected
to ensure no “bridging” artifacts were present.
In postprocessing, data were rereferenced to an elec-
trode placed on the tip of the nose. Both EEG and EOG
were sampled at 512 Hz with a resolution of 24 bits with
an active input range of −262 μV to +262 μV per bit.
The digitized EEG was saved and processed off-line.
Data Processing
The data were preprocessed using Brain Vision Analyzer 2
(Cerebro Productos GmbH). The raw data were first band-pass
filtered from 0.3 a 80 Hz (24 dB), with a second-order But-
terworth filter, and referenced to the tip of the nose. Eye
movements were corrected in the scalp electrode data
using an automated, restricted infomax ocular correction
independent component analysis (for details see the work
of Jung et al., 1998), effectively removing components
heavily correlated with HEOG and VEOG artifacts follow-
ing a meaned slope algorithm. Del 64 independiente
componentes (one per scalp electrode) calculated for each
partícipe, 14 ± 4 were removed on average. Remaining
artifacts exceeding ±100 μV in amplitude or containing a
change of over 100 μV in a period of 50 msec were
rejected. The preprocessed data were then segmented
into epochs ranging from −200 msec before to 800 mseg
after stimulus onset for all conditions. Participants’ data
were entirely excluded from analyses if fewer than 80%
of epochs could be retained following artifact removal.
Six participants’ data were thus excluded, and for the
remaining 30, an average of 1056 ± 63 (equivalent to
91% ± 6%) epochs were retained for all analyses.
ERP Univariate Analysis
The peaks of the P1, N1, and P2 were determined for each
individual participant, by automatically detecting the max-
imal amplitude in predetermined time windows in each
experimental condition (most positive peak between 80
y 130 mseg, most negative peak between 130 y
400
Revista de neurociencia cognitiva
Volumen 34, Número 3
yo
D
oh
w
norte
oh
a
d
mi
d
F
r
oh
metro
h
t
t
pag
:
/
/
d
i
r
mi
C
t
.
metro
i
t
.
mi
d
tu
/
j
/
oh
C
norte
a
r
t
i
C
mi
–
pag
d
yo
F
/
/
/
/
3
4
3
3
9
7
1
9
8
5
0
1
9
/
j
oh
C
norte
_
a
_
0
1
8
1
0
pag
d
.
F
b
y
gramo
tu
mi
s
t
t
oh
norte
0
7
S
mi
pag
mi
metro
b
mi
r
2
0
2
3
200 mseg, and most positive peak between 200 y
320 mseg, respectivamente). The mean latencies of the three
components across conditions were consistent with previ-
ous research (P1: 121 mseg [SEM = 3]; N1: 167 mseg
[SEM = 3]; P2: 230 mseg [SEM = 3]). Analyses were
restricted to posterior lateral sites (averaged across P7,
P5, P9, and PO7 for the left hemisphere, and across P8,
P6, P10, and PO8 for the right hemisphere), where maxi-
mal scene effects were previously observed (Harel et al.,
2016, 2020; Hansen et al., 2018). Mean peak amplitudes
(across participants) were analyzed using a two-way
within-subject ANOVA, with Hemisphere (izquierda, bien) y
Navigability (number of doors) as independent factors.
Because numbers of doors varied parametrically, tenemos
also conducted a linear trend analysis to assess whether
the amplitude of the ERP components scales with decreas-
ing number of doors, implemented as a two-way ANOVA
estructura (for further details on this analysis, see the
work of Pinhas, Tzelgov, & Ganor-Stern, 2012).
ERP Multivariate Analysis
Representational similarity analyses (RSA) were con-
ducted on the segmented EEG data, across all 64 canales
(but not external electrodes), for all artifact-free segments.
Data were exported from Brain Vision Analyzer and pro-
cessed in MATLAB ( Version 2016a, MathWorks) as four-
dimensional matrices (Channel × Condition × Segment ×
Time Point). Subject-level ERP data were submitted to
RSAs against two-model representational dissimilarity
matrices (RDMs), pertaining to the two-alternative
hypotheses: location-specific processing, in which the spa-
tial layout of the doors influenced processing (Figura 4A,
left column); and location-invariant processing, en el cual
only the “global” differences (es decir., between the total num-
ber of doors) influenced processing (Figura 4A, right col-
umn). Experimental conditions were coded as binary
triplets, with ones and zeros corresponding to doors
and paintings, respectivamente. Each digit within a triplet cor-
responded to one of the three possible locations (izquierda,
bien, or center). So, Por ejemplo, condition 1–0–0 would
correspond to a single door on the left, whereas 0–0–1
would correspond to a single door on the right. De este modo, el
identity and location of each door or painting was retained
in these condition codes. Codes were then entered in
ascending order as the rows and columns of a confusion
matrix to construct both model and neural RDMs. Tal como,
each cell corresponded to one pairing of stimulus condi-
ciones. For the location-specific model RDM, values in each
cell were derived from the number and location of doors
(and paintings) in its component row and column stimulus
condiciones (Bonner & Epstein, 2017). Mathematically,
these values were obtained by subtracting the similarity
between stimulus conditions from the maximum possible
semejanza. Similarity was computed as the sum of the dot
product of the condition codes corresponding to the row
and column of the cell, yielding a value between 0
(maximally dissimilar) y 3 (maximally similar). Differ-
ences between conditions in a given cell could therefore
be denoted by the difference between maximal similarity
and the calculated similarity. Mientras tanto, cell values in
the location-invariant model RDM were derived only from
the total number of doors (and paintings) present in each
pair of conditions and not their locations. Mathematically,
they were coded as the difference between sums of the row
and column stimulus conditions. Because the neural RDMs
were constructed using Pearson correlation coefficients of
determination ranging from 0 a 1 (see below), we divided
the cell values in the model RDMs (que van desde 0 a 3) por
3 to achieve common metrics for both. La resultante
model RDMs are presented in Figure 4.
Neural RDMs were constructed for each subject by first
calculating the Pearson correlation coefficient of determi-
nation between the ERP amplitudes of all possible pairs of
conditions across electrodes, separately at each time
punto. As these correlation coefficients denote pairwise
similarities between conditions, differences between con-
ditions were represented by taking the complement of the
coefficient of determination (1 − r2). In order to deter-
mine the level of agreement between the neural and
model RDMs for each subject, we took the Spearman cor-
relation between the two matrices (effectively comparing
the obtained pattern of activation across electrodes in the
neural RDMs to those predicted by each model RDM).
Finalmente, the Spearman correlation coefficients were sub-
mitted to cluster analyses (with a cluster induction param-
eter corresponding to a Type I error rate of .01) across
subjects to correct for multiple comparisons (Benjamini
& Hochberg, 2000).
RESULTADOS
Behavioral Performance: Verification of
Task Independence
We conducted a Task (horizontal change in fixation cross,
vertical change in fixation cross) × Stimulus (8 diferente
door conditions, see above) ANOVA on participants’ accu-
racy scores to verify that the orthogonal fixation task was
indeed independent of stimulus conditions. We found no
significant main effects of either factor (Tarea: F(1, 29) =
2.19, pag = .15, ηp
2 = .07; Stimulus: F(7, 203) = 1.85, pag =
.08, ηp
2 = .06), and notably, there was no interaction
between them, F(7, 203) = 1.02, pag = .42, ηp
2 = .03. En general,
this supports the orthogonality of the fixation cross task
to the stimulus (es decir., door) condiciones.
Univariate Analysis
To examine the extent to which navigational affordances,
operationalized as the number of doors in a room (es decir., el
number of pathways that afford movement in the local
ambiente), are encoded by early neuromarkers of
scene perception, we compared the amplitude of the early
Harel et al.
401
yo
D
oh
w
norte
oh
a
d
mi
d
F
r
oh
metro
h
t
t
pag
:
/
/
d
i
r
mi
C
t
.
metro
i
t
.
mi
d
tu
/
j
/
oh
C
norte
a
r
t
i
C
mi
–
pag
d
yo
F
/
/
/
/
3
4
3
3
9
7
1
9
8
5
0
1
9
/
j
oh
C
norte
_
a
_
0
1
8
1
0
pag
d
.
F
b
y
gramo
tu
mi
s
t
t
oh
norte
0
7
S
mi
pag
mi
metro
b
mi
r
2
0
2
3
visually evoked ERP components (P1, N1, and P2) en
response to the different room conditions (0 doors,
1 door, 2 doors, 3 doors). We conducted a two-way
repeated-measures ANOVA on the amplitude of the indi-
vidually defined peaks of each of the ERP components,
with Hemisphere (izquierda, bien), and Number of Doors
(0 doors, 1 door, 2 doors, 3 doors) as independent
factores. The significant results of these analyses are
reported in Figure 2, and the grand-average waveforms
are depicted in Figure 3.
P2 Component
We found that the P2 amplitude is sensitive to the number
of doors contained in the room scenes, expressed by a sig-
nificant main effect of Number of Doors, F(3, 87) = 4.29,
pag = .007, ηp
2 = .13. To assess the source of the main effect
of Number of Doors, we have conducted two follow-up
analiza: linear trend analysis and pairwise post hoc com-
parisons. The former analysis showed that P2 amplitude
scales with the number of doors present in the scene,
manifesting in a significant linear trend, F(1, 29) = 7.58,
pag = .01, ηp
2 = .21. Post hoc comparisons revealed a signif-
icantly higher P2 amplitude for the 0-doors condition
(m = 5.83 mV, SEM = 0.51) compared to the 3-doors
condición (m = 5.12 mV, SEM = 0.42) , t(29) = 2.45,
pag = .009, and a higher amplitude to the 1-door (m =
5.64 mV, SEM = 0.47) relative to the 2-door condition
(m = 5.36 mV, SEM = 0.45), t(29) = 2.09, pag = .02.
A main effect of Hemisphere was also observed, F(1,
29) = 18.90, pag = .001, ηp
2 = .40, with higher amplitude
in the right hemisphere (m = 6.74 mV, SEM = 0.62)
compared with the left hemisphere (m = 4.23, SEM =
0.42). The effect of Number of Doors, sin embargo, was not
found to significantly differ across hemispheres (interac-
tion of Hemisphere with Number of Doors, F(3, 87) =
1.36, pag = .26, ηp
2 = .05; as standard ERP practice, nosotros
depict both hemispheres in our figures).
Notablemente, the navigational affordances effect persisted
beyond the P2 time window. Figure 3B depicts the dif-
ference ERP waveform contrasting the 3-doors and the
0-doors conditions across the whole scalp using current
source density (CSD) topographical maps. As can be
seen, the signal depicting difference in perceiving min-
imally navigable scenes to maximally navigable scenes
was present from around 200 msec to 350 msec post-
stimulus onset.
N1 Component
An analysis of the N1 component revealed the effect of
Number of Doors did not reach significance, F(3, 87) =
3.02, pag = .06, ηp
2 = .09. No significant linear trend was
noted as a function of number of doors, F(1, 29) < 1.00,
ηp
2 = .02, and the difference between the no-doors condi-
tion and the three doors condition was not found to be
significant (planned t test: t(29) = 0.85, p = .20). No
significant effects of hemisphere, F(1, 29) = 1.30, p =
Figure 2. Grand average ERP
univariate analysis results. Top
row: mean P2 peak amplitudes
in response to the four
navigability conditions (from
least navigable to most
navigable: 0 door, 1 door,
2 doors, 3 doors) presented
separately for the left and right
hemisphere. Bottom row: mean
P1 and N1 peak amplitudes for
the navigability conditions
presented separately for the left
and right hemispheres. Error
bars indicate between-subjects
SE; all data are plotted for the
posterior lateral sites. P2
amplitude showed a significant
linear increase in magnitude
of response as a function of
number of doors, with less
doors resulting in higher
amplitude. Notably, no such
increase was observed on
the amplitude of the P1 or
N1 components (see text
for details).
402
Journal of Cognitive Neuroscience
Volume 34, Number 3
l
D
o
w
n
o
a
d
e
d
f
r
o
m
h
t
t
p
:
/
/
d
i
r
e
c
t
.
m
i
t
.
e
d
u
/
j
/
o
c
n
a
r
t
i
c
e
-
p
d
l
f
/
/
/
/
3
4
3
3
9
7
1
9
8
5
0
1
9
/
j
o
c
n
_
a
_
0
1
8
1
0
p
d
.
f
b
y
g
u
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
l
D
o
w
n
o
a
d
e
d
f
r
o
m
h
t
t
p
:
/
/
d
i
r
e
c
t
.
m
i
t
.
e
d
u
/
j
/
o
c
n
a
r
t
i
c
e
-
p
d
l
f
/
/
/
/
3
4
3
3
9
7
1
9
8
5
0
1
9
/
j
o
c
n
_
a
_
0
1
8
1
0
p
d
.
f
b
y
g
u
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
Figure 3. (A) Group-averaged waveforms for the four navigability conditions (from least navigable to most navigable: 0 door, 1 door, 2 doors,
3 doors) presented separately for the left and right hemisphere at posterior lateral sites. (B) Scalp CSD topographical maps depicting the difference
between the 3-doors and the 0-door conditions show a persistent effect of navigability.
2 = .04, or a Hemisphere × Number of Doors inter-
.26, ηp
action, F(3, 87) < 1.00, ηp
2 = .01, were observed.
P1 Component
Analysis of the amplitude of the P1 component showed
the effects of Number of Doors, and the Number of Doors
x Hemisphere did not reach significance, F(3, 87) = 2.56,
p = .08, ηp
2 = .03, respec-
tively. No significant main effect of Hemisphere was
observed, F(1, 29) = 3.94, p = .06, ηp
2 = .08; F(3, 87) = 1.03, p = .37, ηp
2 = .12.
We also performed a secondary analysis investigating
the extent to which the specific location of the door, which
might provide local information about potential move-
ment path, could have an effect on the early visually
evoked ERPs.4 We conducted two separate two-way ANO-
VAs with Hemisphere and Door Location as independent
variables: one ANOVA for the single-door condition and
another one for the two-doors conditions. For the
single-door condition, we were not able to find any signif-
icant effects of Door Location or Door Location x Hemi-
sphere on either the P1, N1, or P2 peak amplitudes (all
ps > .30). A significant effect of Hemisphere was found
on the amplitude of the P2 and P1 components (P2: F(1,
2 = .40; P1: F(1, 29) = 4.58, pag =
2 = .13), with higher amplitude in the right than in
29) = 18.52, pag = .001, ηp
.04, ηp
the left hemispheres.
For the two-doors condition, a significant main effect of
Door Location was found on the amplitude of the P2 com-
ponent, F(2, 58) = 3.70, pag = .04, ηp
2 = .11, and the interac-
tion between Door Location and Hemisphere was not
found to be significant, F(2, 58) < 1.00, ηp
2 = .01. Post
hoc comparisons showed significantly ( p < .05,
Bonferroni-corrected) lower amplitude to rooms in which
the doors were located in the center and right positions
compared to scenes with doors on the left and right, albeit
the latter were not significantly different than scenes with
left- and center-positioned doors.5 No significant effects of
Door Location or Door Location x Hemisphere were
observed on the N1 amplitude (all ps > .78). For the P1
component, we found a significant main effect of Door
Location, F(2, 58) = 6.00, pag = .006, ηp
2 = .17, mientras que el
interaction between Door Location and Hemisphere was
not found to be significant, F(2, 58) < 1.00, ηp
2 = .00. Lastly,
a main effect of Hemisphere was found for both the P1 and
P2 components (P1: F(1, 29) = 4.21, p = .05, ηp
2 = .13; P2:
F(1, 29) = 17.48, p = .001, ηp
2 = .37), reflecting a right hemi-
sphere advantage (see main analysis reported above).
Harel et al.
403
In summary, we did not observe consistent effects of the
specific location of the door/s in the scene. This may stem
from our univariate approach lacking the sensitivity to
detect what might be subtle differences in the visual input
(for similar results in fMRI, see the work of Bonner &
Epstein, 2017). To address this possibility, we conducted
a multivariate analysis, which allowed us to examine (a)
the extent to which both position-dependent and
position-invariant information can be extracted from the
EEG signal, and (2) the different time windows during
which the signals may be observed.
Multivariate Analysis
To determine the time course of location-specific and
location-invariant navigability-related processing, RSAs
were conducted by comparing the obtained neural RDMs
from each participant at each time point with two-model
RDMs, one corresponding to each hypothesis (see
Methods section and Figure 4A). The pattern of “activa-
tion” across electrodes within their averaged ERP
waveforms was correlated with differences between con-
ditions predicted by either or both of the two-model
RDM. This was followed by a cluster analysis across sub-
jects, allowing us to determine the time intervals in which
the data were best explained by one of the predicted
models. We found significant Spearman correlations
(Figure 4B) between the neural RDMs and the location-
specific RDM at several time windows: from 134–170 msec
(cluster-corrected values of p = .035, two-tailed), 193–
275 msec (cluster-corrected values of p = .0024, two-tailed),
and 295–380 msec (equivalent to all cluster-corrected
values of p = .004, two-tailed). Notably, the cluster analy-
ses also uncovered significant Spearman correlations
between the neural RDMs and location-invariant RDM,
from 196 to 237 msec (Figure 4B). Interestingly, this time
window corresponds with the P2 time window. Together,
the multivariate analyses suggest that local featural differ-
ences between stimuli (i.e., the number and location of
doors) in each navigability condition are processed as
early as 134 msec after stimulus onset, and more global
featural differences (0 doors, 1 door, 2 doors, 3 doors)
l
D
o
w
n
o
a
d
e
d
f
r
o
m
h
t
t
p
:
/
/
d
i
r
e
c
t
.
m
i
t
.
e
d
u
/
j
/
o
c
n
a
r
t
i
c
e
-
p
d
l
f
/
/
/
/
3
4
3
3
9
7
1
9
8
5
0
1
9
/
j
o
c
n
_
a
_
0
1
8
1
0
p
d
.
f
b
y
g
u
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
Figure 4. Results of the
multivariate analysis. (A) The
two-model RDM used for
the analyses, pertaining to the
two-alternative hypotheses:
location-specific processing,
in which the local featural
differences between conditions
(i.e., between doors and
paintings) influenced
processing (left column); and
location-invariant processing,
in which only the “global”
differences (i.e., between the
total number of doors)
influenced processing (right
column). (B) Time course of
decoding (measured using
Spearman correlation
coefficients), comparing the
obtained pattern of activation to
that predicted by the two
models: location-specific (blue)
and location-invariant (red).
Solid lines indicate significant
decoding ( p < .01, corrected:
see text for details).
404
Journal of Cognitive Neuroscience
Volume 34, Number 3
are most likely processed after local featural differences,
no earlier than 196 msec following stimulus onset.6
DISCUSSION
The current study provides novel evidence that the brain
codes information about the potential for navigation in the
scene as early as 200 msec after stimulus onset, and that
this coding involves both a global, position-invariant signal
about the overall navigability of the space and local infor-
mation regarding the positions of navigational pathways. A
standard univariate ERP analysis revealed that the ampli-
tude of the scene-selective P2 ERP component was higher
in response to images of rooms with no doors compared
to rooms with three doors, analogous to the higher P2
amplitude in response to closed relative to open scenes
reported previously (Harel et al., 2016, 2020; Hansen
et al., 2018). Furthermore, P2 peak amplitude scaled
linearly with the constraints on navigation: The more con-
straints on navigation (i.e., less doors in a room), the
higher was its amplitude. And although the effect of navi-
gability was most pronounced on the P2 ERP component,
the difference in amplitude between no-doors and three-
doors condition continued beyond the P2 time window,
lasting for additional 200 msec. Complementing these
findings, a multivariate analysis revealed that the P2 time
window contains significant information about naviga-
tional affordances: It is the first time period in which
position-invariant navigability information is extracted, in
addition to information regarding the location of the diag-
nostic feature (a specific door), which emerges earlier and
persists at this time window as well. Together, these find-
ings suggest that diagnostic information about the poten-
tial for navigation in a scene is present at the early stages of
visual processing and extracted as early as 220 msec after
stimulus onset.
Based on the current findings, we suggest that perceiv-
ing visual environments and navigating through them
should not necessarily be considered as two separate pro-
cesses, but rather as two points on a single continuum,
scene perception being the first step in a sequence of stages
that support navigation. At the neural level, this outlook
has two spatiotemporal corollaries. First, scene-selective
cortex should not only be engaged in the extraction of
scene-diagnostic features but should also carry informa-
tion about the potential for navigation in the scene.
Second, navigability-related neural activity should be
observed in the early stages of visual scene processing.
The current ERP study joins the original fMRI study, which
used the current scene stimuli (Bonner & Epstein, 2017)
to establish these two points. Bonner and Epstein (2017)
showed that information about the potential for move-
ment in the scenes is, indeed, represented in scene-
selective OPA.7 Our study shows that this same information
is represented as early as 200 msec poststimulus onset,
the same latency during which global properties of the
scene are extracted (Harel et al., 2016, 2020; Hansen
et al., 2018). Because the two studies use the exact scene
stimuli, they form a crucial link in connecting the spatial
and temporal aspects of perceptual processing of the
potential for action in scenes. Furthermore, in a follow-
up study, Bonner and Epstein (2018) reanalyzed their
imaging data and showed using deep convolutional net-
works that the affordance properties of scenes could be
represented through just a few stages of purely feedfor-
ward hierarchical computations, implying that computa-
tions of navigational affordances in OPA could be
achieved rapidly, in line with the current findings.
Given the limited temporal precision of fMRI, the func-
tional nature of the observed OPA activation cannot be
determined unequivocally; whereas OPA activity could
indeed reflect stimulus-driven processing of navigability
information available in the scene, it could also potentially
reflect recurrent feedback from posterior parietal cortex as
part of the occipito-parietal circuit (Kravitz, Saleem, et al.,
2011). Our current results suggest the former, rather than
the latter, alternative is more probable, as we show that
by 200 msec, sufficient evidence has accumulated for
determining the potential for navigation in the scene.
Moreover, the fact that the navigability effect on the
EEG manifests without any apparent need for encoding
or planning movement (see below) strengthens the
notion that navigational affordances are encoded manda-
torily, as originally suggested by Bonner and Epstein
(2017). It is important to note, however, that despite
the similarities between the studies, we cannot unequiv-
ocally conclude that OPA is the neural generator of the
observed effects of navigational affordances on the P2
component. We have not performed source localization
in the current study, as the relationship between ERP gen-
erator locations and scalp electrodes is complex (Nunez &
Srinivasan, 2006), and oftentimes the answers to which
cortical area generates a certain ERP effect could vary as
a function of the mathematical solution favored by the
researcher (for a comprehensive discussion, see the work
of Luck, 2014, Online Chapter 14; for an empirical demon-
stration of the limits of localization, see the work of
Petrov, 2012). Future research combining ERP and fMRI
(e.g., simultaneously recording ERPs in an MRI scanner;
see the work of Sadeh, Podlipsky, Zhdanov, & Yovel,
2010) is needed to determine the relationship between
OPA activity and the P2 sensitivity to navigationally rele-
vant information. Although the current data suggest that
navigational affordances are encoded rapidly, it is still an
open question just how rapid “rapid” is. Some recent
studies show earlier encoding of navigable space, specifi-
cally, around 100–120 msec poststimulus onset. A study
combining MEG and fMRI showed significant scene
boundary encoding (i.e., sensitivity to navigation-
constraining large-scale geometrical boundaries) in OPA
with corresponding MEG response patterns emerging as
early as 65 msec and peaking at about 100 msec poststim-
ulus onset (Henriksson et al., 2019). In a similar vein, dif-
ferential processing of closed and open scenes (which
Harel et al.
405
l
D
o
w
n
o
a
d
e
d
f
r
o
m
h
t
t
p
:
/
/
d
i
r
e
c
t
.
m
i
t
.
e
d
u
/
j
/
o
c
n
a
r
t
i
c
e
-
p
d
l
f
/
/
/
/
3
4
3
3
9
7
1
9
8
5
0
1
9
/
j
o
c
n
_
a
_
0
1
8
1
0
p
d
.
f
b
y
g
u
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
confer distinct navigational affordances, see below) has
been reported to manifest not only at the P2 time window
but also earlier, around 120 msec poststimulus onset
(Lowe et al., 2018). Findings from our multivariate analy-
ses support these studies demonstrating early coding of
spatial geometry. Information about the direction of
potential movement (i.e., specific door position) was
found to be represented as early as 140 msec poststimulus
onset, suggesting early encoding of navigational affor-
dances. Notably, position-invariant information, which is
perhaps more related to spatial expanse was found later,
at the P2 time window, between 200 and 240 msec. Con-
sequently, we suggest that spatial affordances are picked
up at different points in times, each characterized by sep-
arate neural representations: Location-specific affordance
information is encoded as early as 134 msec, persists until
170 msec, and re-emerges around 200 msec. This is
followed by more global scene representations that are
affected by the affordances of the space, namely, when
there are more affordances (i.e., less constraints on move-
ment overall), the global space becomes larger and more
“open.” This second stage reflects global processing of the
scene, entailing the integration of features across the
entire scene (for discussion, see the work of Harel et al.,
2020). As we note above, this later time window is coinci-
dent with the univariate results highlighting the sensitivity
of the P2 component to navigation-related information
(although the multivariate analyses detect the onset of this
effect slightly earlier than the peak analyses, this is to be
expected, given that the P2 only peaks around 230 msec).
Our focus on the P2 component, suggesting it is the
electrophysiological marker of navigational affordances,
follows our previous work highlighting the P2 as an index
of the processing of high-level global scene information.
The posterior P2 is the first ERP component to show evi-
dence of scene-selectivity, with higher amplitude to
scenes compared with faces and objects (Harel et al.,
2016), and the only visually evoked component to be mod-
ulated by scene inversion, as would be expected if global
scene information is indeed extracted during this period
(Harel & Al Zoubi, 2019). Furthermore, P2 amplitude is
sensitive to GSPs, such as spatial expanse (closed/open)
and naturalness (man-made/natural), and these effects
are automatic, evident across a variety of stimulus presen-
tation conditions, and are largely unperturbed by manipu-
lations of local texture (Harel et al., 2016, 2020; Hansen
et al., 2018). Together, these studies suggest that the
scene-selective P2 and the P2 time window in general
(approx. 200–250 msec) is indicative of and essential for
the processing of the global spatial structure of scenes
(see also the work of Kaiser, Häberle, & Cichy, 2020;
Kaiser, Turini, & Cichy, 2019; Cichy et al., 2017). Spatial
structure is used here broadly as an umbrella term to cap-
ture related concepts, such as scene layout, expanse, and
boundary. Notably, spatial structure is one of the key cat-
egories of GSPs proposed by Greene and Oliva (2009) to
be central for rapid scene categorization. According to
Greene and Oliva, rapid scene categorization is not pri-
marily mediated through objects and parts, but rather
through global, ecological properties that describe spatial
and functional aspects of scene space (such as navigability
or mean depth). Based on Greene and Oliva’s theory, we
hypothesized that closed and open scenes not only
describe the large-scale geometry of the scenes (i.e., its
spatial structure) but may also capture the functional prop-
erty of navigability. As such, closed and open scenes rep-
resent two ends of a continuum wherein enclosed spaces
posit more constraints on movement than open spaces.
And because closed scenes consistently evoke a stronger
neural response than open scenes (Harel et al., 2020;
Hansen et al., 2018; Lowe et al., 2018), we predicted that
scenes conferring more constraints on navigability should
evoke a higher P2 amplitude than scenes that confer less
such constraints. This prediction bore out, not only in its
dichotomous form (0-doors vs. 3-doors) but also as a con-
tinuous, parametric effect of navigability, evident in a lin-
ear decrease of P2 amplitude as a function of the number
of doors, as well as in a significant decoding of number of
doors independent of their local location around the P2
time window, 200–240 msec poststimulus onset. It is still
an open question to what extent the P2 exclusively indexes
global scene information, or whether it also incorporates
the processing of local scene information. Specifically, in
spite of the studies described above, we still found some
modulation of the P2 amplitude by door location (at least
for the two-door condition, see secondary univariate anal-
yses above), which may suggest it is not entirely indepen-
dent of local diagnostic information. The link, however,
between P2 amplitude and local information is not
straightforward, as no modulation was found for the single
door condition, and the modulation of the two-door con-
dition was not systematic and, furthermore, was not
related to hemisphere, as one might have expected given
a retinotopic mapping of the visual field. Future research
will be needed to elucidate the relationship between local
and global scene processing and how these processes
map onto the P2 amplitude.
Our study reveals another interesting characteristic of
the P2. We found that the P2 navigability effect was later-
alized to the right hemisphere; that is, the effect of the
number of doors on P2 amplitude was more pronounced
in the right than in the left hemisphere. This interaction
resonates and converges with previous ERP studies of
P2, which report laterality effects on the P2 amplitude,
with overall higher amplitude in the right hemisphere,
as well as specific GSPs being more discriminable in the
right hemisphere. Specifically, both scene naturalness
and spatial expanse were reported to have a greater effect
in the right hemisphere (Harel et al., 2016, 2020; Hansen
et al., 2018). The finding that the right hemisphere is more
involved than the left in the processing of global scene
information extends previous research on local/global
processing and hemispheric asymmetries and is in line
with the long-lived proposal that the right hemisphere
406
Journal of Cognitive Neuroscience
Volume 34, Number 3
l
D
o
w
n
o
a
d
e
d
f
r
o
m
h
t
t
p
:
/
/
d
i
r
e
c
t
.
m
i
t
.
e
d
u
/
j
/
o
c
n
a
r
t
i
c
e
-
p
d
l
f
/
/
/
/
3
4
3
3
9
7
1
9
8
5
0
1
9
/
j
o
c
n
_
a
_
0
1
8
1
0
p
d
.
f
b
y
g
u
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
specializes in global processing (Brederoo, Nieuwenstein,
Lorist, & Cornelissen, 2017; Flevaris & Robertson, 2016;
but see the work of Wiesmann, Friederici, Singer, &
Steinbeis, 2020). Thus, the right hemisphere advantage
for GSP effects further supports our proposal that P2
indexes the processing of the spatial structure of the scene
via the processing of global scene information.
Our findings point to the putative connection between
spatial structure and navigability, which, arguably, may
serve as the mechanism by which perceptual information
is transformed into action-relevant information. Given the
close link between spatial structure and navigability, our
findings raise two intriguing questions. First, are navigabil-
ity and spatial structure (or spatial expanse, in the case of
closed vs. open scenes) one and the same thing? Can the
two constructs be used synonymously, or are they two
independent dimensions? Second, given that they are
independent, does the neural encoding of navigational
affordances reflect intermediate levels of representation
(i.e., GSPs), low-level image statistics (Groen, Silson, &
Baker, 2017) or rather the extraction of higher-level eco-
logical scene properties (for discussion, see the work of
Malcolm, Groen, & Baker, 2016)? Our current design can-
not directly address these questions, as we only used
closed indoor scenes (rooms). However, future research
will be able to shed further light on this issue. To test
the independence of the two dimensions, one would have
to vary the amount of movement a scene affords, in both
closed and open scenes. If spatial expanse and navigability
are independent, then varying navigability should have a
similar effect on the neural responses to open scenes (this
could be manipulated, e.g., by adding an increasing num-
ber of obstacles like boulders, or by varying the number of
paths in an open field). In addition, an alternative, more
naturalistic approach could use a large set of real-world
scene images (instead of highly controlled artificial
scenes), in which people would rank these scenes on both
spatial expanse (Zhang et al., 2019) and navigability
(Bonner & Epstein, 2017, Experiment 2) and then in a sep-
arate electrophysiological study integrate these rankings
with the neural response patterns to determine the rela-
tive contribution of each dimension and other image prop-
erties to the neural representations using computational
model-based approaches (e.g., Lescroart & Gallant,
2019; Bonner & Epstein, 2018; Cichy, Kholsa, Pantazis, &
Oliva, 2017; Lescroart, Stansbury, & Gallant, 2015).
At a broader theoretical level, our results support a tight
link between perception and action, a hallmark of senso-
rimotor and embodied cognition theories (e.g., Jelić, Tieri,
De Matteis, Babiloni, & Vecchiato, 2016; Wilson, 2002;
Clark, 1999). One aspect of these theories in the context
of navigation is the constant need for visual updating when
one explores an environment to minimize prediction error
(Kaplan & Friston, 2018; Hassabis & Maguire, 2009; Kurby
& Zacks, 2008; Zacks, Speer, Swallow, Braver, & Reynolds,
2007). The idea is that exploring the environment requires
continuously modeling the potential outcomes of the
intended action, with affordances serving this function
by constraining visual perception to reflect experience-
dependent, observer-relevant information (Sestito, Flach,
& Harel, 2018). At the neural level, this should translate to
dynamic changes in continuous activity in early visual
areas as one walks around and explores their surround-
ings, even before execution of action (e.g., movement;
Jelić et al., 2016). In line with this idea, a recent study using
mobile brain/body imaging technology in which partici-
pants actively walked in a highly engaging immersive
virtual environment demonstrated that environmental
affordances are extracted and encoded throughout the
entirety of the act of exploring one’s surroundings, starting
from early perceptual stages (P1-N1 complex) all the way
to motor planning and execution (Djebbara, Fich, Petrini,
& Gramann, 2019). Notably, our study adds several key
observations to these findings. First, ambulatory, active,
and continuous exploration of the environment is not
necessary for observing electrophysiological responses
representing the extraction of navigational affordances.
Participants in our study were stationary, sitting in front
of a computer screen, and watched briefly presented,
minimalistic room images. The fact that one can observe
similar electrophysiological responses as a function of nav-
igability even without any movement or the presence of an
immersive environment implies that scene affordances are
extracted across multiple contexts and task demands (for a
similar finding with spatial expanse, see the work of
Hansen et al., 2018). Arguably, a lifetime of experience
navigating in the world results in automatic activation of
sensorimotor scene representations when presented with
visual environments, even if these environments are
sparse, minimalistic scenes deprived of rich detail (for
the role of experience in perceiving novel scene affor-
dances, see the work of Sestito, Harel, Nador, & Flach,
2018). The idea that navigational affordances are pro-
cessed mandatorily is also supported by the finding that
navigability effects are evident even without any explicit
task-context or relevant task demands, as participants in
our study were not required to either move about the envi-
ronment, imagine themselves moving about it, or make
any explicit judgments regarding its potential for naviga-
tion. The fact that we found navigability-based modulations
although no movement in space was performed, nor was
movement in space directly relevant to the task, supports
the conclusion that extracting navigational affordances is
rapid, mandatory, and task-independent. Furthermore,
the similar patterns of results between the two studies sug-
gest that our current laboratory-based findings are likely to
generalize to real-world, realistic settings and thus expand
their validity and utility for future research.
In summary, this study demonstrates that navigational
affordances are extracted at the early, perceptual stages of
visual scene processing, suggesting a close link between
scene perception and navigation. Information about the
potential for navigation in the scene is extracted rapidly
and automatically, without any explicit task or movement
Harel et al.
407
l
D
o
w
n
o
a
d
e
d
f
r
o
m
h
t
t
p
:
/
/
d
i
r
e
c
t
.
m
i
t
.
e
d
u
/
j
/
o
c
n
a
r
t
i
c
e
-
p
d
l
f
/
/
/
/
3
4
3
3
9
7
1
9
8
5
0
1
9
/
j
o
c
n
_
a
_
0
1
8
1
0
p
d
.
f
b
y
g
u
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
requirements. Complementing prior neuroimaging stud-
ies showing OPA encodes the structure of navigable
space, the current work establishes the temporal dynam-
ics of processing navigational affordances. Significant nav-
igability information is present in two time windows: an
early time window sensitive to the specific position of
navigability diagnostic stimulus features, and a later
one, which incorporates both position-specific and
position-invariant information. The later time window
overlaps with the univariate P2 ERP component, reflect-
ing global processing of scene structure. Finally, the cur-
rent findings are in line with sensorimotor accounts of
perception, suggesting that perceiving visual environ-
ments and navigating through them should not necessar-
ily be considered as two separate processes, but rather as
two integrated processes.
Reprint requests should be sent to Assaf Harel, Department of
Psychology, Wright State University, 335 Fawcett Hall, 3640 Col.
Glenn Highway, Dayton, Ohio 45435, or via e-mail: assaf
.harel@wright.edu.
4. The reason this analysis was considered secondary is that
we did not expect to find door location effects in the univariate
analysis based on Bonner and Epstein’s (2017) fMRI study. No
robust effects of door location were observed in that study in
the univariate response magnitude analysis, while multivariate
analysis of response patterns did not show sufficient sensitivity
to report such effects.
5. Note that whereas in the 1-door condition, the comparison
between conditions is relatively straightforward with door loca-
tion being the diagnostic cue for movement, in the 2-doors con-
ditions, this is less obvious, as the “odd one out” is not the
door, but rather its absence—a single painting. This difficulty
in interpretation is further exacerbated in the case of no signif-
icant interactions with Hemisphere, meaning that painting loca-
tion is not the consistent source of the effect.
6.
It should be noted, however, that significant clusters do not
contain contiguous intervals of significant Spearman correlation
coefficients; rather, clusters denote the intervals within which
there is a conditional probability of p that the interval contains
at least one significant coefficient, given the Type 1 error rate
(cluster induction parameter) for evaluating the coefficients
independently.
7. Notably, Bonner and Epstein (2017) found OPA sensitivity
to navigability information not only using current stimuli, which
are computer-generated, but also using naturalistic images of
indoor scenes.
Diversity in Citation Practices
REFERENCES
Retrospective analysis of the citations in every article pub-
lished in this journal from 2010 to 2021 reveals a persistent
pattern of gender imbalance: Although the proportions of
authorship teams (categorized by estimated gender identi-
fication of first author/last author) publishing in the Jour-
nal of Cognitive Neuroscience ( JoCN ) during this period
were M(an)/M = .407, W(oman)/M = .32, M/ W = .115,
and W/ W = .159, the comparable proportions for the arti-
cles that these authorship teams cited were M/M = .549,
W/M = .257, M/ W = .109, and W/ W = .085 (Postle and
Fulvio, JoCN, 34:1, pp. 1–3). Consequently, JoCN encour-
ages all authors to consider gender balance explicitly when
selecting which articles to cite and gives them the oppor-
tunity to report their article’s gender citation balance.
Notes
1. The link between scene recognition and navigation reso-
nates the putative link between perception and action pro-
posed by sensorimotor accounts of perception, which posit that
the potential for action in the environment, also known as affor-
dances, is conveyed by the visual stimulus itself (e.g., Gibson,
1979).
2. This is in fact an inherent limitation of fMRI because of its
low temporal resolution. For a more general discussion on the
inferential challenges of fMRI, see the work of Ghuman and
Martin (2019).
3. The extent to which these early signatures reflect the
extraction of local image statistics or more global scene proper-
ties is still an open question. For instance, scene inversion
effects are only observed around 220–250 msec poststimulus
onset (Kaiser, Häberle, & Cichy, 2020; Harel & Al Zoubi,
2019), consistent with the idea that global scene structure infor-
mation is extracted later than 100 msec.
Benjamini, Y., & Hochberg, Y. (2000). On the adaptive
control of the false discovery rate in multiple testing with
independent statistics. Journal of Educational and
Behavioral Statistics, 25, 60–83. https://doi.org/10.3102
/10769986025001060
Bonner, M. F., & Epstein, R. A. (2017). Coding of navigational
affordances in the human visual system. Proceedings of the
National Academy of Sciences, U.S.A., 114, 4793–4798.
https://doi.org/10.1073/pnas.1618228114, PubMed: 28416669
Bonner, M. F., & Epstein, R. A. (2018). Computational
mechanisms underlying cortical responses to the affordance
properties of visual scenes. PLoS Computational Biology,
14, e1006111. https://doi.org/10.1371/journal.pcbi.1006111,
PubMed: 29684011
Brederoo, S. G., Nieuwenstein, M. R., Lorist, M. M., &
Cornelissen, F. W. (2017). Hemispheric specialization for
global and local processing: A direct comparison of linguistic
and non-linguistic stimuli. Brain and Cognition, 119, 10–16.
https://doi.org/10.1016/j.bandc.2017.09.005, PubMed:
28923763
Cichy, R. M., Khosla, A., Pantazis, D., & Oliva, A. (2017).
Dynamics of scene representations in the human brain
revealed by magnetoencephalography and deep neural
networks. Neuroimage, 153, 346–358. https://doi.org/10.1016
/j.neuroimage.2016.03.063, PubMed: 27039703
Clark, A. (1999). An embodied cognitive science? Trends in
Cognitive Sciences, 3, 345–351. https://doi.org/10.1016/S1364
-6613(99)01361-3
Dilks, D. D., Julian, J. B., Kubilius, J., Spelke, E. S., & Kanwisher,
N. (2011). Mirror-image sensitivity and invariance in object
and scene processing pathways. Journal of Neuroscience, 31,
11305–11312. https://doi.org/10.1523/JNEUROSCI.1935-11
.2011, PubMed: 21813690
Dilks, D. D., Julian, J. B., Paunov, A. M., & Kanwisher, N. (2013).
The occipital place area is causally and selectively involved in
scene perception. Journal of Neuroscience, 33, 1331–1336.
https://doi.org/10.1523/JNEUROSCI.4081-12.2013, PubMed:
23345209
408
Journal of Cognitive Neuroscience
Volume 34, Number 3
l
D
o
w
n
o
a
d
e
d
f
r
o
m
h
t
t
p
:
/
/
d
i
r
e
c
t
.
m
i
t
.
e
d
u
/
j
/
o
c
n
a
r
t
i
c
e
-
p
d
l
f
/
/
/
/
3
4
3
3
9
7
1
9
8
5
0
1
9
/
j
o
c
n
_
a
_
0
1
8
1
0
p
d
.
f
b
y
g
u
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
Djebbara, Z., Fich, L. B., Petrini, L., & Gramann, K. (2019).
Sensorimotor brain dynamics reflect architectural
affordances. Proceedings of the National Academy of
Sciences, U.S.A., 116, 14769–14778. https://doi.org/10.1073
/pnas.1900648116, PubMed: 31189596
Ekstrom, A. D. (2015). Why vision is important to how we
navigate. Hippocampus, 25, 731–735. https://doi.org/10.1002
/hipo.22449, PubMed: 25800632
Ekstrom, A. D., Spiers, H. J., Bohbot, V. D., & Rosenbaum, R. S.
(2018). Human spatial navigation. Princeton University
Press. https://doi.org/10.2307/j.ctvc773wg
Epstein, R. A., Higgins, J. S., Jablonski, K., & Feiler, A. M. (2007).
Visual scene processing in familiar and unfamiliar
environments. Journal of Neurophysiology, 97, 3670–3683.
https://doi.org/10.1152/jn.00003.2007, PubMed: 17376855
Epstein, R. A., Higgins, J. S., & Thompson-Schill, S. L. (2005).
Learning places from views: Variation in scene processing as a
function of experience and navigational ability. Journal of
Cognitive Neuroscience, 17, 73–83. https://doi.org/10.1162
/0898929052879987, PubMed: 15701240
Flevaris, A. V., & Robertson, L. C. (2016). Spatial frequency
selection and integration of global and local information in
visual processing: A selective review and tribute to Shlomo
Bentin. Neuropsychologia, 83, 192–200. https://doi.org/10
.1016/j.neuropsychologia.2015.10.024, PubMed: 26485158
Ghuman, A. S., & Martin, A. (2019). Dynamic neural
representations: An inferential challenge for fMRI. Trends in
Cognitive Sciences, 23, 534–536. https://doi.org/10.1016/j.tics
.2019.04.004, PubMed: 31103440
Gibson, J. J. (1979). The ecological approach to visual
perception. Boston, MA: Houghton, Mifflin and Company.
Greene, M. R., & Oliva, A. (2009). The briefest of glances the
time course of natural scene understanding. Psychological
Science, 20, 464–472. https://doi.org/10.1111/j.1467-9280
.2009.02316.x, PubMed: 19399976
Groen, I. I. A., Ghebreab, S., Lamme, V. A. F., & Scholte, H. S.
(2016). The time course of natural scene perception with
reduced attention. Journal of Neurophysiology, 115, 931–946.
https://doi.org/10.1152/jn.00896.2015, PubMed: 26609116
Groen, I. I. A., Ghebreab, S., Prins, H., Lamme, V. A., & Scholte,
H. S. (2013). From image statistics to scene gist: Evoked
neural activity reveals transition from low-level natural image
structure to scene category. Journal of Neuroscience, 33,
18814–18824. https://doi.org/10.1523/JNEUROSCI.3128-13
.2013, PubMed: 24285888
Groen, I. I. A., Silson, E. H., & Baker, C. I. (2017). Contributions
of low- and high-level properties to neural processing of
visual scenes in the human brain. Philosophical
Transactions of the Royal Society of London, Series B:
Biological Sciences, 372, 20160102. https://doi.org/10.1098
/rstb.2016.0102, PubMed: 28044013
Hansen, N. E., Noesen, B. T., Nador, J. D., & Harel, A. (2018).
The influence of behavioral relevance on the processing of
global scene properties: An ERP study. Neuropsychologia,
114, 168–180. https://doi.org/10.1016/j.neuropsychologia
.2018.04.040, PubMed: 29729276
Harel, A., & Al Zoubi, H. (2019). Early electrophysiological
correlates of scene perception are sensitive to inversion.
Journal of Vision, 19, 190. https://doi.org/10.1167/19.10.190
Harel, A., Groen, I. I. A., Kravitz, D. J., Deouell, L. Y., &
Baker, C. I. (2016). The temporal dynamics of scene
processing: A multifaceted EEG investigation. eNeuro, 3,
ENEURO-0139-16.2016. https://doi.org/10.1523/ENEURO
.0139-16.2016, PubMed: 27699208
Harel, A., Mzozoyana, M. W., Al Zoubi, H., Nador, J. D., Noesen,
B. T., Lowe, M. X., et al. (2020). Artificially-generated scenes
demonstrate the importance of global scene properties for
scene perception. Neuropsychologia, 141, 107434. https://
doi.org/10.1016/j.neuropsychologia.2020.107434, PubMed:
32179102
Hassabis, D., & Maguire, E. A. (2009). The construction system
of the brain. Philosophical Transactions of the Royal Society
of London, Series B: Biological Sciences, 364, 1263–1271.
https://doi.org/10.1098/rstb.2008.0296, PubMed: 19528007
Hasson, U., Harel, M., Levy, I., & Malach, R. (2003). Large-scale
mirror-symmetry organization of human occipito-temporal
object areas. Neuron, 37, 1027–1041. https://doi.org/10.1016
/S0896-6273(03)00144-2, PubMed: 12670430
Henriksson, L., Mur, M., & Kriegeskorte, N. (2019). Rapid
invariant encoding of scene layout in human OPA. Neuron,
103, 161–171. https://doi.org/10.1016/j.neuron.2019.04.014,
PubMed: 31097360
Jelić, A., Tieri, G., Me Matteis, F., Babiloni, F., & Vecchiato, G.
(2016). The enactive approach to architectural experience: A
neurophysiological perspective on embodiment, motivation,
and affordances. Frontiers in Psychology, 7, 481. https://doi
.org/10.3389/fpsyg.2016.00481, PubMed: 27065937
Julian, J. B., Keinath, A. T., Marchette, S. A., & Epstein, R. A.
(2018). The neurocognitive basis of spatial reorientation.
Current Biology, 28, R1059–R1073. https://doi.org/10.1016/j
.cub.2018.04.057, PubMed: 30205055
Julian, J. B., Ryan, J., Hamilton, R. H., & Epstein, R. A. (2016).
The occipital place area is causally involved in representing
environmental boundaries during navigation. Current
Biology, 26, 1104–1109. https://doi.org/10.1016/j.cub.2016.02
.066, PubMed: 27020742
Jung, T.-P., Humphries, C., Lee, T.-W., Makeig, S., McKeown,
M. J., Iragui, V., et al. (1998). Extended ICA removes artifacts
from electroencephalographic recordings. Advances in
Neural Information Processing Systems, 10, 894–900.
Kaiser, D., Häberle, G., & Cichy, R. M. (2020). Cortical
sensitivity to natural scene structure. Human Brain
Mapping, 41, 1286–1295. https://doi.org/10.1002/hbm.24875,
PubMed: 31758632
Kaiser, D., Turini, J., & Cichy, R. M. (2019). A neural mechanism
for contextualizing fragmented inputs during naturalistic
vision. eLife, 8, e48182. https://doi.org/10.7554/eLife.48182,
PubMed: 31596234
Kamps, F. S., Julian, J. B., Kubilius, J., Kanwisher, N., & Dilks,
D. D. (2016). The occipital place area represents the local
elements of scenes. Neuroimage, 132, 417–424. https://doi
.org/10.1016/j.neuroimage.2016.02.062, PubMed: 26931815
Kamps, F. S., Lall, V., & Dilks, D. D. (2016). The occipital place
area represents first-person perspective motion information
through scenes. Cortex, 83, 17–26. https://doi.org/10.1016/j
.cortex.2016.06.022, PubMed: 27474914
Kaplan, R., & Friston, K. J. (2018). Planning and navigation as
active inference. Biological Cybernetics, 112, 323–343. https://
doi.org/10.1007/s00422-018-0753-2, PubMed: 29572721
Kravitz, D. J., Peng, C. S., & Baker, C. I. (2011). Real-world scene
representations in high-level visual cortex: It’s the spaces
more than the places. Journal of Neuroscience, 31,
7322–7333. https://doi.org/10.1523/JNEUROSCI.4588-10
.2011, PubMed: 21593316
Kravitz, D. J., Saleem, K. S., Baker, C. I., & Mishkin, M. (2011). A
new neural framework for visuospatial processing. Nature
Reviews Neuroscience, 12, 217–230. https://doi.org/10.1038
/nrn3008, PubMed: 21415848
Harel, A., Kravitz, D. J., & Baker, C. I. (2013). Deconstructing
Kurby, C. A., & Zacks, J. M. (2008). Segmentation in the
visual scenes in cortex: Gradients of object and spatial layout
information. Cerebral Cortex, 23, 947–957. https://doi.org/10
.1093/cercor/bhs091, PubMed: 22473894
perception and memory of events. Trends in Cognitive
Sciences, 12, 72–79. https://doi.org/10.1016/j.tics.2007.11.004,
PubMed: 18178125
Harel et al.
409
l
D
o
w
n
o
a
d
e
d
f
r
o
m
h
t
t
p
:
/
/
d
i
r
e
c
t
.
m
i
t
.
e
d
u
/
j
/
o
c
n
a
r
t
i
c
e
-
p
d
l
f
/
/
/
/
3
4
3
3
9
7
1
9
8
5
0
1
9
/
j
o
c
n
_
a
_
0
1
8
1
0
p
d
.
f
b
y
g
u
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
Lescroart, M. D., & Gallant, J. L. (2019). Human scene-selective
areas represent 3D configurations of surfaces. Neuron, 101,
178–192. https://doi.org/10.1016/j.neuron.2018.11.004,
PubMed: 30497771
Lescroart, M. D., Stansbury, D. E., & Gallant, J. L. (2015). Fourier
power, subjective distance, and object categories all provide
plausible models of BOLD responses in scene-selective
visual areas. Frontiers in Computational Neuroscience, 9,
135. https://doi.org/10.3389/fncom.2015.00135, PubMed:
26594164
Lowe, M. X., Rajsic, J., Ferber, S., & Walther, D. B. (2018).
Discriminating scene categories from brain activity within 100
milliseconds. Cortex, 106, 275–287. https://doi.org/10.1016/j
.cortex.2018.06.006, PubMed: 30037637
Luck, S. J. (2014). An introduction to the event-related
potential technique (2nd ed.). MIT Press.
Malcolm, G. L., Groen, I. I. A., & Baker, C. I. (2016). Making
sense of real-world scenes. Trends in Cognitive Sciences, 20,
843–856. https://doi.org/10.1016/j.tics.2016.09.003, PubMed:
27769727
Nakamura, K., Kawashima, R., Sato, N., Nakamura, A., Sugiura,
M., Kato, T., et al. (2000). Functional delineation of the
human occipito-temporal areas related to face and scene
processing: A PET study. Brain, 123, 1903–1912. https://doi
.org/10.1093/brain/123.9.1903, PubMed: 10960054
Nunez, P. L., & Srinivasan, R. (2006). Electric fields of the brain:
The neurophysics of EEG. Oxford University Press. https://doi
.org/10.1093/acprof:oso/9780195050387.001.0001
Persichetti, A. S., & Dilks, D. D. (2016). Perceived egocentric
distance sensitivity and invariance across scene-selective
cortex. Cortex, 77, 155–163. https://doi.org/10.1016/j.cortex
.2016.02.006, PubMed: 26963085
Persichetti, A. S., & Dilks, D. D. (2018). Dissociable neural
systems for recognizing places and navigating through them.
Journal of Neuroscience, 38, 10295–10304. https://doi.org/10
.1523/JNEUROSCI.1200-18.2018, PubMed: 30348675
Petrov, Y. (2012). Harmony: EEG/MEG linear inverse source
reconstruction in the anatomical basis of spherical
harmonics. PLoS One, 7, e44439. https://doi.org/10.1371
/journal.pone.0044439, PubMed: 23071497
Pinhas, M., Tzelgov, J., & Ganor-Stern, D. (2012). Estimating
linear effects in ANOVA designs: The easy way. Behavior
Research Methods, 44, 788–794. https://doi.org/10.3758
/s13428-011-0172-y, PubMed: 22101656
Rousselet, G., Joubert, O., & Fabre-Thorpe, M. (2005). How
long to get to the “gist” of real-world natural scenes?
Visual Cognition, 12, 852–877. https://doi.org/10.1080
/13506280444000553
Sadeh, B., Podlipsky, I., Zhdanov, A., & Yovel, G. (2010).
Event-related potential and functional MRI measures of
face-selectivity are highly correlated: A simultaneous
ERP-fMRI investigation. Human Brain Mapping, 31,
1490–1501. https://doi.org/10.1002/hbm.20952, PubMed:
20127870
Sato, N., Nakamura, K., Nakamura, A., Sugiura, M., Ito, K.,
Fukuda, H., et al. (1999). Different time course between
scene processing and face processing: A MEG study.
NeuroReport, 10, 3633–3637. https://doi.org/10.1097
/00001756-199911260-00031, PubMed: 10619657
Sestito, M., Flach, J., & Harel, A. (2018). Grasping the world
from a cockpit: Perspectives on embodied neural
mechanisms underlying human performance and
ergonomics in aviation context. Theoretical Issues in
Ergonomics Science, 19, 692–711. https://doi.org/10.1080
/1463922X.2018.1474504
Sestito, M., Harel, A., Nador, J., & Flach, J. (2018). Investigating
neural sensorimotor mechanisms underlying flight expertise
in pilots: Preliminary data from an EEG study. Frontiers in
Human Neuroscience, 12, 489. https://doi.org/10.3389
/fnhum.2018.00489, PubMed: 30618676
Walther, D. B., Caddigan, E., Fei-Fei, L., & Beck, D. M. (2009).
Natural scene categories revealed in distributed patterns of
activity in the human brain. Journal of Neuroscience, 29,
10573–10581. https://doi.org/10.1523/JNEUROSCI.0559-09
.2009, PubMed: 19710310
Walther, D. B., Chai, B., Caddigan, E., Beck, D. M., & Fei-Fei, L.
(2011). Simple line drawings suffice for functional MRI
decoding of natural scene categories. Proceedings of the
National Academy of Sciences, U.S.A., 108, 9661–9666.
https://doi.org/10.1073/pnas.1015666108, PubMed: 21593417
Wiesmann, C. G., Friederici, A. D., Singer, T., & Steinbeis, N.
(2020). Two systems for thinking about others’ thoughts in
the developing brain. Proceedings of the National Academy
of Sciences, U.S.A., 117, 6928–6935. https://doi.org/10.1073
/pnas.1916725117, PubMed: 32152111
Wilson, M. (2002). Six views of embodied cognition.
Psychonomic Bulletin & Review, 9, 625–636. https://doi.org
/10.3758/BF03196322, PubMed: 12613670
Zacks, J. M., Speer, N. K., Swallow, K. M., Braver, T. S., &
Reynolds, J. R. (2007). Event perception: A mind–brain
perspective. Psychological Bulletin, 133, 273–293. https://doi
.org/10.1037/0033-2909.133.2.273, PubMed: 17338600
Zhang, H., Houpt, J. W., & Harel, A. (2019). Establishing
reference scales for scene naturalness and openness.
Behavior Research Methods, 51, 1179–1186. https://doi.org
/10.3758/s13428-018-1053-4, PubMed: 29845553
410
Journal of Cognitive Neuroscience
Volume 34, Number 3
l
D
o
w
n
o
a
d
e
d
f
r
o
m
h
t
t
p
:
/
/
d
i
r
e
c
t
.
m
i
t
.
e
d
u
/
j
/
o
c
n
a
r
t
i
c
e
-
p
d
l
f
/
/
/
/
3
4
3
3
9
7
1
9
8
5
0
1
9
/
j
o
c
n
_
a
_
0
1
8
1
0
p
d
.
f
b
y
g
u
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3