REPORT

REPORT

Spatial Scene Memories Are Biased Towards a
Fixed Amount of Semantic Information

Michelle R. Greene1,2 and Devanshi Trivedi1,3

1Bates College, Program in Neuroscience
2Barnard College, Università della Columbia
3Oxford University

a n o p e n a c c e s s

j o u r n a l

Keywords: boundary extension, scene perception

ABSTRACT

Scene memory has known spatial biases. Boundary extension is a well-known bias whereby
observers remember visual information beyond an image’s boundaries. While recent studies
demonstrate that boundary contraction also reliably occurs based on intrinsic image
properties, the specific properties that drive the effect are unknown. This study assesses the
extent to which scene memory might have a fixed capacity for information. We assessed both
visual and semantic information in a scene database using techniques from image processing
and natural language processing, rispettivamente. We then assessed how both types of
information predicted memory errors for scene boundaries using a standard rapid serial visual
presentation (RSVP) forced error paradigm. A linear regression model indicated that memories
for scene boundaries were significantly predicted by semantic, but not visual, information and
that this effect persisted when scene depth was considered. Boundary extension was observed
for images with low semantic information, and contraction was observed for images with high
semantic information. This suggests a cognitive process that normalizes the amount of
semantic information held in memory.

Citation: Greene, M. R., & Trivedi, D.
(2023). Spatial Scene Memories Are
Biased Towards a Fixed Amount of
Semantic Information. Open Mind:
Discoveries in Cognitive Science,
7, 445–459. https://doi.org/10.1162
/opmi_a_00088

DOI:
https://doi.org/10.1162/opmi_a_00088

INTRODUCTION

Supplemental Materials:
https://doi.org/10.1162/opmi_a_00088

Received: 25 May 2023
Accepted: 3 Giugno 2023

Competing Interests: The authors
declare no conflict of interests.

Corresponding Author:
Michelle R. Greene
mgreene2@bates.edu

Copyright: © 2023
Istituto di Tecnologia del Massachussetts
Pubblicato sotto Creative Commons
Attribuzione 4.0 Internazionale
(CC BY 4.0) licenza

The MIT Press

It is said that pictures are worth one thousand words. If so, how many of these words can we
retain in memory after viewing a scene? Although scene memory is excellent (Brady et al.,
2008), these memories come with specific biases. One’s scene memories can contain more
space than the picture depicted (boundary extension, (per esempio., Intraub & Richardson, 1989)), O
they may be smaller (boundary contraction, e.g. (Bainbridge & Baker, 2020UN)). A given scene
will consistently expand or contract in observers’ memories, yet the scene properties that lead
to one type of scene transformation have not yet been fully characterized. This work
approaches the question from the perspective of information capacity limitations. Here, we
ask whether there is a bias towards storing constant amount of scene information in visual
memory. If this is the case, scenes that exceed the information capacity limit will contract,
while scenes under the information capacity limit will expand. We measured the relative
amounts of visual and semantic information from scenes in a large scene database and
assessed the boundary transformations for each scene. We found that a scene’s semantic infor-
mation predicted boundary transformation scores in the manner predicted by the fixed infor-
mation hypothesis.

l

D
o
w
N
o
UN
D
e
D

F
R
o
M
H

T
T

P

:
/
/

D
io
R
e
C
T
.

M

io
T
.

/

e
D
tu
o
P
M

io
/

l

UN
R
T
io
C
e

P
D

F
/

D
o

io
/

io

/

.

/

1
0
1
1
6
2
o
P
M
_
UN
_
0
0
0
8
8
2
1
4
8
1
8
0
o
P
M
_
UN
_
0
0
0
8
8
P
D

.

/

io

F

B

G
tu
e
S
T

T

o
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3

Spatial Scene Memories Are Biased Towards a Fixed Amount of Semantic Information Greene and Trivedi

Initial studies of spatial biases in scene memory reported only boundary extension (Intraub,
2012; Intraub & Richardson, 1989). Tuttavia, it was a remarkably consistent effect – it was
observed across retention intervals ranging from milliseconds to days, across different testing
modalities, and across observers ranging in age from childhood to late adulthood (for a review,
Vedere (Hubbard et al., 2010)). More recent work has demonstrated that in a larger and more
diverse scene database, boundary extension and contraction are equally likely to occur
(Bainbridge & Baker, 2020UN). Although ongoing discussion is still establishing why the older
image sets only produced boundary extension (Bainbridge & Baker, 2020B; Intraub, 2020),
several recent studies have replicated and extended the basic finding of bidirectional bound-
ary transformations (Hafri et al., 2022; Lin et al., 2022; Park et al., 2021).

In these studies, there is general agreement that scene depth plays a pivotal role. Bainbridge
and Baker (2020UN) noted that scene composition was a significant predictor of the effect:
images composed of a central object in a relatively homogeneous background consistently
led to boundary extension, while more expansive scenes led to boundary contraction. Dopo
creating a set of 3D rendered environments that spanned the object-to-scene continuum, (Park
et al., 2021) found that larger depth was associated with more boundary contraction but that
no absolute viewing distance was associated with the change between boundary extension
and contraction. Piuttosto, the transition point was scene-specific and biased towards common,
canonical views of the given environment. Congruently, (Lin et al., 2022) found that spatial
scene memories were biased towards the modal depth for a given scene category, using
images from a large scene database with measured depth values.

A second emerging hypothesis is that image properties that provide depth cues may be
causally driving the effect. Camera aperture settings that are correlated with scene scale are
highly predictive of boundary transformations (Gandolfo et al., 2022). In a similar vein, (Hafri
et al., 2022) used image processing techniques to manipulate the perceived depth of images.
They found that expansive images that were made to appear small led to less boundary con-
traction, consistent with the idea that image features that give rise to spatial scale drive bound-
ary transformations rather than knowledge of scene categories and their expected mean depth
values.

There are also reasons to believe that scene depth may not be the sole feature responsible
for boundary transformations. Scenes that produce the most boundary extension are rela-
tively sparse, such as a central object on a simple background (Bainbridge & Baker,
2020UN). This suggests that scenes with similar depth may produce different boundary trans-
formations depending on the richness of their content. Images with highly emotional content
can produce boundary contraction (Takarangi et al., 2016), suggesting that specific image
content can drive boundary transformations. The goal of this study is to assess whether
increases in either semantic or visual information play a role in scene boundary
transformations.

The present research measures scene information and tests the extent to which information
predicts scene transformation scores. As scenes contain both visual and semantic information,
we used image processing techniques to measure visual information and natural language pro-
cessing to measure semantic information. Using a standard RSVP boundary extension para-
digm, we show that a scene’s semantic information significantly predicts scene boundary
transformations over and above the effect of scene depth. Specifically, images with low seman-
tic information tend to expand their boundaries, while images with high semantic information
contract. Altogether, this process suggests a scene memory representation that is normalized
towards a fixed amount of semantic information.

OPEN MIND: Discoveries in Cognitive Science

446

l

D
o
w
N
o
UN
D
e
D

F
R
o
M
H

T
T

P

:
/
/

D
io
R
e
C
T
.

M

io
T
.

/

e
D
tu
o
P
M

io
/

l

UN
R
T
io
C
e

P
D

F
/

D
o

io
/

io

/

/

.

1
0
1
1
6
2
o
P
M
_
UN
_
0
0
0
8
8
2
1
4
8
1
8
0
o
P
M
_
UN
_
0
0
0
8
8
P
D

.

/

io

F

B

G
tu
e
S
T

T

o
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3

Spatial Scene Memories Are Biased Towards a Fixed Amount of Semantic Information Greene and Trivedi

MATERIALS AND METHODS

Stimuli

We used 1000 images from recent work on boundary extension (Bainbridge & Baker, 2020UN).
This dataset includes 500 scene-centric images from the SUN database (Xiao et al., 2014) E
500 object-centered images from the Google Open Images Dataset (Kuznetsova et al., 2020).
Each image was center cropped and resized to 350×350 pixels. We computed information
measures on visual and semantic features (see below) on all 1000 images and then selected
120 images that spanned the range of both information metrics. Of these, 68 were from the
Google Open Images Dataset, E 52 were from the SUN database.

Measuring Visual Information

Our goal was to compare the amount of visual information across scene images. Although
there have been efforts to estimate the entropy of natural scenes based on patches of natural
images (Chandler & Field, 2007; Hosseini et al., 2010; Petrov & Zhaoping, 2003), the number
of images and computational resources needed to estimate the entropy of natural scenes has
remained intractable fully. Therefore, we aim to estimate the relative amount of visual infor-
mation within this dataset rather than computing a specific number of bits for each image.

Because there is no single way to compute visual information, our approach was to com-
pute three measures that describe different aspects of scene information and complexity and
then employ dimensionality reduction (principal components analysis, (PCA)) to unify what is
common in these metrics into one measure of visual information. Although some recent work
has employed similar features and a whitening transformation to reduce multicollinearity
(Greene & Hansen, 2020), we decided to use PCA to unify the metrics because we had no
specific hypotheses about the individual metrics and instead wanted an aggregate measure of
visual information. Each of the three features is shown visually in Figure 1A.

The first measure was edge density from a Canny edge detector. Intuitively, an image with
more contours will have more complexity and, Perciò, more information. Edge density has
been shown to correlate with observers’ subjective assessments of image complexity (Ciocca,
Corchs, & Gasparini, 2015UN; Corchs et al., 2016; Nagle & Lavie, 2020), and eye fixations tend
to land on areas of higher edge density (Parkhurst & Niebur, 2003) suggesting that observers
need to foveate these regions to disambiguate their higher information content. Each image
was converted to grayscale. A 5-pixel by 5-pixel Gaussian blur was applied to the image to
reduce pixel noise. The lower and upper bounds for hysteresis thresholding were 30 E 150
pixels, rispettivamente. The number of pixels that corresponded to edges was stored for each
Immagine.

The second metric described how colorful a scene was by computing the entropy of the
image’s color histogram. A similar metric correlates with subjective ratings of colorfulness
(Hasler & Suesstrunk, 2003). The role of colorfulness in image complexity and information
has had mixed results. In a study of abstract textures, images with a greater variety of colors
were rated as more complex (Kocaoğlu & Olguntürk, 2018). Tuttavia, images of real-world
scenes received similar complexity ratings in both color and grayscale versions (Ciocca,
Corchs, Gasparini, et al., 2015B), suggesting that color plays little role in the assessment of sub-
jective visual complexity. Each image was converted from RGB to CIELAB color. Using the A*
and B* color channels, we created a two-dimensional histogram with 20 bins each. The value
stored in each histogram cell reflected the number of pixels in the given A*-B* color range. Noi
converted this histogram to a probability distribution and computed and stored its entropy.

OPEN MIND: Discoveries in Cognitive Science

447

l

D
o
w
N
o
UN
D
e
D

F
R
o
M
H

T
T

P

:
/
/

D
io
R
e
C
T
.

M

io
T
.

/

e
D
tu
o
P
M

io
/

l

UN
R
T
io
C
e

P
D

F
/

D
o

io
/

io

/

/

.

1
0
1
1
6
2
o
P
M
_
UN
_
0
0
0
8
8
2
1
4
8
1
8
0
o
P
M
_
UN
_
0
0
0
8
8
P
D

/

.

io

F

B

G
tu
e
S
T

T

o
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3

Spatial Scene Memories Are Biased Towards a Fixed Amount of Semantic Information Greene and Trivedi

l

D
o
w
N
o
UN
D
e
D

F
R
o
M
H

T
T

P

:
/
/

D
io
R
e
C
T
.

M

io
T
.

/

e
D
tu
o
P
M

io
/

l

UN
R
T
io
C
e

P
D

F
/

D
o

io
/

io

/

/

.

1
0
1
1
6
2
o
P
M
_
UN
_
0
0
0
8
8
2
1
4
8
1
8
0
o
P
M
_
UN
_
0
0
0
8
8
P
D

/

.

io

F

B

G
tu
e
S
T

T

o
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3

(UN) Examples of the three visual information features. For edge density, we summed the number of pixels containing edges. Color
Figura 1.
variability was assessed via the entropy of A*-B* histograms. The entropy of the Gist description provided a metric of the variability of oriented
edge content at various spatial scales and image locations. For each, the lowest and highest values are shown, along with a visualization of the
feature. (B) Examples of the three semantic information features. To illustrate description length, we provide several example descriptions for
the longest and shortest descriptions. Lexical entropy is visualized by word clouds. Lexical entropy increases when the number of unique
words increases. The pairwise similarity is visualized using a simplified 2D graph as a stand-in for the high-dimensional vector space repre-
senting each word. A small average cosine distance was obtained for related concepts. For each, the lowest and highest values are shown,
along with a visualization of the feature.

Entropy is a measurement of uncertainty (Shannon, 1948) and is computed with the following
formula:

H ¼ −1

Xn

i¼1

pi log 2pi

Where n is the number of bins in the discrete probability distribution, and pi is the proba-
bility associated with the ith bin. Entropy is maximized for uniform distributions and
approcci 0 as any pi approaches 1.

To choose the bin size, we selected a separate set of six images containing three images
with subjectively low color variability and three images with subjectively high color variabil-
ità. We computed color histograms with bin sizes between 2 E 200 and examined the
entropy differences between the two image groups. We found that a bin size of 20 provided
the largest and most reliable differences between the two groups.

OPEN MIND: Discoveries in Cognitive Science

448

Spatial Scene Memories Are Biased Towards a Fixed Amount of Semantic Information Greene and Trivedi

The third metric was entropy in the Gist descriptor of (Oliva & Torralba, 2001). This metric
describes the dominant orientations and spatial frequencies at different image locations. Questo
descriptor has been shown to be helpful in scene classification (Oliva & Torralba, 2001) E
correlates with neural responses to scenes ( Watson et al., 2017). Here, we used three spatial
scales with 8, 6, E 4 orientations per scale at each of the 64 image locations for a total of
1152 caratteristiche. We normalized the response vector to sum to 1 and computed entropy over this
representation.

Così, each image’s visual information was represented by three numbers: edge density,
color entropy, and gist entropy. We computed principal components analysis (PCA) on this
matrix and projected all images onto the first principal component to create a unified measure
of visual information. This component accounted for 97% of the variance in the original data.

Measuring Semantic Information

While the visual information metrics describe the quantity and variety of physical image fea-
tures, we would also like to describe the amount of meaning in an image as well. These cannot
be computed directly from the image, so we conducted a preliminary study to obtain
observers’ descriptions of images. From these, we used natural language processing to com-
pute several metrics of the complexity of these descriptions. We will detail each of these
below; see Figure 1B for visual descriptions.

For the description experiment, 1112 participants on Amazon’s Mechanical Turk (mTurk)
platform viewed trials in which they were shown one of the 1000 photographs and asked to
describe the image such that another person could pick it out of a lineup of similar images. Noi
restricted the experiment to US-based persons who had previously completed at least 1000
previous hits with at least 95% accuracy. Each individual could complete as many of the
1000 trials as desired. The median number of trials per participant was 10 (range: 1 A 1000).

As with visual information, we extracted three features that reflected semantic information
in these descriptions and used PCA to combine them into one unified measurement. The first
metric was each image’s median description length (in words). We reasoned that participants’
image descriptions followed the Gricean maxim of informativeness – that is, they only con-
tained content that would not be taken for granted by a reader (Grice, 1991). Così, all things
being equal, images that consistently elicited longer descriptions contained more semantic
information than those with shorter descriptions. High-frequency words (Piantadosi et al.,
2011) and words with highly predictive contexts (Mahowald et al., 2013) tend to be shorter,
reflecting communication strategies that maximize information transmission.

The second metric was the entropy of a bag-of-words description ( Wang et al., 2020). Noi
used the NLTK python library (Bird et al., 2009) to convert each description to lowercase and
to remove punctuation and the default stopwords from the nltk library (common non-content
words such as ‘the’, ‘and’, ‘I’). From these, we created a dictionary of all unique word tokens (UN
total of 19,556 words across the 100,000 unique descriptions). For each image, we created a
histogram of word token use across the 100 descriptions and computed entropy over this his-
togram. Under this metric, an image with more varied descriptions across observers or more
unique work tokens will receive a higher entropy measurement. This metric has been shown to
correlate with subjective complexity ratings of abstract art ( Wang et al., 2020).

The last metric we computed was the mean semantic distance between pairs of words as
evaluated by a Word2vec, a word embedding model (Mikolov et al., 2013) trained on the
Google News dataset. The logic behind this metric is that larger distances between words will

OPEN MIND: Discoveries in Cognitive Science

449

l

D
o
w
N
o
UN
D
e
D

F
R
o
M
H

T
T

P

:
/
/

D
io
R
e
C
T
.

M

io
T
.

/

e
D
tu
o
P
M

io
/

l

UN
R
T
io
C
e

P
D

F
/

D
o

io
/

io

/

.

/

1
0
1
1
6
2
o
P
M
_
UN
_
0
0
0
8
8
2
1
4
8
1
8
0
o
P
M
_
UN
_
0
0
0
8
8
P
D

/

.

io

F

B

G
tu
e
S
T

T

o
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3

Spatial Scene Memories Are Biased Towards a Fixed Amount of Semantic Information Greene and Trivedi

indicate less standard context and, così, more semantic information. Encoding models based
on word2vec have shown to be helpful in predicting neural responses in object-selective cor-
tex (Bonner & Epstein, 2021). For each description, we removed stopwords using NLTK (Bird
et al., 2009). With the remaining words, we computed the cosine distance between each pair
of remaining word tokens. Per esempio, for the description “The quick brown fox jumps over
the lazy dog”, we would have the following tokens after omitting stopwords: “Quick brown
fox jumps over lazy dog”. These seven words contain 21 unique word pairs, and cosine dis-
tance was computed between each pair (Per esempio, quick-brown, quick-fox, quick-jumps,
eccetera.). We computed the mean similarity across all distance pairs and across the 100 descrip-
tions for each image. High similarity scores are obtained when two words appear close
together in language contexts.

Così, each image’s semantic information was represented by three metrics: median
description length, average lexical entropy, and average pairwise similarity. We performed
PCA on this matrix and projected each image onto the first principal component for a single
metric of semantic information. This first component explained about 86% of the original var-
iability across images.

Comparing Image Metrics

In order to compare our semantic and visual information metrics to existing metrics on the
same images, we computed correlation coefficients between each of the four metrics measured
by (Bainbridge & Baker, 2020UN) and our metrics. Come mostrato in figura 2, Pearson correlation
coefficients were low to moderate across the board. We observed higher absolute correla-
tions between Bainbridge and Baker’s metrics and our semantic information metrics,

l

D
o
w
N
o
UN
D
e
D

F
R
o
M
H

T
T

P

:
/
/

D
io
R
e
C
T
.

M

io
T
.

/

e
D
tu
o
P
M

io
/

l

UN
R
T
io
C
e

P
D

F
/

D
o

io
/

io

/

/

.

1
0
1
1
6
2
o
P
M
_
UN
_
0
0
0
8
8
2
1
4
8
1
8
0
o
P
M
_
UN
_
0
0
0
8
8
P
D

.

/

io

F

B

G
tu
e
S
T

T

o
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3

Figura 2. Pearson correlation coefficients between the image metrics from Bainbridge and Baker
(2020UN), shown in rows, and our semantic and visual information metrics shown in columns.

OPEN MIND: Discoveries in Cognitive Science

450

Spatial Scene Memories Are Biased Towards a Fixed Amount of Semantic Information Greene and Trivedi

compared to our visual information metrics. Further, the highest absolute correlations with the
Bainbridge and Baker metrics were found between lexical entropy, which was positively cor-
related with the number of objects, object eccentricity, and subjective scene distance, E
negatively correlated with object size.

Experimental Procedure

We assessed boundary transformations using a standard rapid serial visual presentation (RSVP)
paradigm (Bainbridge & Baker, 2020UN; Hafri et al., 2022; Lin et al., 2022; Park et al., 2021).
We recruited 360 US-based participants on the online experiment platform Prolific (mean age =
38 years, 288 female, 61 male, 11 nonbinary). The experiment was approved by the Bates
College IRB, and volunteers provided written informed consent before participating. Our sam-
ple size was determined via a power analysis (conducted in the pwr library of R) using the
effect size reported in (Bainbridge & Baker, 2020UN). Our sample size is in line with (Hafri
et al., 2022).

Each participant viewed each of the 120 images in random order. The trial structure for this
experiment is shown in Figure 3. Each of the 120 trials consisted of the target image shown for
250 ms, followed by a rapid serial visual presentation (RSVP) stream of five block-scrambled
masks shown for 50 ms each. None of the mask images were from the experimental image set.
Following the mask series, participants were presented with the same target image for 250 ms
and asked to assess whether the second version was more contracted (zoomed in) or expanded
(zoomed out) compared to the previous image. As the two images were identical, participants
were making a forced error. For this reason, no performance feedback was provided.

If there were no systematic memory biases for scene boundaries, we would expect a
roughly equal number of people to make each error type. If boundary extension is the primary
mode of boundary transformation, we would expect participants to indicate that the second
presentation of each image was more contracted (zoomed in). Tuttavia, if some images are
prone to boundary extension and others to boundary contraction, we should see systematic
biases in both directions but differing by image, as found by (Bainbridge & Baker, 2020UN). Questo
experiment tests the hypothesis that memory for scene boundaries might have a fixed informa-
tion capacity. Così, we predicted that images with more information would lead to contrac-
zione, while images with less information would lead to extension.

RESULTS

We correlated the visual and semantic information metrics to assess their relative indepen-
dence. Across the 120 images in this experiment, we found no significant correlation between
them (r = 0.03, T(118) = 0.31, p = 0.76). Therefore, these two metrics tap into relatively inde-
pendent sources of image information.

For each of the 120 images, we computed a boundary transformation score across the 360
observers. This score was the average of the number of observers who made an extension error
(coded as +1) to the number of observers who made a contraction error (coded as −1). Così,
scores larger than 0 indicate boundary extension, while negative scores indicate boundary
contraction. If no systematic tendency is observed, this score would be about 0.

Boundary transformation scores ranged from −0.22 to 0.55. 97 images (81%) showed
boundary extension, while 22 images showed boundary contraction, and one image had a
transformation ratio of 0.0. Our results were highly similar to the results obtained by (Bainbridge
& Baker, 2020UN): our scores were highly correlated with their boundary transformation scores

OPEN MIND: Discoveries in Cognitive Science

451

l

D
o
w
N
o
UN
D
e
D

F
R
o
M
H

T
T

P

:
/
/

D
io
R
e
C
T
.

M

io
T
.

/

e
D
tu
o
P
M

io
/

l

UN
R
T
io
C
e

P
D

F
/

D
o

io
/

io

/

/

.

1
0
1
1
6
2
o
P
M
_
UN
_
0
0
0
8
8
2
1
4
8
1
8
0
o
P
M
_
UN
_
0
0
0
8
8
P
D

.

/

io

F

B

G
tu
e
S
T

T

o
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3

Spatial Scene Memories Are Biased Towards a Fixed Amount of Semantic Information Greene and Trivedi

Figura 3. Trial structure for the experiment. We employed a forced error paradigm in which observers were asked to indicate whether the
second presentation of an image was closer or farther than the originally-presented image after a dynamic mask.

(r = 0.86, T(118) = 17.93, P < 2.2e–16). Thus, this work also provides an independent replica- tion of their work. To compute the noise ceiling of the data, we estimated how well half of the participants’ boundary transformation scores predicted the scores of the other participants. In 10,000 iter- ations, we randomly split participants into two groups and computed the boundary transfor- mation scores for each group, and then predicted the first scores with the second via linear regression. We found that 95% of the splits had R2 values between 0.69 and 0.79. This sets an upper bound to how well any model can predict boundary transformation scores. Finally, we computed a baseline model with only depth (rankings provided by (Bainbridge & Baker, 2020a) predicting boundary transformation scores. A significant regression equation was found (F(1,116) = 210.3, p < 2.2e–16) with an R2 of 0.64. In other words, increasing the mean depth of an environment decreases the probability that the scene’s boundaries will expand in memory, as has been found in other recent work (Bainbridge & Baker, 2020a; Hafri et al., 2022; Park et al., 2021). Our main hypothesis is that the information contained in visual memory is relatively con- stant. This would predict that images with too much information will contract in memory while images with too little information will expand. We tested this through multiple linear regres- sion analysis, including the visual or semantic information scores as additional predictors for the boundary transformation score of each image. l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . / e d u o p m i / l a r t i c e - p d f / d o i / i / / . 1 0 1 1 6 2 o p m _ a _ 0 0 0 8 8 2 1 4 8 1 8 0 o p m _ a _ 0 0 0 8 8 p d / . i When semantic information was added to the model, a significant regression equation was found (F(2,115) = 123.4, p < 2.2e–16) with an adjusted R2 of 0.68 (99% of the lower bound of the noise ceiling). However, when visual information was added to the depth model, we did not observe an increase in R2 over the baseline depth model (F(2,115) = 105, p < 2.2e–16, adjusted R2 = 0.64). Similarly, a model with all three predictors had a sightly lower adjusted R2 than the model containing only semantic information and depth (F(3,114) = 82, p < 2.2e–16, adjusted R2 = 0.68, see Table 1). Therefore, visual information does not contribute significantly to memory distortions of scene boundaries. f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 Intercept Depth Semantic Information Visual Information Table 1. Regression coefficients and statistics. Coefficient 0.52 –0.13 –0.03 –0.001 t-value 20.6 –14.8 –3.70 0.732 OPEN MIND: Discoveries in Cognitive Science p-value 2e–16 2e–16 0.0003 0.47 452 Spatial Scene Memories Are Biased Towards a Fixed Amount of Semantic Information Greene and Trivedi Euler diagram of variance partition analysis. The size of the ellipses are proportional to Figure 4. the explained variance of each model, and the degree of shared variance between model is illus- trated with the overlap of ellipses. In order to understand which semantic features drove this result, we ran separate regression analyses for each of the three semantic features. We found that median description length was a significant predictor of boundary transformation scores (β = −0.06, F(1, 118) = 18.99, p = 2.83e–05, R2 = 0.13, as was the lexical entropy measurement (β = −0.23, F(1,118) = 30.6, p = 1.9e–7, R2 = 0.20). However, we did not find mean pairwise semantic similarity to be a sig- nificant predictor of boundary transformation scores (F(1,118) < 1). Therefore, images that elic- ited longer descriptions and also descriptions that were more variable across observers were associated with boundary contractions, while images that elicited shorter and more homoge- neous descriptions were associated with boundary extension. The previous results indicated that semantic information provides additional predictive power over and above that provided by a scene’s depth. However, we do not net know if the variability accounted for with semantic information is independent of that accounted for by depth. To address this question, we used depth, semantic, and visual information as pre- dictors to perform a variance partitioning analysis (Greene et al., 2016; Groen et al., 2018; Lescroart et al., 2015). In total, seven regression models were constructed: (1) the three features together; (2–4) each pair of features; (5–7) each individual feature. From these, simple arith- metic calculations on the R2 values between the models can allow us to infer each model’s shared and independently explained variance. We visualized the result in an Euler diagram (eulerr package, R) in Figure 4. Congruent with our previous analyses, we can see that seman- tic (but not visual) information provides some additional explained variance for boundary transformation scores. Therefore, in addition to confirming previous findings that larger depth scenes are associ- ated with boundary contraction instead of extension, our results show that higher semantic content is associated with boundary contraction, even when image depth is held constant. DISCUSSION It is said that pictures are worth one thousand words, but not all pictures evoke similarly rich descriptions. In other words, scenes differ in their semantic visual information. Inspired by recent studies showing that observers’ memories of the spatial boundaries of real-world scenes can either expand or contract (Bainbridge & Baker, 2020a; Hafri et al., 2022; Lin et al., 2022; OPEN MIND: Discoveries in Cognitive Science 453 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . / e d u o p m i / l a r t i c e - p d f / d o i / i . / / 1 0 1 1 6 2 o p m _ a _ 0 0 0 8 8 2 1 4 8 1 8 0 o p m _ a _ 0 0 0 8 8 p d / . i f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 Spatial Scene Memories Are Biased Towards a Fixed Amount of Semantic Information Greene and Trivedi Park et al., 2021), this paper tested the hypothesis that scene memories are biased towards a fixed amount of information. Using image processing and natural language processing tech- niques to create proxies for visual and semantic information, we found that semantic (but not visual) information predicted whether a scene’s boundaries would contract or expand in mem- ory and that this relationship held even when scene depth was held constant. Specifically, scenes with less semantic information tended to expand. In contrast, scenes with more seman- tic information tended to contract, consistent with the hypothesis that the amount of semantic information in visual memory is fixed. Many recent studies have pointed to the role of scene depth in boundary extension (Bainbridge & Baker, 2020a; Bertamini et al., 2005; Gandolfo et al., 2022; Hafri et al., 2022; Lin et al., 2022; Park et al., 2021). Specifically, scenes with small fields of view tend to produce boundary extension, while panoramic spaces can produce boundary contraction. Using the scene depth ratings provided by (Bainbridge & Baker, 2020a), we have both replicated and extended their result. We have found that semantic information, in addition to scene depth, predicts the amount of remembered visual space. Scenes with small mean depth also likely limit semantic information, as they tend to depict a single object against a background. How- ever, a scene with a large depth of field does not necessarily have a large amount of semantic information. For example, an empty warehouse may have very little semantic information. Thus, scene depth may be necessary, but not sufficient for boundary contraction, and seman- tic information may explain why not all large-scale scenes produce boundary contraction or why only boundary extension was previously reported in the literature (Intraub & Richardson, 1989). There are several theoretical accounts of the boundary extension phenomenon. Early the- ories linked boundary extension to predictive cognition and mental simulation. According to this account, representing space beyond the scene’s boundaries aids in planning eye movements and other actions within the scene (Hubbard, 1996; Intraub, 2002). More recent theories have posited that boundary extension helps build an immersive egocentric spatial framework that integrates visual input from fixation to fixation (Intraub, 2012). These ideas, while attractive, are not congruent with boundary contraction. Other theories posit that those scene boundary transformations result from systematic memory biases toward the statistics of one’s visual experience (Bainbridge & Baker, 2020a; Bartlett, 1932; Hafri et al., 2022; Lin et al., 2022; Park et al., 2021). In other words, atypically, narrow-field views will extend in memory while larger-than-normal views will contract. Although there is some disagreement about whether these viewpoint statistics are category-specific (Bainbridge & Baker, 2020a; Lin et al., 2022; Park et al., 2021) or generalize to scenes as a whole (Hafri et al., 2022), this theoretical account is consistent with both boundary extension and contraction. It is worth noting that this view does not necessarily conflict with our information theoretic account. Photographs are acts of visual communication (Sontag, 2001). As such, the known viewpoint and compo- sitional biases in photographs (Parkhurst & Niebur, 2003; Tatler et al., 2005) that lead to some viewpoints being more typical may reflect the desire of the photographer to optimize the amount of information that a viewer can apprehend at a glance. We observed that memory for semantically sparse scenes was expanded while memory for semantically complex scenes contracted. This implies a cognitive process that normalizes the amount of semantic information in memory. Memories for scenes that exceed a semantic information capacity limit may be “zoomed in”, focusing primarily on central content as most of the informative content of a photograph is located there (Tatler, 2007). By contrast, semantically sparse scenes may be inflated via automatic inferences based on the scene gist (Friedman, 1979; Greene, 2016), or other prominent objects (Draschkow & Võ, 2017). These OPEN MIND: Discoveries in Cognitive Science 454 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . / e d u o p m i / l a r t i c e - p d f / d o i / i / / . 1 0 1 1 6 2 o p m _ a _ 0 0 0 8 8 2 1 4 8 1 8 0 o p m _ a _ 0 0 0 8 8 p d / . i f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 Spatial Scene Memories Are Biased Towards a Fixed Amount of Semantic Information Greene and Trivedi inferential processes may serve to guide subsequent saccades or other behavior (Intraub & Richardson, 1989), or may reflect mechanisms to improve the accuracy of noisy representa- tions by leveraging prior knowledge (Hemmer & Steyvers, 2009; Kersten et al., 2004). This may be similar to other inferential processes that have been noted in boundary transforma- tions. For example, scene boundary memories are biased toward the most canonical view- points of a given location (Lin et al., 2022). Our results suggest that visual memory for scene boundaries is normalized towards a fixed capacity for semantic information. This is congruent with information-based capacity limita- tions for other forms of memory (Brady et al., 2009; Miller, 1956). For example, observers’ working memory capacities vary depending on the complexity of the objects held in memory (Alvarez & Cavanagh, 2004), suggesting that visual working memory has a fixed information capacity rather than a fixed capacity for objects or features. There is also evidence that visual experience can create compressed representations to increase working memory capacity. This has been shown through both training studies (Brady et al., 2009) and is reflected in the fact that working memory for real-world objects is higher than for abstract laboratory stimuli (Brady et al., 2016; Curby et al., 2009). This can be understood in a framework of semantic information. Familiar objects have consistent category labels (Murphy, 2004). This is a form of cognitive compression, allowing interlocutors to reason about a wide range of features that are associ- ated with the category label. As there is a bidirectional interplay between categories and the features that give rise to them (Schyns, 1998), it is plausible that rare content leads observers to “zoom in” to gain the same amount of information that could have been offloaded to a cate- gory label, if one exists. While accounts of capacity limitations speak to what can be lost in a memory represen- tation, they cannot speak to what can be gained. Our memories are augmented through prior knowledge in the form of schemas (Bartlett, 1932). The question of what specific sche- matic knowledge may be filled in is open, but previous literature suggests several possibil- ities. First, we know that memory representations for specific scene details decay more quickly than memory for the gist (Zeng et al., 2021). Activating scene gist activates a set of probable objects (Friedman, 1979) shared across viewers (Greene, 2016). Further, we know that observers will often falsely remember seeing objects that are consistent with a particular scene gist (Brewer & Treyans, 1981; Castelhano & Henderson, 2008). Therefore, it may be the case that specific objects are expanded in cases with low semantic informa- tion. Another possibility is that the perception of some objects activates memory represen- tations for other co-occurring objects (Koehler & Eckstein, 2017) and that these relations are expanded in memory. It seems likely that large anchor objects may drive the expansion of smaller, content objects (Draschkow & Võ, 2017). We believe that these possibilities are not mutually exclusive. Bayesian models of memory reconstruction have shown that prior knowledge in multiple hierarchical forms interacts (Hemmer & Steyvers, 2009). Thus, scene memory representations likely involve prior knowledge about objects, object relationships, viewpoints, and events. Another open question is why the mind would be biased towards a fixed amount of seman- tic information rather than biased towards compressing memory whenever possible. The func- tional role of boundary extension has not yet been established (Bainbridge & Baker, 2020a, 2020b; Intraub, 2020). The classic view is that boundary extension enables the observer to make predictions about what is beyond the immediate view and to facilitate subsequent sac- cades (Intraub & Richardson, 1989). The idea that boundary transformations may serve pre- dictive cognition is not incongruent with our view. Upon viewing a scene, the gist is activated OPEN MIND: Discoveries in Cognitive Science 455 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . / e d u o p m i / l a r t i c e - p d f / d o i / i / . / 1 0 1 1 6 2 o p m _ a _ 0 0 0 8 8 2 1 4 8 1 8 0 o p m _ a _ 0 0 0 8 8 p d . / i f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 Spatial Scene Memories Are Biased Towards a Fixed Amount of Semantic Information Greene and Trivedi (Oliva, 2005). This gist itself then activates the set of probable objects that could be found in the scene (Friedman, 1979; Greene, 2016). In this way, the gist increases the amount of seman- tic information, regardless of whether the predicted objects are fixated, or even if they are present (Brewer & Treyans, 1981). The semantic information capacity limitation may also be related to previously described capacity limitations. Boundaries of images with highly emotional content (which are likely also images with a large amount of semantic information) have been shown to contract in memory (Takarangi et al., 2016) (although see (Beighley et al., 2019)). These types of images can also lead to an “attentional rubbernecking” effect that mimics the attentional blink (Most et al., 2005). It may be the case that images with highly emotional content have higher levels of semantic information as we define it. Let us consider two hypothetical images: a man hold- ing a cell phone and a man holding a gun. We posit that the descriptions of the latter picture may center around the gun, the man’s motivations for holding it, and so on. By contrast, the act of holding a cell phone is so common as not even to warrant a mention. Therefore, the first picture would generate descriptions with fewer words, and thus contain less semantic infor- mation than the second. Finally, it may also be the case that images with a large amount of semantic information also impair early visual perception: images that depicted low-probability real-world events led to lower detection and categorization performance compared to perceptually-matched images of more probable content (Greene et al., 2015). In this study, observers generated descriptions of probable and improbable images that were presented for varying durations ranging from 24 ms to unlimited. It is worth noting that the descriptions of improbable images in the unlimited viewing duration were longer than those of the matched probable images and thus would register as higher semantic information using the metrics in the present study. We did not necessarily expect that memory for scene boundaries did not show a fixed infor- mation limit for visual information. Previous work has shown that images with high levels of visual information (indexed by similar features to the ones used here) led to less efficient visual search and lower performance on target detection in low contrast (Rosenholtz et al., 2007). It may be the case that while ongoing visual processes such as visual search are limited by visual information, scene memory representations are more conceptual and thus limited by semantic information. It could also be the case that we saw no effect of visual information because we observed a quadratic relationship between our visual metrics and scene depth. Scenes with a mid-level depth had the most visual information, while very shallow and deep scenes had lower visual information. This is congruent with previous results that show systematic differ- ences in specific image statistics, such as the slope of the amplitude spectrum, with scene depth (Torralba & Oliva, 2003). Our metrics of visual information are also related to these lower-level image statistics. Thus, visual information may play a nonlinear role in the memory for scene boundaries. This possibility is congruent with the findings of (Sun & Firestone, 2022) who found that visual stimuli of medium visual complexity led to longer verbal descriptions than images with lower or higher complexity levels. A final possibility is that the three metrics of visual information that we used, although motivated by previous literature, may not fully tap into biologically and psychologically relevant features. Future work may expand the number of visual features to include other metrics that have been shown to influence behavioral and neural responses, such as Weibull statistics of oriented filters (Groen et al., 2012). While our results demonstrate that memory errors for scene boundaries can be predicted by each image’s semantic information, much more predictive power is provided by scene depth. However, combining both depth and semantic information allows us to explain about 90% of OPEN MIND: Discoveries in Cognitive Science 456 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . / e d u o p m i / l a r t i c e - p d f / d o i / i . / / 1 0 1 1 6 2 o p m _ a _ 0 0 0 8 8 2 1 4 8 1 8 0 o p m _ a _ 0 0 0 8 8 p d / . i f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 Spatial Scene Memories Are Biased Towards a Fixed Amount of Semantic Information Greene and Trivedi explainable variability in scene boundary transformation scores. Thus, semantic information provides a critical element in the memory representation of scenes. ACKNOWLEDGMENTS We would like to thank Alon Hafri and one anonymous reviewer for helpful feedback. Thanks to Wilma Bainbridge and Lindsay Houck for useful discussions of this work. AUTHOR CONTRIBUTIONS Michelle R. Greene: Conceptualization, Methodology, Project Administration, Supervision, Visualization, Writing. Devanshi Trivedi: Formal Analysis, Methodology, Software, Visualization, Writing. FUNDING INFORMATION This project was supported by the Bates College Program in Neuroscience. DATA AVAILABILITY STATEMENT Data and stimuli are available at https://osf.io/wn7v2/. REFERENCES Alvarez, G. A., & Cavanagh, P. (2004). The capacity of visual short-term memory is set both by visual information load and by number of objects. Psychological Science, 15(2), 106–111. https://doi.org/10.1111/j.0963-7214.2004.01502006.x, PubMed: 14738517 Bainbridge, W. A., & Baker, C. I. (2020a). Boundaries extend and contract in scene memory depending on image properties. Cur- rent Biology, 30(3), 537–543. https://doi.org/10.1016/j.cub.2019 .12.004, PubMed: 31983637 Bainbridge, W. A., & Baker, C. I. (2020b). Reply to Intraub. Current Biology, 30(24), R1465–R1466. https://doi.org/10.1016/j.cub .2020.10.032, PubMed: 33352123 Bartlett, S. F. C. (1932). Remembering: A Study in Experimental and Social Psychology. Cambridge University Press. (2019). Remembering: Does Beighley, S., Sacco, G. R., Bauer, L., Hayes, A. M., & Intraub, H. the emotional content of a photograph affect boundary extension? Emotion, 19(4), 699–714. https://doi.org/10.1037/emo0000477, PubMed: 30138008 Bertamini, M., Jones, L. A., Spooner, A., & Hecht, H. (2005). Boundary extension: The role of magnification, object size, con- text, and binocular information. Journal of Experimental Psychol- ogy: Human Perception and Performance, 31(6), 1288–1307. https://doi.org/10.1037/0096-1523.31.6.1288, PubMed: 16366790 Bird, S., Klein, E., & Loper, E. (2009). Natural Language Processing with Python—Analyzing Text with the Natural Language Toolkit. O’Reilly Media, Inc. (2021). Object representations Bonner, M. F., & Epstein, R. A. in the human brain reflect the co-occurrence statistics of vision and language. Nature Communications, 12(1), Article 4081. https://doi.org/10.1038/s41467-021-24368-2, PubMed: 34215754 Brady, T. F., Konkle, T., & Alvarez, G. A. (2009). Compression in visual working memory: Using statistical regularities to form more efficient memory representations. Journal of Experimental Psychology: General, 138(4), 487–502. https://doi.org/10.1037 /a0016797, PubMed: 19883132 Brady, T. F., Konkle, T., Alvarez, G. A., & Oliva, A. (2008). Visual long-term memory has a massive storage capacity for object details. Proceedings of the National Academy of Sciences of the United States of America, 105(38), 14325–14329. https:// doi.org/10.1073/pnas.0803390105, PubMed: 18787113 Brady, T. F., Störmer, V. S., & Alvarez, G. A. (2016). Working mem- ory is not fixed-capacity: More active storage capacity for real-world objects than for simple stimuli. Proceedings of the National Academy of Sciences of the United States of America, 113(27), 7459–7464. https://doi.org/10.1073/pnas.1520027113, PubMed: 27325767 Brewer, W. F., & Treyans, J. C. (1981). Role of schemata in memory for places. Cognitive Psychology, 13(2), 207–230. https://doi.org /10.1016/0010-0285(81)90008-6 Castelhano, M. S., & Henderson, J. M. (2008). The influence of color on the perception of scene gist. Journal of Experi- mental Psychology: Human Perception and Performance, 34(3), 660–675. https://doi.org/10.1037/0096-1523.34.3.660, PubMed: 18505330 Chandler, D. M., & Field, D. J. (2007). Estimates of the information content and dimensionality of natural scenes from proximity distributions. Journal of the Optical Society of America, 24(4), 922–941. https://doi.org/10.1364/josaa.24.000922, PubMed: 17361279 Ciocca, G., Corchs, S., & Gasparini, F. (2015a). Complexity Per- ception of Texture Images. In V. Murino, E. Puppo, D. Sona, M. Cristani, & C. Sansone (Eds.), New Trends in Image Analysis and Processing – ICIAP 2015 Workshops (pp. 119–126). Springer, Cham. https://doi.org/10.1007/978-3-319-23222-5_15 Ciocca, G., Corchs, S., Gasparini, F., Bricolo, E., & Tebano, R. (2015b). Does Color Influence Image Complexity Perception? In A. Trémeau, R. Schettini, & S. Tominaga (Eds.), Computational Color Imaging (pp. 139–148). Springer, Cham. https://doi.org/10 .1007/978-3-319-15979-9_14 OPEN MIND: Discoveries in Cognitive Science 457 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . / e d u o p m i / l a r t i c e - p d f / d o i / i / / . 1 0 1 1 6 2 o p m _ a _ 0 0 0 8 8 2 1 4 8 1 8 0 o p m _ a _ 0 0 0 8 8 p d . / i f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 Spatial Scene Memories Are Biased Towards a Fixed Amount of Semantic Information Greene and Trivedi Corchs, S. E., Ciocca, G., Bricolo, E., & Gasparini, F. (2016). Pre- dicting complexity perception of real world images. PLoS One, 11(6), e0157986. https://doi.org/10.1371/journal.pone .0157986, PubMed: 27336469 Curby, K. M., Glazek, K., & Gauthier, I. (2009). A visual short-term memory advantage for objects of expertise. Journal of Experimental Psychology. Human Perception and Performance, 35(1) 94–107. https://doi.org/10.1037/0096-1523.35.1.94, PubMed: 19170473 Draschkow, D., & Võ, M. L.-H. (2017). Scene grammar shapes the way we interact with objects, strengthens memories, and speeds search. Scientific Reports, 7(1), 16471. https://doi.org/10.1038 /s41598-017-16739-x, PubMed: 29184115 Friedman, A. (1979). Framing pictures: The role of knowledge in automatized encoding and memory for gist. Journal of Experi- mental Psychology: General, 108(3), 316–355. https://doi.org /10.1037/0096-3445.108.3.316, PubMed: 528908 Gandolfo, M., Nägele, H., & Peelen, M. V. (2022). Predictive pro- cessing of scene layout depends on naturalistic depth of field. bioRxiv. https://doi.org/10.1101/2021.11.08.467670 Greene, M. R. (2016). Estimations of object frequency are fre- quently overestimated. Cognition, 149, 6–10. https://doi.org/10 .1016/j.cognition.2015.12.011, PubMed: 26774103 Greene, M. R., Baldassano, C., Esteva, A., Beck, D. M., & Fei-Fei, L. (2016). Visual scenes are categorized by function. Journal of Experimental Psychology: General, 145(1), 82–94. https://doi.org /10.1037/xge0000129, PubMed: 26709590 Greene, M. R., Botros, A. P., Beck, D. M., & Fei-Fei, L. (2015). What you see is what you expect: Rapid scene understanding benefits from prior experience. Attention, Perception & Psychophysics, 77(4), 1239–1251. https://doi.org/10.3758/s13414-015-0859-8, PubMed: 25776799 Greene, M. R., & Hansen, B. C. (2020). Disentangling the inde- pendent contributions of visual and conceptual features to the spatiotemporal dynamics of scene categorization. Journal of Neuroscience, 40(27), 5283–5299. https://doi.org/10.1523 /JNEUROSCI.2088-19.2020, PubMed: 32467356 Grice, P. (1991). Studies in the Way of Words. Harvard University Press. Groen, I. I., Greene, M. R., Baldassano, C., Fei-Fei, L., Beck, D. M., & Baker, C. I. (2018). Distinct contributions of functional and deep neural network features to representational similarity of scenes in human brain and behavior. ELife, 7, e32962. https://doi.org/10 .7554/eLife.32962, PubMed: 29513219 Groen, I. I. A., Ghebreab, S., Lamme, V. A. F., & Scholte, H. S. (2012). Spatially pooled contrast responses predict neural and perceptual similarity of naturalistic image categories. PLoS Com- putational Biology, 8(10), e1002726. https://doi.org/10.1371 /journal.pcbi.1002726, PubMed: 23093921 Hafri, A., Wadhwa, S., & Bonner, M. F. (2022). Perceived distance alters memory for scene boundaries. Psychological Science, 3 3 ( 1 2 ) , 2 0 4 0 – 2 0 5 8 . h t t p s : / / d o i . o r g / 1 0 . 1 1 7 7 /09567976221093575, PubMed: 36206190 Hasler, D., & Suesstrunk, S. E. (2003). Measuring colorfulness in natural images. Human Vision and Electronic Imaging, 5007, 87–95. https://doi.org/10.1117/12.477378 Hemmer, P., & Steyvers, M. (2009). A Bayesian account of recon- structive memory. Topics in Cognitive Science, 1(1), 189–202. https://doi.org/10.1111/j.1756-8765.2008.01010.x, PubMed: 25164805 Hosseini, R., Sinz, F., & Bethge, M. (2010). Lower bounds on the redundancy of natural images. Vision Research, 50(22), 2213–2222. https://doi.org/10.1016/j.visres.2010.07.025, PubMed: 20705084 Hubbard, T. L. (1996). Displacement in depth: Representational momen- tum and boundary extension. Psychological Research, 59(1), 33–47. https://doi.org/10.1007/BF00419832, PubMed: 8693049 Hubbard, T. L., Hutchison, J. L., & Courtney, J. R. (2010). Boundary extension: Findings and theories. Quarterly Journal of Experi- mental Psychology, 63(8), 1467–1494. https://doi.org/10.1080 /17470210903511236, PubMed: 20432178 Intraub, H. (2002). Anticipatory spatial representation of natural scenes: Momentum without movement? Visual Cognition, 9(1–2), 93–119. https://doi.org/10.1080/13506280143000340 Intraub, H. (2012). Rethinking visual scene perception. Wiley Inter- disciplinary Reviews: Cognitive Science, 3(1), 117–127. https:// doi.org/10.1002/wcs.149, PubMed: 26302476 Intraub, H. (2020). Searching for boundary extension. Current Biol- ogy, 30(24), R1463–R1464. https://doi.org/10.1016/j.cub.2020 .10.031, PubMed: 33352122 Intraub, H., & Richardson, M. (1989). Wide-angle memories of close-up scenes. Journal of Experimental Psychology: Learning, Memory, and Cognition, 15(2), 179–187. https://doi.org/10 .1037/0278-7393.15.2.179, PubMed: 2522508 Kersten, D., Mamassian, P., & Yuille, A. (2004). Object percep- tion as Bayesian inference. Annual Review of Psychology, 55, 271–304. https://doi.org/10.1146/annurev.psych.55.090902 .142005, PubMed: 14744217 Kocaoğlu, R., & Olguntürk, N. (2018). Color and visual complexity in abstract images: Part II. Color Research & Application, 43(6), 952–957. https://doi.org/10.1002/col.22266 Koehler, K., & Eckstein, M. P. (2017). Beyond scene gist: Objects guide search more than scene background. Journal of Experimen- tal Psychology: Human Perception and Performance, 43(6), 1177–1193. https://doi.org/10.1037/xhp0000363, PubMed: 28287759 Kuznetsova, A., Rom, H., Alldrin, N., Uijlings, J., Krasin, I., Pont- Tuset, J., Kamali, S., Popov, S., Malloci, M., Kolesnikov, A., Duerig, T., & Ferrari, V. (2020). The open images dataset V4. International Journal of Computer Vision, 128, 1956–1981. https://doi.org/10.1007/s11263-020-01316-z Lescroart, M. D., Stansbury, D. E., & Gallant, J. L.(2015). Fourier power, subjective distance, and object categories all provide plausible models of BOLD responses in scene-selective visual areas. Frontiers in Computational Neuroscience, 9. 135. https:// doi.org/10.3389/fncom.2015.00135, PubMed: 26594164 Lin, F., Hafri, A., & Bonner, M. F. (2022). Scene memories are biased toward high-probability views. Journal of Experimental Psychology: Human Perception and Performance, 48(10), 1116–1129. https://doi.org/10.1037/xhp0001045, PubMed: 35980704 Mahowald, K., Fedorenko, E., Piantadosi, S. T., & Gibson, E. (2013). Info/information theory: Speakers choose shorter words in pre- dictive contexts. Cognition, 126(2), 313–318. https://doi.org/10 .1016/j.cognition.2012.09.010, PubMed: 23116925 Mikolov, T., Sutskever, I., Chen, K., Corrado, G. S., & Dean, J. (2013). Distributed Representations of Words and Phrases and their Compositionality. Advances in Neural Information Process- ing Systems, 26. Miller, G. (1956). Human memory and the storage of information. IRE Transactions on Information Theory, 2(3), 129–137. https:// doi.org/10.1109/TIT.1956.1056815 Most, S. B., Chun, M. M., Widders, D. M., & Zald, D. H. (2005). Attentional rubbernecking: Cognitive control and personality in emotion-induced blindness. Psychonomic Bulletin & Review, 12(4), 654–661. https://doi.org/10.3758/BF03196754, PubMed: 16447378 OPEN MIND: Discoveries in Cognitive Science 458 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . / e d u o p m i / l a r t i c e - p d f / d o i / i / / . 1 0 1 1 6 2 o p m _ a _ 0 0 0 8 8 2 1 4 8 1 8 0 o p m _ a _ 0 0 0 8 8 p d / . i f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 Spatial Scene Memories Are Biased Towards a Fixed Amount of Semantic Information Greene and Trivedi Murphy, G. L. (2004). The big book of concepts. MIT Press. Nagle, F., & Lavie, N. (2020). Predicting human complexity percep- tion of real-world scenes. Royal Society Open Science, 7(5), 191487. https://doi.org/10.1098/rsos.191487, PubMed: 32537189 Oliva, A. (2005). Gist of the Scene. In L. Itti, G. Rees, & J. K. Tsotsos (Eds.), Neurobiology of Attention (pp. 251–256). Academic Press. https://doi.org/10.1016/B978-012375731-9/50045-8 Oliva, A., & Torralba, A. (2001). Modeling the shape of the scene: A holistic representation of the spatial envelope. International Jour- nal of Computer Vision, 42, 145–175. https://doi.org/10.1023 /A:1011139631724 Park, J., Josephs, E. L., & Konkle, T. (2021). Systematic transition from boundary extension to contraction along an object-to-scene continuum. PsyArXiv. https://doi.org/10.31234/osf.io/84exs Parkhurst, D. J., & Niebur, E. (2003). Scene content selected by active vision. Spatial Vision, 16(2), 125–154. https://doi.org/10 .1163/15685680360511645, PubMed: 12696858 Petrov, Y., & Zhaoping, L. (2003). Local correlations, information redundancy, and sufficient pixel depth in natural images. Journal of the Optical Society of America, 20(1), 56–66. https://doi.org /10.1364/josaa.20.000056, PubMed: 12542318 Piantadosi, S. T., Tily, H., & Gibson, E. (2011). Word lengths are optimized for efficient communication. Proceedings of the National Academy of Sciences of the United States of America, 108(9), 3526–3529. https://doi.org/10.1073/pnas.1012551108, PubMed: 21278332 Rosenholtz, R., Li, Y., & Nakano, L. (2007). Measuring visual clut- ter. Journal of Vision, 7(2), 1–22. https://doi.org/10.1167/7.2.17, PubMed: 18217832 Schyns, P. G. (1998). Diagnostic recognition: Task constraints, object information, and their interactions. Cognition, 67(1–2), 147–179. https://doi.org/10.1016/S0010-0277(98)00016-X, PubMed: 9735539 Shannon, C. E. (1948). A mathematical theory of communication. The Bell System Technical Journal, 27, 379–423, https://doi.org /10.1002/j.1538-7305.1948.tb01338.x Sontag, S. (2001). On Photography (1st edition). Picador. Sun, Z., & Firestone, C. (2022). Seeing and speaking: How verbal “description length” encodes visual complexity. Journal of Exper- imental Psychology: General, 151(1), 82–96. https://doi.org/10 .1037/xge0001076, PubMed: 34498910 Takarangi, M. K. T., Oulton, J. M., Green, D. M., & Strange, D. (2016). Boundary restriction for negative emotional images is an example of memory amplification. Clinical Psychological Sci- ence, 4(1), 82–95. https://doi.org/10.1177/2167702615569912 Tatler, B. W. (2007). The central fixation bias in scene viewing: Selecting an optimal viewing position independently of motor biases and image feature distributions. Journal of Vision, 7(14), 1–17. https://doi.org/10.1167/7.14.4, PubMed: 18217799 Tatler, B. W., Baddeley, R. J., & Gilchrist, I. D. (2005). Visual corre- lates of fixation selection: Effects of scale and time. Vision Research, 45(5), 643–659. https://doi.org/10.1016/j.visres.2004 .09.017, PubMed: 15621181 Torralba, A., & Oliva, A. (2003). Statistics of natural image catego- ries. Network, 14(3), 391–412. https://doi.org/10.1088/0954 -898X_14_3_302, PubMed: 12938764 Wang, X., Bylinskii, Z., Hertzmann, A., & Pepperell, R. (2020). To w a r d Q u a n t i f y i n g A m b i g u i t i e s i n A r t i s t i c I m a g e s . arXiv:2008.09688 [cs.CV]. https://arxiv.org/abs/2008.09688, https://doi.org/10.1145/3418054 Watson, D. M., Andrews, T. J., & Hartley, T. (2017). A data driven approach to understanding the organization of high-level visual cortex. Scientific Reports, 7(1), 3596. https://doi.org/10.1038 /s41598-017-03974-5, PubMed: 28620238 Xiao, J., Ehinger, K. A., Hays, J., Torralba, A., & Oliva, A. (2014). SUN database: Exploring a large collection of scene categories. International Journal of Computer Vision, 119, 3–22. https://doi .org/10.1007/s11263-014-0748-y Zeng, T., Tompary, A., Schapiro, A. C., & Thompson-Schill, S. L. (2021). Tracking the relation between gist and item memory over the course of long-term memory consolidation. eLife, 10, e65588. https://doi.org/10.7554/eLife.65588, PubMed: 34259626 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . / e d u o p m i / l a r t i c e - p d f / d o i / i . / / 1 0 1 1 6 2 o p m _ a _ 0 0 0 8 8 2 1 4 8 1 8 0 o p m _ a _ 0 0 0 8 8 p d / . i f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 OPEN MIND: Discoveries in Cognitive Science 459REPORT image
REPORT image
REPORT image
REPORT image
REPORT image
REPORT image
REPORT image
REPORT image
REPORT image
REPORT image
REPORT image
REPORT image
REPORT image
REPORT image
REPORT image
REPORT image
REPORT image
REPORT image
REPORT image
REPORT image
REPORT image

Scarica il pdf