REPORT

REPORT

Zipfian Distributions in Child-Directed Speech

Ori Lavi-Rotbain1

and Inbal Arnon2

1The Edmond and Lilly Safra Center for Brain Sciences, Hebrew University, Jerusalem, Israel
2Département de psychologie, Hebrew University, Jerusalem, Israel

Mots clés: Child-Directed Speech, Zipfian distribution, language learning

un accès ouvert

journal

ABSTRAIT

Across languages, word frequency and rank follow a power law relation, forming a distribution
known as the Zipfian distribution. There is growing experimental evidence that this well-
studied phenomenon may be beneficial for language learning. Cependant, most investigations
of word distributions in natural language have focused on adult-to-adult speech: Zipf’s law has
not been thoroughly evaluated in child-directed speech (CDS) across languages. If Zipfian
distributions facilitate learning, they should also be found in CDS. En même temps, several
unique properties of CDS may result in a less skewed distribution. Ici, we examine the
frequency distribution of words in CDS in three studies. We first show that CDS is Zipfian
across 15 languages from seven language families. We then show that CDS is Zipfian from
early on (six-months) and across development for five languages with sufficient longitudinal
data. Enfin, we show that the distribution holds across different parts of speech: Nouns, verbs,
adjectives and prepositions follow a Zipfian distribution. Ensemble, the results show that the
input children hear is skewed in a particular way from early on, providing necessary (but not
sufficient) support for the postulated learning advantage of such skew. They highlight the need
to study skewed learning environments experimentally.

je

D
o
w
n
o
un
d
e
d

F
r
o
m
h

t
t

p

:
/
/

d
je
r
e
c
t
.

m

je
t
.

/

e
d
toi
o
p
m

je
/

je

un
r
t
je
c
e

p
d

F
/

d
o

je
/

je

/

.

/

1
0
1
1
6
2
o
p
m
_
un
_
0
0
0
7
0
2
0
6
8
1
1
0
o
p
m
_
un
_
0
0
0
7
0
p
d

.

/

je

F

b
oui
g
toi
e
s
t

t

o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3

INTRODUCTION

When it comes to frequency, not all words are created equal: it is known that words in lan-
guage are not uniformly distributed. Plutôt, a small number of word types accounts for most
of the tokens across languages. In English for example, 0.3% of word types account for 50% de
all tokens, with the remaining words having very low frequencies ( Lequel, 2013). The relation
between a word’s rank and its frequency can be described using a power law: Differences in
frequency are large among the most frequent words, and almost non-existent among the least
frequent words. This power law relation results in a highly skewed frequency distribution with
a narrowed peak for the most frequent words, and a very long tail for the low frequency words
(Chiffre 1). This distribution is known as the Zipfian distribution (Zipf, 1949), and is character-
ized by a linear relation between log frequency and log rank (Chiffre 2). Natural language does
not follow Zipf’s law completely, with consistent prediction errors on both edges of the fre-
quency scale (Montemurro, 2001; Piantadosi, 2014). Par exemple, the most frequent words
are not as frequent as they are expected to be. We follow Piantadosi (2014) and use the term
“near-Zipfian” to refer to a distribution where Zipf’s law holds approximately.

Near-Zipfian distributions are found across languages, for multiple linguistic domains: Ils
are found when looking at the entire lexicon, at different parts of speech separately, and even

Citation: Lavi-Rotbain, O., & Arnon, je.
(2023). Zipfian Distributions in Child-
Directed Speech. Open Mind:
Discoveries in Cognitive Science,
7, 1–30. https://est ce que je.org/10.1162
/opmi_a_00070

EST CE QUE JE:
https://doi.org/10.1162/opmi_a_00070

Supplemental Materials:
https://doi.org/10.1162/opmi_a_00070

Reçu: 15 May 2022
Accepté: 30 Novembre 2022

Intérêts concurrents: The authors
declare no conflict of interest.

Auteur correspondant:
Ori Lavi-Rotbain
orilavirotbain@gmail.com

droits d'auteur: © 2022
Massachusetts Institute of Technology
Publié sous Creative Commons
Attribution 4.0 International
(CC PAR 4.0) Licence

La presse du MIT

Zipfian Distribution in CDS

Lavi-Rotbain and Arnon

je

D
o
w
n
o
un
d
e
d

F
r
o
m
h

t
t

p

:
/
/

d
je
r
e
c
t
.

m

je
t
.

/

e
d
toi
o
p
m

je
/

je

un
r
t
je
c
e

p
d

F
/

d
o

je
/

je

/

.

/

1
0
1
1
6
2
o
p
m
_
un
_
0
0
0
7
0
2
0
6
8
1
1
0
o
p
m
_
un
_
0
0
0
7
0
p
d

/

.

je

F

b
oui
g
toi
e
s
t

t

o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3

Chiffre 1. The observed word distributions for each language.

OPEN MIND: Discoveries in Cognitive Science

2

Zipfian Distribution in CDS

Lavi-Rotbain and Arnon

je

D
o
w
n
o
un
d
e
d

F
r
o
m
h

t
t

p

:
/
/

d
je
r
e
c
t
.

m

je
t
.

/

e
d
toi
o
p
m

je
/

je

un
r
t
je
c
e

p
d

F
/

d
o

je
/

je

/

.

/

1
0
1
1
6
2
o
p
m
_
un
_
0
0
0
7
0
2
0
6
8
1
1
0
o
p
m
_
un
_
0
0
0
7
0
p
d

.

/

je

F

b
oui
g
toi
e
s
t

t

o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3

Chiffre 2. The observed (in green) and expected (in red) word distributions for each language on a log-log scale.

OPEN MIND: Discoveries in Cognitive Science

3

Zipfian Distribution in CDS

Lavi-Rotbain and Arnon

for specific semantic classes such as taboo words (Bentz, Alikaniotis, Samardžić, & Buttery,
2017; Ferrer-i-Cancho, 2005; Mehri & Jamaati, 2017; Piantadosi, 2014). There are many dif-
ferent explanations for the origin of Zipfian distributions in language, with ongoing controversy
about the significance of this law and whether it tells us something fundamental about lan-
guage or not (par exemple., Ferrer-i-Cancho et al., 2020). D'une part, such distributions are found
in numerous domains across the physical worldfrom the distribution of visual objects
(Clerkin et al., 2017) and their co-occurrences (Lavi-Rotbain & Arnon, 2021), to population
size in American cities (Clauset et al., 2009) and crater size on the moon (see Newman,
2005 for a review) – where they are thought to reflect general mathematical principles not
unique to language (par exemple., scale-invariance: Chater & Brun, 1999). On the other hand, lan-
guage, unlike the physical world, is created and shaped by humans (Ferrer-i-Cancho et al.,
2020; Piantadosi, 2014; Semple et al., 2022). Par conséquent, the recurrence and preservation
of Zipfian distributions in language may reflect foundational properties of human language or
cognition (Christiansen & Chater, 2008; Ferrer-i-Cancho et al., 2020; Gibson et al., 2019;
Piantadosi, 2014; Semple et al., 2022).

Some explanations emphasize the impact of language being a communication system
under cognitive pressures, and suggest that Zipfian distributions emerge spontaneously under
the pressure to minimize listener and speaker efforts (Ferrer-i-Cancho & Sole, 2003), or from
the need for easy and fast communication (Ferrer-i-Cancho, 2016). A similar perspective
emphasizes the role of Shannon’s information theory (Shannon, 1948): Given that language
is a noisy communication channel, there is pressure for optimal coding of the lexicon (Coupé
et coll., 2019; Ferrer-i-Cancho et al., 2020). Zipfian distributions are seen as a form of optimal
coding under such accounts. These explanations attribute the presence of Zipfian distributions
to the communicative and cognitive pressures that impact language learning and use.

A different line of work highlights semantics as the driving force of Zipfian distributions.
Manin (2008) suggests a mechanism by which the tendency to avoid excessive synonymy
results in a Zipfian distribution. According to his model, the measure of specificity of a word
is closely related to frequency: Words that are more generic tend to be more frequent (e.g.
“city”), while words that are more specific are less frequent (par exemple., “London”). The variance
in specificity level across words, together with the changes in words’ meaning over time
and with the pressure to avoid excessive synonymy (using a word that is too specific for a
certain scenario), lead to a Zipfian distribution in which the semantic space can be divided
to layers of varying specificity (Manin, 2008). In a more recent model, Lestrade (2017) adds to
Manin’s specificity model and claims that semantics alone is not enough: it takes both seman-
tics and syntax in order to form a Zipfian distribution. According to Lestrade, syntax leads to
division of words into parts-of-speech categories that differ greatly in size. As in Manin (2008),
Lestrade claims that semantics leads to a variation in specificity level across words from the
same part-of-speech category. Using a computational model, Lestrade shows that only when
accounting for both semantics and syntax, the resulted distribution is Zipfian (Lestrade, 2017).

A host of additional explanations tie the presence of Zipfian distributions in language to
learnability pressures (Bentz, Alikaniotis, Cysouw, & Ferrer-i-Cancho, 2017; Coupé et al.,
2019; Lavi-Rotbain & Arnon, 2019, 2020, 2022). Skewed distributions, like the Zipfian one,
are more predictable than uniform distributions, making it easier to guess the next word. Ce
increased predictability may confer a learnability advantage for certain aspects of language
learning, like segmenting and learning words, for several reasons. D'abord, when it is easier to
predict upcoming elements, processing resources can be used more efficiently. En outre,
high frequency words can be learned early on, and used to facilitate learning of lower fre-
quency elements, as can be seen in infants use of their own name to segment adjacent words

OPEN MIND: Discoveries in Cognitive Science

4

je

D
o
w
n
o
un
d
e
d

F
r
o
m
h

t
t

p

:
/
/

d
je
r
e
c
t
.

m

je
t
.

/

e
d
toi
o
p
m

je
/

je

un
r
t
je
c
e

p
d

F
/

d
o

je
/

je

/

/

.

1
0
1
1
6
2
o
p
m
_
un
_
0
0
0
7
0
2
0
6
8
1
1
0
o
p
m
_
un
_
0
0
0
7
0
p
d

.

/

je

F

b
oui
g
toi
e
s
t

t

o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3

Zipfian Distribution in CDS

Lavi-Rotbain and Arnon

(Bortfeld et al., 2005), or in the use of familiar words to segment novel phonologically similar
words (Altvater-Mackensen & Mani, 2013). Enfin, lower frequency words may also benefit
from appearing next to high frequency ones: the contrast between high and low frequency
could make the lower frequency words easier to identify or learn. The facilitative potential
of the contrast between high and low frequency is supported by findings from several domains.
In a phenomenon known as stimulus-specific adaptation, rare tones cause an increased neu-
ronal reaction in rats, but only when they were rare enough (Rubin et al., 2016). De la même manière, dans
memory tasks, adults show improved memory for novel items only when these items were
relatively rare (Reggev et al., 2018).

The idea that Zipfian distributions may facilitate learning is starting to gain empirical sup-
port. The vast majority of lab-based studies looking at word segmentation and learning use a
uniform distribution where each item appears the same number of times (par exemple., Saffran et al.,
1996). This allows researchers to better control for frequency effects, but does not reflect the
natural skew of words in language. Cependant, several recent studies compared learning in
Zipfian and uniform distributions and found facilitative effects: Across different aspects of
language learning, including mapping between phrasal form and meaning, word segmenta-
tion, noun learning, and category formation, exposure to Zipfian distributions did not harm
learning, despite the lower frequency of certain elements (Kurumada et al., 2013; Schuler
et coll., 2017) and even led to facilitation in certain cases (Goldberg et al., 2004; Hendrickson
& Perfors, 2019; Lavi-Rotbain & Arnon, 2019, 2020, 2022). Surtout, if the propensity of
this distribution in language can be explained (even partially) by its learning advantage, alors
language directed to learners should also be Zipfian. If it is not, then the facilitation found in
the lab may be less relevant for actual language learning. More broadly, if CDS is not Zipfian,
this could undermine the role of learnability pressures in the emergence of Zipfian distribu-
tion. For both reasons, it is important to ask whether CDS is Zipfian or not.

What reason is there to think that CDS will not follow a Zipfian distribution? While CDS is
similar in certain aspects to adult-to-adult speech, some of its’ unique and well-documented
properties may lead word distributions to be less skewed. En particulier, CDS has a smaller
lexicon than adult-to-adult speech (Fernald & Simon, 1984; Roy et al., 2015), which may lead
to a more uniform distribution of types. CDS also includes many repeated sequences, tel que
variation sets (where consecutive sentences have overlapping words, par exemple., “Where is the
bunny? Here is the bunny”; Brodsky et al., 2007; Tal & Arnon, 2018), and recurring frequent
frames (Cameron-Faulkner et al., 2003; Mintz, 2003). These differences could lead to word
distributions that have fewer types and more similar frequencies across ranks, in contrast with
the steep change in frequency that is a hallmark of Zipfian distributions. C'est, the lexical and
structural differences between CDS and adult-to-adult speech could theoretically lead to a
word distribution that differs from the Zipfian one. In such a scenario, word distribution might
change over time, becoming “more Zipfian”: Several aspects of CDS, such as mean length of
utterance and type-token ratio, change across development and become more “adult-like”
with age (Roy et al., 2009). Word distributions may show a similar trend, starting with a less
skewed distribution and becoming more Zipfian as the child grows and their vocabulary
expands.

What do we currently know about word distributions in CDS? Fait intéressant, there is little
research on this question. Even though Zipfian distributions are assumed to be a basic property
of natural language, the existing findings documenting them are based on adult-to-adult
speech (most often written language), where both speakers and listeners are proficient
language users. To our knowledge, only few prior studies explored the distribution of words
in the input directed to young children (Dale & Spivey, 2006; Hendrickson & Perfors, 2019;

OPEN MIND: Discoveries in Cognitive Science

5

je

D
o
w
n
o
un
d
e
d

F
r
o
m
h

t
t

p

:
/
/

d
je
r
e
c
t
.

m

je
t
.

/

e
d
toi
o
p
m

je
/

je

un
r
t
je
c
e

p
d

F
/

d
o

je
/

je

/

.

/

1
0
1
1
6
2
o
p
m
_
un
_
0
0
0
7
0
2
0
6
8
1
1
0
o
p
m
_
un
_
0
0
0
7
0
p
d

.

/

je

F

b
oui
g
toi
e
s
t

t

o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3

Zipfian Distribution in CDS

Lavi-Rotbain and Arnon

Lavi-Rotbain & Arnon, 2022; Montag et al., 2018). Montag et al. (2018) focused on whether
the amount of the available input or its diversity are more important for learning. One of the
measurements the authors looked at was the distribution of words. The authors found that sam-
ples of 50,000, 20,000 et 2,000 words in CDS had a skewed frequency distribution that was
close to linear on a log-log scale. Hendrickson and Perfors (2019) focused on the facilitative
effect of Zipfian distributions for cross-situational word learning in an experimental setting. Dans
the appendix, they provide an analysis of the frequency distribution of nouns in a corpus of
English child-directed speech, as a validation of the use of this skewed distribution in the
experimental setting. For this dataset, they find that English nouns in CDS follow a near-Zipfian
distribution (as reflected in a linear relationship between log frequency and log rank). Dale and
Spivey (2006) analysed the correlations of different linguistic measures between child’s and
care-giver’s utterances within conversations by using three English corpora from CHILDES
(MacWhinney, 2000). One of these measures was frequency distributions of n-grams (pour
n = 2, 3 et 4). They showed that on a log-log scale there is a linear relation between the
n-grams frequency and their rank (Dale & Spivey, 2006). Lavi-Rotbain and Arnon (2022)
looked at the effect of skewed frequency distribution on speech segmentation. D'abord, they ana-
lysed corpora of CDS using measures from Shannon’s information theory and found that across
languages CDS has similar entropy levels, meaning it is similarly skewed. Then they used these
values to create artificial language distributions (Lavi-Rotbain & Arnon, 2022).

Surtout, in these studies, analysing the frequency distribution of words was not the
main focus of the study, leaving many open questions. We do not know whether a similar
distribution holds for the entire lexicon, across languages, across development and for different
parts of speech. En outre, the fit to the Zipfian distribution was illustrated by finding a close
fit to a linear distribution on a log-log scale. While informative, this does not necessarily indi-
cate that the original distribution is a Zipfian one, since several different distributions appear
linear on a log-log scale (Clauset et al., 2009).

The Current Study

In the current study, we expand on the existing literature by exploring word distributions in
child-directed speech across 15 languages with three goals in mind. The first, and primary one,
is to see if words follow a Zipfian distribution in child-directed speech across different
languages (Étude 1). Such a finding would support the generality of Zipfian distributions in
language and their possible role in language learning. We can also use this data to see whether
the parameters of the Zipfian distributionα and βare similar across languages. A previous
study estimated the parameters of the Zipfian distribution based on Holy Bible translations
across a large number of languages and found differences across languages (Mehri & Jamaati,
2017). Cependant, they looked at written language, rather than spoken, and tried to fit only the
middle region of the distribution, excluding the most and least frequent words (Mehri &
Jamaati, 2017). We will ask whether a similar variation will be found in CDS. The second goal
is to ask whether there is a developmental change in the distribution of words in child-directed
speechas was found for other structural propertiesor whether the distribution is Zipfian
from the very start (Étude 2). Enfin, we want to see if words follow a Zipfian distribution in
child-directed speech across different parts of speech (Étude 3), as was found for adult-to adult
speech (Piantadosi, 2014).

We decided to look at four parts of speech: nouns, verbs, adjectives and prepositions. Le
skewed distribution of content words like nouns, verbs and adjectives, can be attributed in part
to meaning: We talk with infants more about teddy-bears, Par exemple, than stocks. While this

OPEN MIND: Discoveries in Cognitive Science

6

je

D
o
w
n
o
un
d
e
d

F
r
o
m
h

t
t

p

:
/
/

d
je
r
e
c
t
.

m

je
t
.

/

e
d
toi
o
p
m

je
/

je

un
r
t
je
c
e

p
d

F
/

d
o

je
/

je

/

.

/

1
0
1
1
6
2
o
p
m
_
un
_
0
0
0
7
0
2
0
6
8
1
1
0
o
p
m
_
un
_
0
0
0
7
0
p
d

/

.

je

F

b
oui
g
toi
e
s
t

t

o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3

Zipfian Distribution in CDS

Lavi-Rotbain and Arnon

does not explain the particular way frequency decreases, it could explain the presence of few
highly frequency words and the long tail of the distribution. Cependant, finding such a distribu-
tion for prepositions would suggest the impact of factors beyond meaning on frequency for
several reasons. D'abord, prepositions are not expected to vary so much in their real-world fre-
quency: there is no reason to think people talk more about objects being “on” other object
than “behind” other objects. En outre, different prepositions can be used to describe the
same event: We can describe things as being placed ‘under’ or ‘above’, or happening ‘before’
or ‘after’, et ainsi de suite, meaning that the presence of an event is not fully predictive of the prep-
osition used to describe it. Enfin, and most importantly, the meaning of prepositions varies
across languages, with languages carving up the meaning space differently (par exemple., Bowerman &
Choi, 2001; Christiansen & Chater, 2008; Levinson et al., 2003). Par exemple, the difference
between ‘in’ and ‘at’ in English is not reflected in Hebrew where the same preposition is used
for both meanings. C'est, despite the physical world being the same, the choice of preposi-
tions differs across languages, meaning the world itself cannot be the leading cause for finding
a Zipfian distribution in the use of prepositions.

We explore the first question by asking how closely word distributions adhere to Zipf’s law
in CDS across 15 languages from seven language families. We address the second question by
analysing word distributions across development for the five languages for which we have
large enough longitudinal corpora. We address the third question by analysing word distribu-
tions across different parts of speech for the five languages for which we have morphologically
tagged data. Ensemble, the studies aim to provide a comprehensive assessment of word distri-
butions in CDS. To pre-empt the results, we find that word distributions follow a near-Zipfian
distribution across languages, across development and across parts of speech illustrating their
prominence in children’s learning environment.

STUDY 1: DO WORDS IN CDS FOLLOW A ZIPFIAN DISTRIBUTION
ACROSS LANGUAGES?

We analysed the distribution of words in CDS for all the languages whose corpora on
CHILDES (MacWhinney, 2000) matched our selection criteria (see details below). This resulted
in the analysis of 15 languages, from seven language families: Germanic (including English
British and North-American, German, Dutch, Swedish, Danish and Norwegian); Latinate
(including French, Spanish, Portuguese and Catalan); Uralic (Estonian); Slavic (Polish); Semitic
(Hebrew), Japonic ( Japonais), and Sino-Tibetan (Mandarin), (see Table 1 for more information
about the corpora). We performed two analyses: D'abord, we looked at the entire corpora avail-
able for each language and tried to fit it to a Zipfian distribution in order to see if CDS is Zipfian
across languages. We then used small samples from languages with large enough corpora to
assess the stability of our estimates. This additional analysis allowed us to also ask how similar
the parameters of the Zipfian distribution are across languages, or whether they differ between
languages, as was previously found for the Holy Bible translations (Mehri & Jamaati, 2017).

Methods

We included all the corpora available in these languages with the following restrictions: le
data was collected during parent-child interactions (as opposed to investigator-child or peer
interactions); from typically developing children who were 3;6 years old or under during the
recording. We only looked at younger ages since we wanted to focus on the input to young
learners. We extracted the child’s age from the transcripts directly or from the corpora descrip-
tion available on CHILDES. The age criterion was not applied in Catalan since the data was

OPEN MIND: Discoveries in Cognitive Science

7

je

D
o
w
n
o
un
d
e
d

F
r
o
m
h

t
t

p

:
/
/

d
je
r
e
c
t
.

m

je
t
.

/

e
d
toi
o
p
m

je
/

je

un
r
t
je
c
e

p
d

F
/

d
o

je
/

je

/

.

/

1
0
1
1
6
2
o
p
m
_
un
_
0
0
0
7
0
2
0
6
8
1
1
0
o
p
m
_
un
_
0
0
0
7
0
p
d

.

/

je

F

b
oui
g
toi
e
s
t

t

o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3

Child’s Age
Range
0;2–3;6

0;5–3;6

0;5–3;6

0;11–3;6

1;5–3;6

0;6–3;6

0;10–6;11

0;11–3;6

1;0–3;6

1;5–3;6

0;8–3;6

1;1–2;9

0;9–3;6

0;11–3;5

1;8–2;3

1;1–4;2

Language
British English

North-American

English

German

French

Dutch

Japonais

Polish*

Spanish

Swedish

Portuguese

Hebrew

Norwegian

Estonian

Danish

Mandarin

Catalan

Résumé

Tableau 1.

Summary of corpora measures across languages for Study 1.

Non.
Corpora
12

Non.
Tokens
6311249

Non.
Types Word Frequency (per million)
1 – 303005 (0.16 – 48010.31)
27476

34

4876774

24573

1 – 230355 (0.21 – 47235.12)

7

8

5

6

8

2168002

37018

1 – 71238 (0.46 – 32858.83)

1540284

19327

1 – 59852 (0.65 – 38857.77)

1036586

18717

1 – 41646 (0.96 – 40176.12)

941006

25648

1 – 45886 (1.06 – 48762.71)

794183.7

44425

1 – 32172.19 (1.26 – 40509.76)

12

353104

10057

1 – 13437 (2.83 – 38053.94)

2

2

6

2

5

1

2

4

341280

9466

1 – 16924 (2.93 – 49589.78)

309296

7562

1 – 21416 (3.23 – 69241.12)

300766

13801

1 – 16048 (3.32 – 53357.09)

183658

8306

1 – 9135 (5.44 – 49739.19)

167666

10057

1 – 10344 (5.96 – 61694.08)

155826

4102

1 – 8421 (6.42 – 54041.05)

150852

7095

1 – 6305 (6.63 – 41795.93)

132410

6416

1 – 8051 (7.55 – 60803.56)

un

1.57

1.52

1.42

1.53

1.48

1.30

1.16

1.39

1.40

1.37

1.19

1.32

1.21

1.50

1.42

1.30

β

19.48

18.34

12.69

17.98

12.63

7.12

4.04

9.96

7.58

5.44

3.69

6.92

6.02

7.71

11.94

6.27

Pearson’s-r
0.97

0.965

0.998

0.99

0.998

0.993

0.997

0.994

0.998

0.981

0.996

0.996

0.956

0.997

0.989

0.985

Mean = 1.38;
SD = 0.13

Mean = 9.86;
SD = 5.13

Mean = 0.988;
SD = 0.013

* For Polish the data is taken from a list including a summary of words and their relative frequencies. Hence the one digit precision.

Z
je
p
F
je
un
n
D
je
s
t
r
je
b
toi
t
je
o
n

je
n
C
D
S

L
un
v
je

R.
o
t
b
un
je
n

un
n
d
UN
r
n
o
n

je

D
o
w
n
o
un
d
e
d

F
r
o
m
h

t
t

p

:
/
/

d
je
r
e
c
t
.

m

je
t
.

/

e
d
toi
o
p
m

je
/

je

un
r
t
je
c
e

p
d

F
/

d
o

je
/

je

/

.

/

1
0
1
1
6
2
o
p
m
_
un
_
0
0
0
7
0
2
0
6
8
1
1
0
o
p
m
_
un
_
0
0
0
7
0
p
d

.

/

je

F

b
oui
g
toi
e
s
t

t

o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3

Ô
P.
E
N
M.
N
D

je

:

je

D
s
c
o
v
e
r
je
e
s

je

n
C
o
g
n

je
t
je
v
e
S
c
e
n
c
e

je

8

Zipfian Distribution in CDS

Lavi-Rotbain and Arnon

too limited to be screened out by age, nor in Polish since the data we have is a summary of
words and their relative frequencies instead of the original transcripts (see Table 1 for details).

In all analyses, we looked only at utterances produced by adults (both care-givers and
experimenters) and removed utterances produced by the child. Enfin, we only looked at lan-
guages that had at least 100,000 tokens after applying the previous restrictions. This minimal
threshold was applied to increase the reliability of the estimations of the Zipfian distribution
parameters we are calculating (see Equation 1). Previous work has shown that other distribu-
tional measures such as entropy are reliable with a corpus of 50,000 tokens and above (Bentz,
Alikaniotis, Cysouw, et coll., 2017). We chose a more stringent threshold of 100,000 tokens to
ensure reliable results. After applying these criteria, we were left with 15 languages: English
(British and North-American), German, Dutch, Swedish, Danish, Norwegian, French, Spanish,
Portuguese, Catalan, Polish, Estonian, Hebrew, Japanese and Mandarin.

For each language, we assessed words distributions for the entire corpus (collapsing over
individual conversations/dyads, as was done in previous studies of adult speech (Piantadosi,
2014). We did not assess word distributions within single conversations because (un) there is no
expectation that each individual conversation will follow a Zipfian distribution (in the same
way that word frequencies in a single conversation may not reflect their frequency in a lan-
guage), et (b) because each conversation does not provide enough types/tokens data to
accurately estimate the distribution. We started by extracting single word frequencies from
all of the available corpora. A word was defined by its orthographic form. While this definition
has its limitations (we discuss them further in the discussion), such as treating “dog” and
“dog’s” as two independent words, it is the standard definition of a word in the study of Zipfian
distributions and in many cross-linguistic corpus studies (par exemple., Bentz, Alikaniotis, Cysouw,
et coll., 2017; Geertzen et al., 2016; Tal & Arnon, 2018). We cleaned the data by removing
comments made by the transcriber (see available Python code at https://osf.io/bp62q/). After
creating a list of words and their frequencies, we calculated the rank for each word (the most
frequent word was ranked 1, the 2nd frequent was ranked 2 et ainsi de suite). Words with the same
frequency were assigned ranks randomly (par exemple., if two words appeared 100 times, one was
given rank X and the second rank X + 1)1.

How to Assess Whether Word Distributions Are Zipfian? After obtaining a frequency distribution
of words, we evaluated “how Zipfian” it is. It is not straightforward to estimate whether a
particular word distribution is Zipfian. The simplest approach would be to look at the corre-
lation between frequency and rank on a log-log basis and assess how linear this relation is
(under a Zipfian distribution it should be linear). This is the criteria used in some studies
(par exemple., Hendrickson & Perfors, 2019). Cependant, since other distributions, beside a power
law, are linear on a log-log scale, this cannot tell us if the original distribution follows a power
law or not (Clauset et al., 2009). Donc, another approach is needed. One method, pro-
posed by Clauset et al. (2009) involves estimating the minimal rank from which the power-law
distribution is supposed to hold; estimating only alpha and then measuring the maximal dis-
tance between the observed distribution and the expected power-law distribution using
Kolmogorov–Smirnov statistic. This method assumes that word distributions follow a true
power-law. We opted not to use it to assess the fit for several reasons: (un) word distributions
in language are only nearly-Zipfian, diverging from from the expected Zipfian distribution in

1 We did not use the split-to-half method proposed by Piantadosi (2014) where the corpus is divided into half
by a binomial distribution with one half used to assess frequency and the second used to assess rank (lequel
reduces the dependency between the measures of rank and frequency) since we did not want to reduce the
number of languages we can look at (see details on page 15 and Appendix A).

OPEN MIND: Discoveries in Cognitive Science

9

je

D
o
w
n
o
un
d
e
d

F
r
o
m
h

t
t

p

:
/
/

d
je
r
e
c
t
.

m

je
t
.

/

e
d
toi
o
p
m

je
/

je

un
r
t
je
c
e

p
d

F
/

d
o

je
/

je

/

.

/

1
0
1
1
6
2
o
p
m
_
un
_
0
0
0
7
0
2
0
6
8
1
1
0
o
p
m
_
un
_
0
0
0
7
0
p
d

/

.

je

F

b
oui
g
toi
e
s
t

t

o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3

Zipfian Distribution in CDS

Lavi-Rotbain and Arnon

systematic ways (Piantadosi, 2014), (b) Clauset’s method has not been applied broadly to
linguistic data (Clauset et al only use it to analyze Moby Dick in English) and we wanted
our investigation of child-directed speech to be comparable to existing analyses of the fit to
the Zipfian distribution in adult speech, et (c) most importantly, we conducted several anal-
yses of our data using this method, which suggest that it generates inconclusive results that do
not reflect meaningful variation (see Appendix B). Plutôt, we opted to use the method used in
previous studies of adult speech (Piantadosi, 2014) where the Zipfian nature of the distribution
is assessed by (1) estimating the parameters of the distribution (as described below, Équation 1),
(2) finding the expected frequency distribution based on the estimated parameter values, et
(3) evaluating the goodness of the fit between the observed frequency distribution and the
expected one under a Zipfian distribution (Piantadosi, 2014).

Ici, we use the same method. We start by estimating the parameters of the Zipfian distri-
bution (see Equation 1). In this equation, r is the word’s rank; α is the exponent of the power
law while β is a correction added to the original Zipf’s law by Mandelbrot to create a better fit
to actual language data (Mandelbrot, 1953). The sign “∝” indicates proportionality: F(r) is pro-
portional to the right-hand side of the equation.

f rð Þ ∝

1
r þ β

ð

Þα

(1)

Our application differs from that of Piantadosi (2014) only in that we did not use the split-to-
half method where the corpus is divided into half, with one half used to assess frequency and
the second used to assess rank. This is claimed to reduce the dependency between the mea-
sures of rank and frequency. Cependant, we did not use it here since we did not want to reduce
the number of languages we can look at. We need at least 100,000 tokens to reliably estimate
the parameters, and did not want to exclude corpora with fewer than 200,000 tokens (nous
would have to remove five of the languages we currently use). To validate our use of the entire
sample, we conducted the split-to-half method for the ten languages for which there was
enough data, showing that the parameter estimates and the fit to a Zipfian distribution are
almost identical when using this method and using the entire sample (Pearson’s-r between
the alpha based on the entire sample and the alpha estimated using the split-half method =
0.999; Pearson’s-r for beta from the entire sample and the split-half method = 0.999, voir
Appendix A for details).

In order to estimate the α and β that give the best fit of the data to Equation 1, we used the
maximum likelihood estimator (MLE, using the following code in R: https://osf.io/bp62q/), lequel
is a commonly-used algorithm to solve parameter estimation problems (Linders & Louwerse,
2020; Piantadosi, 2014). Finding α and β allows us to (1) estimate the fit of a Zipfian distribution
to the data, et (2) see if they are similar across the different languages, despite differences in
corpus size, lexicon size, and morphological complexity. For each language, we find the
expected Zipfian distribution, based on the α and β values we found. We used the formula for
the probability mass function of a Zipfian distribution to calculate the expected frequency (voir
Équation 2): multiplying each probability by the total number of tokens in the language.

p rð Þ ¼

1
ð
r þ β

Þα *

XN

r¼1

1
ð
r þ β

Þα

(2)

We correlated the observed frequencies with the expected frequencies using Pearson correla-
tions and obtained the Pearson’s-r of the correlation. Note that the correlation we report is the
one between observed and expected frequency, and not between frequency and rank: We want
to know how much of the observed frequency can be explained by the predicted frequency

OPEN MIND: Discoveries in Cognitive Science

10

je

D
o
w
n
o
un
d
e
d

F
r
o
m
h

t
t

p

:
/
/

d
je
r
e
c
t
.

m

je
t
.

/

e
d
toi
o
p
m

je
/

je

un
r
t
je
c
e

p
d

F
/

d
o

je
/

je

/

/

.

1
0
1
1
6
2
o
p
m
_
un
_
0
0
0
7
0
2
0
6
8
1
1
0
o
p
m
_
un
_
0
0
0
7
0
p
d

/

.

je

F

b
oui
g
toi
e
s
t

t

o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3

Zipfian Distribution in CDS

Lavi-Rotbain and Arnon

based on a Zipfian distribution (and not how much the change in rank explains the change in
frequency). While α and β were estimated from the data, they are not expected to predict the
observed data perfectly. As in any attempt to explain real-world behaviour, there is a certain
degree of noise in our estimates. By calculating the fit between the observed (réel) distribution
and the expected one (generated by plugging the α and β we found into Equation 2), we can see
how close they are to each other. In other word, we can see how “Zipfian” the observed distri-
bution is. We present graphs of the observed and expected frequency given rank on a regular and
a log-log scale (Chiffre 1 et 2 correspondingly). The regular scale shows the skewness of the
distribution, the sharp decrease in frequency, and the long tail for the least frequent words. Le
log-log plots show the linear relation between log(rank) and log (frequency).

We wanted to check the reliability of our estimates and also to see if the parameters of the
distribution will differ across languages in CDS, as was found for bible translations (Mehri &
Jamaati, 2017). Pour faire ça, we divided the full corpora into smaller samples of 100,000 tokens.
We only examined languages that provided at least three samples of this size, so we would
have a reasonable number of data points. This resulted in the use of nine languages from four
language families (British English, Dutch, German, Swedish, French, Portuguese, Spanish,
Japanese and Hebrew). We did not use North-American English even though it had a large
enough corpus, since we did not want too many English samples (British alone had 69 sam-
ples). We created the small samples for each language by reading each transcription file from
beginning to end until a sample of at least 100,000 tokens was created. This way the samples
were conversationally continuous. For each sample we estimated α and β as described above.
This yielded several estimations of each parameter for each language, which enabled us to see
whether the parameters from the entire corpus were similar to those from the sub-samples, et
to compare our estimations within and across languages.

Results

Is CDS Zipfian Across Languages? The observed and expected distributions for each language
are plotted in Figure 2. In these plots, we can see that the two curves are close to one another,
indicating that the observed distribution is very close to the expected Zipfian one. This is
reflected in the high and significant Pearson’s-r for the fit between the two, indicating that α
and β we found provide a good fit to the data2. C'est, the distribution of words in CDS is very
close to a Zipfian distribution across languages.

Are Word Distributions in CDS Similar Across Languages? To see how similar word distributions
are across languages, we first compared the parameter values we estimated based on the full
corpora for each language to the other languages. These parameters, and especially α, directly
affect the shape of the distribution. β affects the distribution less and is not expected to be
similar across language. Cependant, since α is the exponent, it dictates the slope of the curve:
For larger values of α, the resulting distribution will be steeper with a faster decrease in
frequency along the rank axis. A similar α across languages, will suggest that word distribu-
tions show a similar decrease in frequency, resulting in a steady difference in frequency along
the slope.

When we look at these values, we see that α has a small SD across languages compared to
β (un: mean = 1.38, SD = 0.13, range = 1.16–1.57; β: mean = 9.86, SD = 5.13), suggesting that

2 The reported Pearson’s-r values were obtained by correlating the observed and expected distribution. Dans
addition, a similar calculation was performed on a log-log scale of these variables, with similar results.

OPEN MIND: Discoveries in Cognitive Science

11

je

D
o
w
n
o
un
d
e
d

F
r
o
m
h

t
t

p

:
/
/

d
je
r
e
c
t
.

m

je
t
.

/

e
d
toi
o
p
m

je
/

je

un
r
t
je
c
e

p
d

F
/

d
o

je
/

je

/

/

.

1
0
1
1
6
2
o
p
m
_
un
_
0
0
0
7
0
2
0
6
8
1
1
0
o
p
m
_
un
_
0
0
0
7
0
p
d

.

/

je

F

b
oui
g
toi
e
s
t

t

o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3

Zipfian Distribution in CDS

Lavi-Rotbain and Arnon

the decrease in frequencies by rank is relatively similar across languages. Par exemple, pour
Catalan we received the same α as for Japanese (α = 1.3) even though our Catalan corpus
had only 130,000 tokens and 6,400 les types, while the Japanese corpus had over 900,000 tokens
et plus 25,000 les types. To see whether α is similar across languages and language families, nous
estimated the parameters again using smaller samples of 100,000 tokens each (as described in
the Methods section). We generated a total of 149 samples with the following division across
languages: 69 for British English; 21 for German; 19 for French; 14 for Japanese; 11 for Dutch;
6 for Spanish; et 3 for Portuguese, Hebrew and Swedish. We estimated the parameters for
each sample and plotted the results per language (Chiffre 3). A one-way ANOVA showed that
the effect of language on α is significant (F(8, 140) = 28.26, p < 0.001), as well as the effect of language family on α (F(3, 145) = 36.03, p < 0.001). Both effects remained significant after excluding Hebrew that has the lowest α. In addition, α differed across languages from the same language family: the effect of language was significant also when looking only at the four Ger- manic or at the three Latinate languages (Germanic: F(3, 100) = 11.02, p < 0.001; Latinate: F(2, 25) = 21098, p < 0.001). That is, while α spans a relatively small range across languages in CDS, it still shows variation between languages, even ones from the same language family. These results are consistent with the variation found for bible translations across languages (Mehri & Jamaati, 2017), and when comparing word frequency distributions for the Swadesh list across languages (Piantadosi, 2014). To examine the stability of the parameters we estimated for each language based on the entire corpus, we wanted to compare them to the α estimates based on the sub-samples. Finding that the parameters are similar, would validate the use of the smaller samples for com- paring α across languages, and for estimating α from relatively smaller corpora (each of our sub-samples had only 100,000 tokens). To explore this, we compared α calculated for the entire corpora to the range of α values estimated for the sub-samples. For visualisation, we added to Figure 3 the α from the entire corpora (represents by “X”). The correlation between the two is very high (Pearson’s-r = 0.98). Interestingly, for all tested languages, the α calculated based on the full corpora is slightly higher than the ones averaged across samples (e.g., for l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . / e d u o p m i / l a r t i c e - p d f / d o i / i . / / 1 0 1 1 6 2 o p m _ a _ 0 0 0 7 0 2 0 6 8 1 1 0 o p m _ a _ 0 0 0 7 0 p d / . i f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 Figure 3. The estimated α by language. Color represents language family. Boxes indicate quartiles. “X” represents α value obtained based on the full corpora. OPEN MIND: Discoveries in Cognitive Science 12 Zipfian Distribution in CDS Lavi-Rotbain and Arnon Figure 4. The correlation between α calculated based on CDS (y-axis) and α calculated based on bible translations for 12 languages (x-axis, taken from Mehri & Jamaati, 2017, Table 1, p. 2473). Hebrew α for full corpora is 1.19 while the average of the three smaller samples is 1.17). How- ever, the values for the entire corpora fall within the range obtained from the different samples for all languages except Portuguese (see details and more discussion of the relation between α and corpus size in Study 2, Figure 5)3. Is the Word Distribution in CDS Similar to That of Adult-to-Adult Speech? Another interesting com- parison is between the parameters of the distribution obtained for CDS and for adult-to-adult speech. The α values found here are generally similar to what was previously found for adult- to adult speech, though there is large variation between different adult corpora. For example, based on the American National Corpus (Reppen & Ide, 2004), Piantadosi reports that α = 1.13 (Piantadosi, 2014). However, other adult-to-adult spoken and written English corpora yielded higher values for English (range = 1.728–1.940; Arnold, 2015, page 14; α = 1.95 for words appearing in Moby Dick; Clauset et al., 2009). Putting aside the variation between different corpora of the same language, the previous section showed differences in α across languages (e.g., English has the highest α while Hebrew has the lowest). We can ask if α varies in similar ways across languages in both adult and child-directed speech. If so, it can suggest that the variation is driven by linguistic properties of the languages shared by all speech registers (such as morphological complexity or lexicon size). To examine this more closely, we compared the α we received based on the entire corpus to the ones generated from bible translations (Mehri & Jamaati, 2017, Table 1, p. 2473). There were 12 overlapping languages between the studies, with two estimations for each (Figure 4). Note that there are differences in how α was assessed in each of the studies: In our study we tried to fit the entire distribution to Equation 1 (including both α and β as parameters), while Mehri & Jamaati tried to fit only the middle portion of the distribution to the simplified version of Zipf’s law (including only α). Interestingly, despite these differences, the correlation between the two α values is positive and significant. Languages 3 The difference in α may be related to corpus size, for example, larger samples tend to have larger vocab- ularies, but may also have higher frequencies for the frequent words making the slope slightly steeper. We dis- cuss the relation between corpus size and α further in Study 2 (Figure 5). OPEN MIND: Discoveries in Cognitive Science 13 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . / e d u o p m i / l a r t i c e - p d f / d o i / i / / . 1 0 1 1 6 2 o p m _ a _ 0 0 0 7 0 2 0 6 8 1 1 0 o p m _ a _ 0 0 0 7 0 p d . / i f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 Zipfian Distribution in CDS Lavi-Rotbain and Arnon Figure 5. α by number of tokens per sample. Each point represents ten samples. Bars indicate the SD of the samples. that have a higher α based on bible translations also have a higher α in CDS (n = 12, Pearson’s-r = 0.67, p < 0.05), suggesting the variation is impacted (at least in part) by linguistic/structural properties of the language itself. In the next study, we ask whether the good fit to a Zipfian distribution is found across development. STUDY 2: DOES CDS FOLLOW A ZIPFIAN DISTRIBUTION ACROSS DEVELOPMENT? In Study 1, we found that CDS follows a Zipfian distribution when we collapse across ages. However, it is possible that the skew is driven by speech directed to older children. To test this, we now analyze the distribution of words in CDS separately for different ages. This enabled us to see if the shape of the distribution is already found in infancy or whether it changes over time, like other linguistic measures that become more “adult-like” as the child grows (Roy et al., 2015). We only looked at languages where there was sufficient data to divide into age bins, and that had an age range that started in early infancy. We also wanted to look at languages from diverse language families. Based on these two parameters, we focused on five languages: English (British), German, French, Japanese and Hebrew. Methods For each language, we divided the available transcripts into fixed age bins of six months, rang- ing from birth up to five years old (i.e., the 1st bin included all available data between 0;0–0;5 months, the 2nd bin included 0;6–0;11 months and so on). Because the available data in each language began and ended at different ages (i.e., the youngest available age in French was eleven months, while in English it was 2-months), we could not look at the entire range for all languages (see Table 2 for details). For each age bin, we plotted the observed distribution of words (Figure 6). To further analyze the data, we needed to determine the minimal corpus size (in tokens) for which the parameter estimates would be stable. This step was required since the amount of data we had was not uniformly distributed across age: We had more data (in terms of both OPEN MIND: Discoveries in Cognitive Science 14 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . / e d u o p m i / l a r t i c e - p d f / d o i / i . / / 1 0 1 1 6 2 o p m _ a _ 0 0 0 7 0 2 0 6 8 1 1 0 o p m _ a _ 0 0 0 7 0 p d . / i f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 Child’s Age Range 0;2–0;5 0;6–0;11 1;0–1;5 1;6–1;11 2;0–2;5 2;6–2;11 3;0–3;5 3;6–3;11 4;0–4;5 4;6–4;11 Summary Table 2. Summary of corpora measures across development for Study 2 for British English. No. Tokens No. Types 5672 15011 24047 222896 2432536 2276518 1272241 321097 262627 186765 694 1147 2954 5627 15012 16728 12695 8176 7694 6375 Word Frequency (per million) 1 - 336 (176.3 - 59238.36) 1 - 735 (66.62 - 48964.09) 1 - 1916 (41.59 - 79677.3) 1 - 9309 (4.49 - 41763.87) 1 - 116879 (0.41 - 48048.21) 1 - 110405 (0.44 - 48497.31) 1 - 62026 (0.79 - 48753.34) 1 - 12771 (3.11 - 39773.03) 1 - 10464 (3.81 - 39843.58) 1 - 7288 (5.35 - 39022.3) α – 1.42 1.14 1.54 1.55 1.54 1.52 1.46 1.44 1.41 β – 16.43 3.18 18.33 18.38 18.15 17.42 16.4 14.9 13.56 Pearson’s-r – 0.969 0.96 0.984 0.971 0.971 0.972 0.979 0.982 0.988 Mean = 1.49; SD = 0.06 Mean = 16.73; SD = 1.88 Mean = 0.978; SD = 0.006 Z i p f i a n D i s t r i b u t i o n i n C D S L a v i - R o t b a i n a n d A r n o n l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . / e d u o p m i / l a r t i c e - p d f / d o i / i / . / 1 0 1 1 6 2 o p m _ a _ 0 0 0 7 0 2 0 6 8 1 1 0 o p m _ a _ 0 0 0 7 0 p d . / i f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 O P E N M N D I : i D s c o v e r i e s i n C o g n i t i v e S c e n c e i 1 5 Zipfian Distribution in CDS Lavi-Rotbain and Arnon l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . / e d u o p m i / l a r t i c e - p d f / d o i / i . / / 1 0 1 1 6 2 o p m _ a _ 0 0 0 7 0 2 0 6 8 1 1 0 o p m _ a _ 0 0 0 7 0 p d . / i f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 Figure 6. Word distributions in British English across development, in bins of six months. Note that the age bin of 0;2–0;5 did not pass the minimal size requirement of 50,000 tokens, but still shows a highly skewed distribution. number of tokens and types) for ages 24–36-months, and less for the other ages (see Figure S1). This inherent bias in our data could influence our variables estimates. We used the largest age bin (2;0–2;6 years old, British English, containing 2.4 million tokens) to find the minimal bound for reliable parameter estimation. We sampled data from this age bin to create samples of varying size: We started from 5,000 words, and increased the sample by 5,000 words until we reached 100,000 words. We created the samples by randomly selecting transcripts, and then reading each transcription file from beginning to end until we reached the required num- ber of tokens. This created samples that were conversationally continuous. We created ten OPEN MIND: Discoveries in Cognitive Science 16 Zipfian Distribution in CDS Lavi-Rotbain and Arnon different samples for each sample size (e.g., ten 5,000-word samples, ten 10,000-word sam- ples, etc.) Figure 5 shows the change in α as a result of changing the sample size: α increased with sample size, and began to stabilize at 50,000 tokens (as reflected in the small SD of 0.02 and the smaller changes in its value). This finding is in the same direction as in Study 1: α calcu- lated on the full corpus was slightly higher than α calculated on samples of 100,000 tokens (Figure 3). While in Study 1 we chose a minimal size of 100,000 tokens per sample, here we could not choose such a stringent minimal bound since it would have left us with too little samples. Therefore, we chose 50,000 tokens as our minimal bound. This left us with thirty usable age bins. For age bins with at least 50,000 tokens, we conducted the same analyses as in Study 1: (1) evaluating α and β; (2) calculating Pearson’s-r between the observed and the expected distri- butions; (3) plotting the two curves together. We used mixed effect regression models to see if α or Pearson’s-r are affected by age. Specifically, we ask two questions. The first, does the dis- tribution become more Zipfian during development, or is it near-Zipfian from the start? To test this, we looked at whether Pearson’s-r changes across the different age bins, and whether it is high even for the youngest age bins. The second question was whether the slope of the dis- tribution changes across development. Such a change would indicate that the distribution becomes more or less skewed as the child grows. To test this, we looked at whether α changes across the different age bins. l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . / e d u o p m i / l a r t i c e - p d f / d o i / i . / / 1 0 1 1 6 2 o p m _ a _ 0 0 0 7 0 2 0 6 8 1 1 0 o p m _ a _ 0 0 0 7 0 p d . / i f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 Results Due to lack of space, we only show the results for British English (Table 2), where we had the largest age range (the full results are provided in Table S1 and Figure S1–S5 in the Supplementary section). Figure 6 shows word distributions for British English across develop- ment. Both Table 2 and Figure 6 include all age bins for British English, including the one containing less data than our minimal bound. All age bins (even the age bin of two-to-five months that included only 5,000 tokens), followed a skewed frequency distribution. This held for all age bins in the other languages as well, meaning that even small samples of CDS are skewed. Thirty age bins passed the minimal size restriction (twelve were removed since they were too small) and were further analyzed. Figure 7 shows the observed and expected word distributions of the age bins that passed the minimal size restriction (three age bins per lan- guage, other bins appear in Figure S6–S10). Importantly, there was a good fit between the observed frequency distribution and the expected Zipfian distribution for all age bins, reflected in a very high Pearson’s-r (see Table 2 and Table S1 for details). We wanted to further ask whether the distribution becomes “more Zipfian” as the child grows (has higher Pearson’s-r), and whether the slope of the distribution changes with age. To answer these questions, we ran two mixed-effect regression models. Following Barr et al. (2013), the models had the maximal random effect structure justified by the data that would converge. In the first model, our dependent variable was Pearson’s-r; age (in months, centred) was our fixed effect, and we had a random intercept for language as well as by-age random slope for the effect of language (Table 3). Age was defined as the minimal age per bin (e.g., for an age bin of 1;0–1;5 years, age for the regression was 12 months). While CDS follows a Zip- fian distribution from the very start, it does not become “more Zipfian” with age: Pearson’s-r did not change with age (β = 0.00004, SE = 0.00008, p > 0.6). This trend can be seen in
Figure 8A: Pearson’s-r is significantly high throughout development, without any visible
changement. The second model had the same fixed and random effects as the previous one, mais

OPEN MIND: Discoveries in Cognitive Science

17

Zipfian Distribution in CDS

Lavi-Rotbain and Arnon

je

D
o
w
n
o
un
d
e
d

F
r
o
m
h

t
t

p

:
/
/

d
je
r
e
c
t
.

m

je
t
.

/

e
d
toi
o
p
m

je
/

je

un
r
t
je
c
e

p
d

F
/

d
o

je
/

je

/

.

/

1
0
1
1
6
2
o
p
m
_
un
_
0
0
0
7
0
2
0
6
8
1
1
0
o
p
m
_
un
_
0
0
0
7
0
p
d

/

.

je

F

b
oui
g
toi
e
s
t

t

o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3

Chiffre 7. Observed (in green) and expected (in red) word distribution on a log-log scale. Three age bins per language are presented. All age
bins here passed the minimal size restriction of 50,000 tokens.

OPEN MIND: Discoveries in Cognitive Science

18

Zipfian Distribution in CDS

Lavi-Rotbain and Arnon

Tableau 3. Mixed-effect regression model looking at Pearson’s-r as a function of age. Variables in
bold were significant. Significance obtained using the lmerTest.

(Intercept)

Estimate
9.903e−01

Std. Error
3.631e−03

df

3.982e+00

t value
272.740

p-value
<.001*** Age (in months) 4.395e−05 8.249e−05 6.098e+00 0.533 >0.1

this time α was our dependent variable (Tableau 4). Here we saw a minor effect of age on α: Le
slope of the distribution became somewhat more moderate across development, as α
decreased with age (β = −0.0021, SE = 0.0007, p = 0.062, Figure 8B). While this effect did
not reach significance, it may indicate that speech directed to younger infants is more skewed
than speech directed to older children.

Étude 2 showed that CDS is skewed from the very start, and throughout development. Ce
skewed nature is seen in all of our samples, including the ones that did not reach the minimal
size requirement. En outre, while CDS does not become more Zipfian, its exponent
becomes mildly smaller with age, suggesting the contrast between high and low frequency
decreases somewhat with age (see Discussion). In the next study, we will ask if different parts
of speech also show such a skewed frequency distribution.

je

D
o
w
n
o
un
d
e
d

F
r
o
m
h

t
t

p

:
/
/

d
je
r
e
c
t
.

m

je
t
.

/

e
d
toi
o
p
m

je
/

je

un
r
t
je
c
e

p
d

F
/

d
o

je
/

je

.

/

/

1
0
1
1
6
2
o
p
m
_
un
_
0
0
0
7
0
2
0
6
8
1
1
0
o
p
m
_
un
_
0
0
0
7
0
p
d

/

.

je

F

b
oui
g
toi
e
s
t

t

o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3

Chiffre 8.

(UN) Pearson’s-r and (B) α as a function of age.

OPEN MIND: Discoveries in Cognitive Science

19

Zipfian Distribution in CDS

Lavi-Rotbain and Arnon

Tableau 4. Mixed-effect regression model looking at α as a function of age. Variables in bold were
significant. Significance obtained using the lmerTest.

(Intercept)

Estimate
1.3511334

Std. Error
0.0701455

df

3.9882635

Age (in months)

−0.0020710

0.0007186

3.0599781

t value
19.262

−2.882

p-value
<.001*** =0.062 STUDY 3: DOES CDS FOLLOW A ZIPFIAN DISTRIBUTION ACROSS PARTS OF SPEECH? In the third study, we ask whether different parts of speech in child-directed speech will also show a Zipfian distribution, as was previously found for adult directed speech (Piantadosi, 2014; Ramscar, 2020). We looked at the following five languages - British English, German, French, Spanish and Hebrew – because they all had large morphologically tagged corpora, and represent different language families. We looked at word distributions in these languages across four parts of speech: nouns, verbs, adjectives and prepositions. While the distribution of the first three parts of speech can be attributed, at least in part, to meaning (we talk about certain entities more than on others, or do certain actions more than others), it is hard to attribute a similar real-world meaning to prepositions, because they do not overlap in meaning across languages (e.g., Bowerman & Choi, 2001; Levinson et al., 2003), and because different prepositions can be used to describe the same event. Finding that prepositions also follow a Zipfian distribution will emphasize the role of pressures beside meaning – such as communicative efficiency or learnability – in the propen- sity of such skewed distributions in language (see Discussion for more on this). Methods We used all the available corpora in British English, German, French, Spanish and Hebrew. We used the morphological tagging provided in the corpora to automatically extract the relevant words (nouns, verbs, adjectives, and prepositions). However, we wanted to make sure the tagging in the corpora was accurate. To examine this, all the data was manually examined by research assistants, who were proficient speakers of the languages in question. The research assistants were asked to go over the extracted lists for each POS, and remove the words that did not belong to that part of speech. The words were not added to their correct part of speech list, since the part of speech for many words cannot be fully determined out of the context (i.e., “run” can be either a verb or a noun) and we did not want to introduce additional noise. After the manual cleaning, we calculated the error rate per language and per POS by dividing the number of types that were mistakenly tagged in each category by the total number of types in each category (e.g., 37 types erroneously classified as Spanish adjectives out of a total of 754 Spanish adjectives). Error rates were very high for prepositions across all languages, but much lower for nouns, verbs and adjec- tives (see Table 5 for full details). We do not know why the error rates are so high, this may reflect our automatic extraction of the tagged data (rather than errors in the tagged data itself ). We used the manually cleaned dataset to calculate the frequency distribution for each part of speech in each language and then did the same analyses as in the previous studies: (1) eval- uating α and β; (2) calculating Pearson’s-r between the observed and the expected frequency distributions; (3) plotting the two curves together. Results All four POS in the five languages had a skewed frequency distribution (Figure 9). To ask if this distribution is Zipfian, we estimated the parameters and compared the expected frequency OPEN MIND: Discoveries in Cognitive Science 20 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . / e d u o p m i / l a r t i c e - p d f / d o i / i . / / 1 0 1 1 6 2 o p m _ a _ 0 0 0 7 0 2 0 6 8 1 1 0 o p m _ a _ 0 0 0 7 0 p d . / i f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 O P E N M N D I : i D s c o v e r i e s i n C o g n i t i v e S c e n c e i German French Spanish Hebrew Summary Language British English Part of Speech nouns No. Tokens 1,088,319 No. Types 17,555 Word Frequency (1 – 17,274) Error Rate 0.084 Table 5. Summary of corpora measures across parts of speech for Study 3. verbs adjectives prepositions nouns verbs adjectives prepositions nouns verbs adjectives prepositions nouns verbs adjectives prepositions nouns verbs adjectives prepositions 796,976 226,907 341,910 3,978 1,987 74 116,944 10,951 56,399 28,388 31,724 169,214 189,203 28,517 69,489 39,155 42,059 6,320 12,672 43,294 27,161 6,915 23,537 2,660 1,259 38 7,305 3,844 1,349 46 3,446 2,556 716 17 2,543 2,705 671 172 (1 – 46,170) (1 – 19,925) (2 – 58,821) (1 – 13,086) (1 – 2,986) (1 – 2,628) (1 – 5,817) (1 – 3,347) (1 – 18,234) (1 – 4,240) (1 – 15,717) (1 – 1,099) (1 – 1,514) (1 – 426) (1 – 4552) (1 – 2,309) (1 – 1,038) (1 – 281) (1 – 3,519) 0.037 0.083 0.75 0.098 0.136 0.36 0.957 0.102 0.034 0.017 0.92 0.075 0.006 0.0491 0.77 0.25 0.067 0.203 0.687 α 1.53 1.74 1.6 8.59 0.9 1.54 1.58 49.77 1.35 1.46 1.42 β 75.52 9.93 7.13 33.97 −0.41 8.09 8.81 265.13 37.54 4.33 3.62 115.07 476.04 1.15 1.32 1.39 11.85 8.56 12.06 129.8 235.1 1.17 1.19 1.54 4.2 9.47 11.89 16.32 18.74 Pearson’s-r 0.926 0.992 0.962 0.985 0.934 0.992 0.962 0.986 0.947 0.995 0.927 0.974 0.991 0.946 0.921 0.966 0.995 0.986 – – nouns verbs Mean (SD) Mean (SD) adjectives Mean (SD) prepositions Mean (SD) 1.22 (0.24) 26.8 (30.6) 0.94 (0.02) 1.45 (0.21) 8.56 (2.79) 0.99 (0.01) 1.5 (0.1) 9.59 (4.84) 0.96 (0.025) 61.5 (58.6) 20.9 (25.1)* 205.8 (188.4) 105.9 (138.1)* 0.99 (0.002)* * Measured only on English, German and Hebrew. 2 1 Z i p f i a n D i s t r i b u t i o n i n C D S L a v i - R o t b a i n a n d A r n o n l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . / e d u o p m i / l a r t i c e - p d f / d o i / i / . / 1 0 1 1 6 2 o p m _ a _ 0 0 0 7 0 2 0 6 8 1 1 0 o p m _ a _ 0 0 0 7 0 p d . / i f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 Zipfian Distribution in CDS Lavi-Rotbain and Arnon l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . / e d u o p m i / l a r t i c e - p d f / d o i / i . / / 1 0 1 1 6 2 o p m _ a _ 0 0 0 7 0 2 0 6 8 1 1 0 o p m _ a _ 0 0 0 7 0 p d / . i f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 Figure 9. The observed distributions for four parts of speech across five languages. Note that Spanish prepositions had to little types to be analyzed further, but still shows a highly skewed distribution. OPEN MIND: Discoveries in Cognitive Science 22 Zipfian Distribution in CDS Lavi-Rotbain and Arnon with the observed one. For Spanish and French prepositions, the estimated α was very high (115 and 129.8 respectively) compared to all other estimated α (Table 5). As mentioned, α is the exponent of the distribution, with higher α for more skewed distributions. In this case, the high value of α leads to a binary expected distribution: The most frequent word was assigned a probability of one while the rest of the words were assigned a probability of zero. Therefore, we could not correlate the observed and the expected distributions for Spanish and French prepositions (so they are absent from Table 5). Beside these two datasets, all tested parts of speech, in all languages, showed a very good fit to the Zipfian distribution, as indicated by the high Pearson’s-r scores (Figures S11 in the Supplementary shows results on a log-log scale). Importantly, prepositions in English, German and Hebrew also showed a very good fit to the Zipfian distribution (mean = 0.99). Across parts of speech, α is relatively similar across languages (nouns: mean = 1.22, SD = 0.24; verbs: mean = 1.45, SD = 0.21; adjectives: mean = 1.5, SD = 0.09, Table 5). However, prepositions seem to have a higher α and larger SD compared to nouns, verbs and adjectives (mean = 61.4, SD = 58.6, for only English, German and Hebrew: mean = 20.9, SD = 25.1). This finding may be related to several unique properties of prepositions: prepositions have the smallest number of types. At the same time, high frequency prepositions are of the most frequent words in the language (looking only at the four categories used here, the most frequent word in each lan- guage was a preposition), while the minimum frequency for this group is the same as for the others (1–2 appearances). This means that frequency must decrease rapidly for prepositions, resulting in a higher α. However, it is important to note that the small number of preposition types might impair the reliability of the calculation of α: Our results clearly show that the dis- tribution of prepositions is indeed skewed (see Figure 9), but whether they follow a Zipfian distribution and what are its parameters need to be further examined. In sum, the examination of the distribution of word frequencies across four POS categories and in five languages shows a strong recurrence of near-Zipfian distributions, and highlights the prevalence of this kind of skew in children’s linguistic environment. DISCUSSION We set out to ask whether words in child-directed speech (CDS) follow a Zipfian distribution (Zipf, 1949). This distribution, where word frequencies follow a power law distribution, is one of the striking commonalities between languages, and has been found across languages (Piantadosi, 2014). However, almost all previous documentation of this phenomenon has focused on adult-to-adult speech, most often written language (Lestrade, 2017; Mehri & Jamaati, 2017; Moreno-Sánchez et al., 2016; Petersen et al., 2012; Piantadosi, 2014; Ramscar, 2020). If this distribution is a hallmark of language, and in particular, if it is driven, at least partially, by learnability pressures (Bentz, Alikaniotis, Cysouw, et al., 2017; Lavi-Rotbain & Arnon, 2022), it should also be found in the input children hear. At the same time, CDS has several unique lexical and structural characteristics, which could lead to a less skewed distribution. Importantly, no study to date has systematically examined word distributions in CDS across languages and parts of speech (but see: Dale & Spivey, 2006; Hendrickson & Perfors, 2019; Lavi-Rotbain & Arnon, 2022; Montag et al., 2018). Here, we expand on current findings to ask whether words follow a Zipfian distribution in CDS across languages, across development, and for different parts of speech. We assessed how Zipfian a distribution is by evaluating the parameters of the power law equation using the maximum-likelihood estimator, and computing the Pearson correlation between the observed and the expected frequency distribution. We find that words follow a Zipfian distribution in CDS across 15 languages from OPEN MIND: Discoveries in Cognitive Science 23 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . / e d u o p m i / l a r t i c e - p d f / d o i / i / . / 1 0 1 1 6 2 o p m _ a _ 0 0 0 7 0 2 0 6 8 1 1 0 o p m _ a _ 0 0 0 7 0 p d / . i f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 Zipfian Distribution in CDS Lavi-Rotbain and Arnon seven language families (Study 1); that they are Zipfian across development, even in speech directed to infants under the age of one (Study 2); and across different parts of speech (Study 3). What is the importance of these findings? First, these results support the generality of Zipfian distributions in language, showing it is found in another speech register besides adult-to-adult speech. Second, finding that different parts of speech, and specifically prepositions4, also follow a Zipfian distribution indicates that Zipfian distributions are not driven only by semantics. Finally, these results highlight the skewed nature of infants’ early learning environ- ments. Our findings join others in showing the skewness of infant’s environment (Clerkin et al., 2017; Lavi-Rotbain & Arnon, 2021, 2022). This skewness might have implications for learning, as suggested by recent studies showing improved learning in such distributions (e.g., Hendrickson & Perfors, 2019; Kurumada et al., 2013; Lavi-Rotbain & Arnon, 2020, 2021, 2022). This learnability advantage is also supported by findings showing that speakers turn uniform distributions into skewed ones in the process of cultural transmission, a pattern con- sistent with the presence of a cognitive bias for them (Shufaniya & Arnon, 2022). It has been shown that word segmentation is facilitated in language-like distributions. Since segmentation is one of the first challenges infants face when breaking into the speech stream, finding that these distributions are present when infants are segmenting speech strengthens their poten- tially facilitative role in the process. The results also show that α – which determines the slope of the distribution - spans a rel- atively narrow range (1.16–1.57) across languages, highlighting the similarity between them. This relatively narrow range of values raises the possibility that these values of α are more optimal than others for learning, and are therefore preserved across languages. The slope of the distribution dictates how fast frequency will decrease with rank: Steeper slopes will lead to a faster decrease resulting in a larger contrast between high and low frequencies. This might have implications for learning: It has been suggested that the slope found in natural languages is uniquely beneficial for language learning tasks, compared to more moderate slopes (Lavi- Rotbain & Arnon, 2022). At the same time, there is still consistent variation in α across and within language families. Previous studies examining bible translations also found variation across languages (Mehri & Jamaati, 2017; Piantadosi, 2014). Interestingly, it seems that α varies in a similar way across languages in CDS and in the bible translations (Figure 4), suggesting that some of the variation may reflect structural/linguistic differences between the languages. Further research is needed to know which linguistic properties affect α, and in what way. Methodologically, our findings highlight the importance of using learning environments similar to those of the real world. Despite the prevalence of Zipfian distributions in the wild, most lab-based explorations of word segmentation and learning use experimental paradigms where participants are exposed to uniform distributions, where each novel item appears the same number of times. For example, in the auditory statistical learning task used to assess word segmentation in the lab (Saffran et al., 1996), learners are exposed to novel words in a uniform distribution. While the use of uniform distributions allows researcher to control for frequency effects, they differ drastically from natural input and may lead to difficulties in learning. Inter- estingly, the presence and utility of skewed distributions is not limited to language: the objects infants see follow a Zipfian distribution (Clerkin et al., 2017), as do object combinations (Lavi- Rotbain & Arnon, 2021). Accordingly, children and adults are better at learning relations 4 While prepositions clearly follow a skewed distribution, the small number of types in the corpora we looked at makes our parameter estimations less reliable than the ones for nouns, verbs and adjectives. More data is needed to establish the fit to a Zipfian distribution. OPEN MIND: Discoveries in Cognitive Science 24 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . / e d u o p m i / l a r t i c e - p d f / d o i / i . / / 1 0 1 1 6 2 o p m _ a _ 0 0 0 7 0 2 0 6 8 1 1 0 o p m _ a _ 0 0 0 7 0 p d . / i f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 Zipfian Distribution in CDS Lavi-Rotbain and Arnon between objects (better visual statistical learning) when they are presented in a Zipfian distri- bution (Lavi-Rotbain & Arnon, 2021). These findings point to the importance of using learning environments that provide a more ecological reflection of children’s learning environment, and the need to further explore the impact of skewed distributions on learning. One of the limitations of the current study is how we defined what a word is. We defined a word by its orthographic form as is commonly done in investigations of word distributions (Cohen Priva & Gleason, 2016; Petersen et al., 2012; Piantadosi et al., 2011). This means that we treated words such as ‘walk’ and ‘walks’ as two unrelated items. In morphologically rich languages, such as German and Hebrew, this will result in a larger lexicon compared to less morphologically rich languages. For example, while English only has around five verb forms for each verb, Hebrew can have up to twenty separate items for a given verb. That is, defining words orthographically means increasing the number of types more for morphologically rich languages compared to less morphologically rich ones. Defining words in this way also doesn’t account for the semantic relatedness between word forms. The ecological validity of using orthographic form to define a word is supported by evidence that speakers do repre- sent inflected forms, and not only stems and morphemes, and that they are sensitive to whole- form frequency (Baayen et al., 1997). However, further work is needed to see whether the distribution shape or slope change across languages when a different word definition is used. Of particular interest is looking at lemmas to see whether this changes the distribution more for languages with rich morphology. The results presented here are limited to speech produced by adults: We did not look at the productions of the children themselves for both methodological and theoretical reasons. From a methodological perspective, children at the ages we examined simply do not provide enough linguistic output to reliably estimate the parameters of the distribution. This is espe- cially true for the younger ages where infants often produce only single words. Moreover, care- givers who speak the same language share much of their lexicon, making it is reasonable to combine different conversations into one larger database. However, this may not hold for the younger ages: children who have just begun to speak produce only few words, and these do not always overlap between different infants, making it problematic to collapse frequency counts over different infants. There are also theoretical reasons not to collapse the speech of children at different ages: The kind of language produced by a two-year old (mostly content words: Caselli et al., 1995; Goodman et al., 2008) is very different from that produced by a five-year-old making it hard to understand the patterns of an aggregated distribution. To under- stand whether the language produced by individual learners is Zipfian and when it becomes so, two interesting and open questions, would require conducting longitudinal analyses of dense individual corpora. In sum, we showed in this paper that word distributions in CDS are Zipfian across lan- guages, across parts of speech, and from early on. These findings support the generality of Zipf’s law in language and highlight the skewed nature of infant’s early linguistic environment. While many questions remain open, these results strengthen the suggestion that such skewed distribution may be relevant for learning. ACKNOWLEDGMENTS We thank Zohar Aizenbud, Hila Merha and, Maya Ravid for help with extracting and cleaning the data. We thank Shira Tal for her helpful comments. We thank all researchers who made their corpora available through CHILDES. The research was funded by the Israeli Science Foundation grant number 584/16 and grant number 445/20 awarded to the second author. OPEN MIND: Discoveries in Cognitive Science 25 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . / e d u o p m i / l a r t i c e - p d f / d o i / i / . / 1 0 1 1 6 2 o p m _ a _ 0 0 0 7 0 2 0 6 8 1 1 0 o p m _ a _ 0 0 0 7 0 p d . / i f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 Zipfian Distribution in CDS Lavi-Rotbain and Arnon REFERENCES Altvater-Mackensen, N., & Mani, N. (2013). Word-form familiarity bootstraps infant speech segmentation. Developmental Science, 16(6), 980–990. https://doi.org/10.1111/desc.12071, PubMed: 24118722 Arnold, R. A. (2015). Changes in the coefficients of Zipf’s law for English corpora of different contexts. University of Nevada, Reno. Baayen, R. H., Dijkstra, T., & Schreuder, R. (1997). Singulars and plurals in Dutch: Evidence for a parallel dual-route model. Jour- nal of Memory and Language, 37(1), 94–117. https://doi.org/10 .1006/jmla.1997.2509 Barr, D. J., Levy, R., Scheepers, C., & Tily, H. J. (2013). Random effects structure for confirmatory hypothesis testing: Keep it max- imal. Journal of Memory and Language, 68(3), 255–278. https:// doi.org/10.1016/j.jml.2012.11.001, PubMed: 24403724 Bentz, C., Alikaniotis, D., Cysouw, M., & Ferrer-i-Cancho, R. (2017). The entropy of words—Learnability and expressivity across more than 1000 languages. Entropy, 19(6), 275. https:// doi.org/10.3390/e19060275 Bentz, C., Alikaniotis, D., Samardžić, T., & Buttery, P. (2017). Vari- ation in word frequency distributions: Definitions, measures and implications for a corpus-based language typology. Journal of Quantitative Linguistics, 24(2–3), 128–162. https://doi.org/10 .1080/09296174.2016.1265792 Bortfeld, H., Morgan, J. L., Golinkoff, R. M., & Rathbun, K. (2005). Mommy and me: Familiar names help launch babies into speech- stream segmentation. Psychological Science, 16(4), 298–304. https://doi.org/10.1111/j.0956-7976.2005.01531.x, PubMed: 15828977 Bowerman, M., & Choi, S. (2001). Shaping meanings for language: Universal and language-specific in the acquisition of spatial semantic categories. In Language acquisition and conceptual development (pp. 475–511). Cambridge University Press. https://doi.org/10.1017/CBO9780511620669.018 Brodsky, P., Waterfall, H., & Eldermen, S. (2007). Characterizing motherese: On the computational structure of child-directed language. In Proceedings of the 29th Cognitive Science Society Conference (pp. 833–838). Cognitive Science Society. Cameron-Faulkner, T., Lieven, E., & Tomasello, M. (2003). A construction based analysis of child directed speech. Cognitive Sci- ence, 27(6), 843–873. https://doi.org/10.1207/s15516709cog2706_2 Caselli, M. C., Bates, E., Casadio, P., Fenson, J., Fenson, L., Sanderl, L., & Weir, J. (1995). A cross-linguistic study of early lexical development. Cognitive Development, 10(2), 159–199. https:// doi.org/10.1016/0885-2014(95)90008-X Chater, N., & Brown, G. D. A. (1999). Scale-invariance as a unifying psychological principle. Cognition, 69(3), 17–24. https://doi.org /10.1016/S0010-0277(98)00066-3, PubMed: 10193053 Christiansen, M. H., & Chater, N. (2008). Language as shaped by the brain. The Behavioral and Brain Sciences, 31(5), 489–508; discus- sion 509–558. https://doi.org/10.1017/S0140525X08004998, PubMed: 18826669 Clauset, A., Shalizi, C. R., & Newman, M. E. J. (2009). Power-law distributions in empirical data. SIAM Review, 51(4), 661–703. https://doi.org/10.1137/070710111 Clerkin, E. M., Hart, E., Rehg, J. M., Yu, C., & Smith, L. B. (2017). Real-world visual statistics and infants’ first-learned object names. Philosophical Transactions of the Royal Society B: Biolog- ical Sciences, 372(1711), 20160055. https://doi.org/10.1098/rstb .2016.0055, PubMed: 27872373 Cohen Priva, U., & Gleason, E. (2016). Simpler structure for more informative words: A longitudinal study. In A. Papafragou, D. Grodner, D. Mirman, & J. Trueswell (Eds.), Proceedings of the the Cognitive Science Society 38th Annual Conference of (pp. 1895–1900). Cognitive Science Society. Coupé, C., Oh, Y., Dediu, D., & Pellegrino, F. (2019). Different languages, similar encoding efficiency: Comparable information rates across the human communicative niche. Science Advances, 5(9), eaaw2594. https://doi.org/10.1126/sciadv.aaw2594, PubMed: 32047854 Dale, R., & Spivey, M. J. (2006). Unraveling the dyad: Using recur- rence analysis to explore patterns of syntactic coordination between children and caregivers in conversation. Language Learning, 56(3), 391–430. https://doi.org/10.1111/j.1467-9922 .2006.00372.x Du Bois, J. W., Chafe, W. L., Meyer, C., Thompson, S. A., & Martey, N. (2000). Santa Barbara corpus of spoken American English [CD-ROM]. Linguistic Data Consortium. Fernald, A., & Simon, T. (1984). Expanded intonation contours in mothers’ speech to newborns. Developmental Psychology, 20(1), 104–113. https://doi.org/10.1037/0012-1649.20.1.104 Ferrer-i-Cancho, R. (2005). The variation of Zipf’s law in human language. European Physical Journal B, 44(2), 249–257. https:// doi.org/10.1140/epjb/e2005-00121-8 Ferrer-i-Cancho, R. (2016). Compression and the origins of Zipf’s law for word frequencies. Complexity, 21(S2), 409–411. https:// doi.org/10.1002/cplx.21820 Ferrer-i-Cancho, R., Bentz, C., & Seguin, C. (2020). Optimal coding and the origins of Zipfian laws. Journal of Quantitative Linguis- tics, 29(2), 165–194. https://doi.org/10.1080/09296174.2020 .1778387 Ferrer-i-Cancho, R., & Sole, R. V. (2003). Least effort and the origins of scaling in human language. Proceedings of the National Acad- emy of Sciences, 100(3), 788–791. https://doi.org/10.1073/pnas .0335980100, PubMed: 12540826 Geertzen, J., Blevins, J. P., & Milin, P. (2016). The informativeness of linguistic unit boundaries. Italian Journal of Linguistics, 28(1), 25–48. https://doi.org/10.17863/CAM.69 Gibson, E., Futrell, R., Piantadosi, S. T., Dautriche, I., Bergen, L., & Levy, R. (2019). How efficiency shapes human language. Trends in Cognitive Sciences, 23(5), 389–407. https://doi.org/10.1016/j .tics.2019.02.003, PubMed: 31006626 Gillespie, C. S. (2015). Fitting heavy tailed distributions: The poweRlaw package. Journal of Statistical Software, 64(2), 1–16. https://doi.org/10.18637/jss.v064.i02 Goldberg, A. E., Casenhiser, D. M., & Sethuraman, N. (2004). Learning argument structure generalizations. Cognitive Linguis- tics, 15(3), 289–316. https://doi.org/10.1515/cogl.2004.011 Goodman, J. C., Dale, P. S., & Li, P. (2008). Does frequency count? Parental input and the acquisition of vocabulary. Journal of Child L a n g u a g e , 3 5 ( 3 ) , 5 1 5 – 5 3 1 . h t t p s : / / d o i . o r g / 1 0 . 1 0 1 7 /S0305000907008641, PubMed: 18588713 Hendrickson, A. T., & Perfors, A. (2019). Cross-situational learning in a Zipfian environment, Cognition, 189, 11–22. https://doi.org /10.1016/j.cognition.2019.03.005, PubMed: 30903853 Kurumada, C., Meylan, S. C., & Frank, M. C. (2013). Zipfian fre- quency distributions facilitate word segmentation in context. Cognition, 127(3), 439–453. https://doi.org/10.1016/j.cognition .2013.02.002, PubMed: 23558340 Lavi-Rotbain, O., & Arnon, I. (2019). Children learn words better in low entropy. In Proceedings of the 41th Annual Conference of the Cognitive Science Society (pp. 631–637). Cognitive Science Society. OPEN MIND: Discoveries in Cognitive Science 26 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . / e d u o p m i / l a r t i c e - p d f / d o i / i / / . 1 0 1 1 6 2 o p m _ a _ 0 0 0 7 0 2 0 6 8 1 1 0 o p m _ a _ 0 0 0 7 0 p d / . i f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 Zipfian Distribution in CDS Lavi-Rotbain and Arnon Lavi-Rotbain, O., & Arnon, I. (2020). The learnability consequences of Zipfian distributions: Word segmentation is facilitated in more predictable distributions. PsyArXiv. https://doi.org/10.31234/osf .io/xwgpk Lavi-Rotbain, O., & Arnon, I. (2021). Visual statistical learning is facil- itated in Zipfian distributions. Cognition, 206, 104492. https://doi .org/10.1016/j.cognition.2020.104492, PubMed: 33157380 Lavi-Rotbain, O., & Arnon, I. (2022). The learnability consequences of Zipfian distributions in language. Cognition, 223, 105038. https:// doi.org/10.1016/j.cognition.2022.105038, PubMed: 35123219 Lestrade, S. (2017). Unzipping Zipf’s law. PLOS ONE, 12(8), e0181987. https://doi.org/10.1371/journal.pone.0181987, PubMed: 28792963 Levinson, S., Meira, S., Bohnemeyer, J., Terrill, A., Guirardello, R., Enfield, N., … De Witte, C. (2003). “Natural concepts” in the spa- tial topological domain-adpositional meanings in crosslinguistic perspective: An exercise in semantic typology. Language, 79(3), 485–516. https://doi.org/10.1353/lan.2003.0174 Linders, G. M., & Louwerse, M. M. (2020). Zipf’s law in human-machine dialog. In Proceedings of the 20th ACM Interna- tional Conference on Intelligent Virtual Agents (pp. 1–8). Associ- ation for Computing Machinery. https://doi.org/10.1145 /3383652.3423878 MacWhinney, B. (2000). The CHILDES Project: Tools for analyzing talk (3rd ed.). Lawrence Erlbaum Associates. Mandelbrot, B. (1953). An informational theory of the statistical structure of language. In W. Jackson (Ed.), Communication theory (pp. 486–502). Academic Press. Manin, D. (2008). Zipf’s law and avoidance of excessive synon- ymy. Cognitive Science, 32(7), 1075–1098. https://doi.org/10 .1080/03640210802020003, PubMed: 21585444 Mehri, A., & Jamaati, M. (2017). Variation of Zipf’s exponent in one hundred live languages: A study of the Holy Bible translations. Physics Letters, Section A: General, Atomic and Solid State Phys- ics, 381(31), 2470–2477. https://doi.org/10.1016/j.physleta.2017 .05.061 Mintz, T. H. (2003). Frequent frames as a cue for grammatical cat- egories in child directed speech. Cognition, 90(1), 91–117. https://doi.org/10.1016/S0010-0277(03)00140-9, PubMed: 14597271 Montag, J. L., Jones, M. N., & Smith, L. B. (2018). Quantity and diversity: Simulating early word learning environments. Cogni- tive Science, 42, 375–412. https://doi.org/10.1111/cogs.12592, PubMed: 29411899 Montemurro, M. A. (2001). Beyond the Zipf–Mandelbrot law in quantitative linguistics. Physica A: Statistical Mechanics and Its Applications, 300(3–4), 567–578. https://doi.org/10.1016/S0378 -4371(01)00355-7 Moreno-Sánchez, I., Font-Clos, F., & Corral, Á. (2016). Large-scale analysis of Zipf’s law in English texts. PLOS ONE, 11(1), e0147073. https://doi.org/10.1371/journal.pone.0147073, PubMed: 26800025 Newman, M. E. J. (2005). Power laws, Pareto distributions and Zipf’s law. Contemporary Physics, 46(5), 323–351. https://doi .org/10.1080/00107510500052444 Petersen, A. M., Tenenbaum, J. N., Havlin, S., Stanley, H. E., & Perc, M. (2012). Languages cool as they expand: Allometric scal- ing and the decreasing need for new words. Scientific Reports, 2, 943. https://doi.org/10.1038/srep00943, PubMed: 23230508 Piantadosi, S. T. (2014). Zipf’s word frequency law in natural lan- guage: A critical review and future directions. Psychonomic Bulletin & Review, 21(5), 1112–1130. https://doi.org/10.3758 /s13423-014-0585-6, PubMed: 24664880 Piantadosi, S. T., Tily, H., & Gibson, E. (2011). Word lengths are optimized for efficient communication. Proceedings of the National Academy of Sciences of the United States of America, 108(9), 3526–3529. https://doi.org/10.1073/pnas.1012551108, PubMed: 21278332 Ramscar, M. (2020). The empirical structure of word frequency dis- tributions. arXiv:2001.05292. https://doi.org/10.48550/arXiv .2001.05292 R Core Team. (2021). R: A language and environment for statistical computing. R Foundation for Statistical Computing. Retrieved June 26, 2020 from https://www.r-project.org/ Reggev, N., Sharoni, R., & Maril, A. (2018). Distinctiveness benefits novelty (and not familiarity), but only up to a limit: The prior knowledge perspective. Cognitive Science, 42(1), 103–128. https://doi.org/10.1111/cogs.12498, PubMed: 28503806 Reppen, R., & Ide, N. (2004). The American National Corpus: Overall goals and the first release. Journal of English Linguistics, 32(2), 105–113. https://doi.org/10.1177/0075424204264856 Roy, B. C., Frank, M. C., DeCamp, P., Miller, M., & Roy, D. (2015). Predicting the birth of a spoken word. Proceedings of the National Academy of Sciences, 112(41), 12663–12668. https:// doi.org/10.1073/pnas.1419773112, PubMed: 26392523 Roy, B. C., Frank, M. C., & Roy, D. (2009). Exploring word learning in a high-density longitudinal corpus. In Proceedings of the 31st Annual Meeting of the Cognitive Science Society (pp. 2106–2111). Cognitive Science Society. Rubin, J., Ulanovsky, N., Nelken, I., & Tishby, N. (2016). The rep- resentation of prediction error in auditory cortex. PLOS Compu- tational Biology, 12(8), e1005058. https://doi.org/10.1371 /journal.pcbi.1005058, PubMed: 27490251 Saffran, J. R., Aslin, R. N., & Newport, E. L. (1996). Statistical learn- ing by 8-month-old infants. Science, 274(5294), 1926–1928. https://doi.org/10.1126/science.274.5294.1926, PubMed: 8943209 Schuler, K. D., Reeder, P. A., Newport, E. L., & Aslin, R. N. (2017). The effect of Zipfian frequency variations on category formation in adult artificial language learning. Language Learning and Development, 13(4), 357–374. https://doi.org/10.1080 /15475441.2016.1263571, PubMed: 30405323 Semple, S., Ferrer-i-Cancho, R., & Gustison, M. L. (2022). Lin- guistic laws in biology. Trends in Ecology and Evolution, 37(1), 53–66. https://doi.org/10.1016/j.tree.2021.08.012, PubMed: 34598817 Shannon, C. E. (1948). A mathematical theory of communication. Bell System Technical Journal, 27, 379–423. https://doi.org/10 .1002/j.1538-7305.1948.tb01338.x Shufaniya, A., & Arnon, I. (2022). A cognitive bias for Zipfian distributions? Uniform distributions become more skewed via cultural transmission. Journal of Language Evolution, 7(1), 59–80. https://doi.org/10.1093/jole/lzac005 Tal, S., & Arnon, I. (2018). SES effects on the use of variation sets in child-directed speech. Journal of Child Language, 45(6), 1423–1438. https://doi.org/10.1017/S0305000918000223, PubMed: 29974833 Yang, C. (2013). Who’s afraid of George Kingsley Zipf? Or: Do chil- dren and chimps have language? Significance, 10(6), 29–34. https://doi.org/10.1111/j.1740-9713.2013.00708.x Zipf, G. K. (1949). Human behavior and the principle of least effort. Addison-Wesley Press. OPEN MIND: Discoveries in Cognitive Science 27 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . / e d u o p m i / l a r t i c e - p d f / d o i / i / / . 1 0 1 1 6 2 o p m _ a _ 0 0 0 7 0 2 0 6 8 1 1 0 o p m _ a _ 0 0 0 7 0 p d / . i f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 Zipfian Distribution in CDS Lavi-Rotbain and Arnon APPENDIX A: COMPARING PARAMETERS ESTIMATED USING THE ENTIRE CORPUS AND THE HALF-SPLIT METHOD FOR TEN LANGUAGES To ensure that our use of the entire corpus to estimate the parameters did not impact the results, we applied the half-split method (Piantadosi, 2014) to the ten languages where we had enough data. We first used a binomial split to cut the data into half: the first half was used to estimate word frequency and the second half was used to estimate its rank, to reduce the dependency between the two measures. The binomial split works as follows: For each word, we sample from a binomial the same number of times as its frequency (with a probability for success – p – equals to 0.5) and receive a number (close to half of the overall frequency but not equal to it) that will now be used as the word’s frequency. The difference between the original frequency and the new one is used to assign each word its rank. For example, the word “up” in British English appeared 23,251 times. If we would assign it a rank based on this frequency, it would receive r = 55. However, the frequency after the binomial split was 11,731. The fre- quency for the rank estimation is the difference between the two (23251 − 11731= 11520), which gave it rank 57. The ten languages we analysed are the ones with at least 200,000 tokens so that each half of the sample has approximately more than 100,000 tokens (depends on the way the data was split). These were: North-American and British English, German, French, Dutch, Japanese, Spanish, Swedish, Portuguese and Hebrew. Table A1 shows the estimated parameters for each language using the full corpus and the split-half method. Importantly, the parameters estimated by the two methods – with and without binomial split – are highly correlated (Pearson’s-r for α = 0.999; Pearson’s-r for β = 0.999) with α values almost identical between the two methods. β values also show high similarity. This analysis validates our use of the entire corpus to estimate the parameters. Table A1. Summary of corpora measures across languages with and without binomial split. Without binomial split (taken from Study 1) α β Pearson’s-r 0.97 Language British English North-American English German French Dutch Japanese Spanish Swedish Portuguese Hebrew α 1.56 1.52 1.41 1.52 1.47 1.29 1.37 1.38 1.35 1.17 With binomial split β 19.21 18.04 11.61 17.47 11.98 6.73 9.34 7.17 5.01 3.27 Pearson’s-r 0.97 0.965 0.997 0.99 0.998 0.993 0.994 0.997 0.983 0.996 1.57 1.52 1.42 1.53 1.48 1.30 1.39 1.40 1.37 1.19 19.48 18.34 12.69 17.98 12.63 7.12 9.96 7.58 5.44 3.69 0.965 0.998 0.99 0.998 0.993 0.994 0.998 0.981 0.996 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . / e d u o p m i / l a r t i c e - p d f / d o i / i / . / 1 0 1 1 6 2 o p m _ a _ 0 0 0 7 0 2 0 6 8 1 1 0 o p m _ a _ 0 0 0 7 0 p d . / i f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 OPEN MIND: Discoveries in Cognitive Science 28 Zipfian Distribution in CDS Lavi-Rotbain and Arnon APPENDIX B: EXAMINING THE STABILITY OF THE CLAUSET METHOD We used the ‘poweRlaw’ library (Gillespie, 2015) in R (R Core Team, 2021) to assess the fit of our child-directed corpora and of several adult corpora (ADS) to a true power law. We found that the p-value determining if the data was taken from a power-law distribution was not con- sistent across analyses in ways that do not seem to reflect meaningful variation in the data. In a nutshell, using 100 or 1000 bootstrap iterations, about half of the languages in our sample had a p < .05, meaning they deviate from a power-law distribution but whether they deviated did not correspond to our other estimates of the fit to a Zipfian distribution and was not consistent even within the same language (e.g., British English deviated but North-American English did not). To see if the presence of deviating languages reflects a true difference in word distributions between CDS and ADS, and to further explore the seeming inconsistencies the method gen- erates, we also analyzed two American-English adult speech corpora using the Clauset method (The Santa Barbra corpus; Du Bois et al., 2000) and CABank English CallFriend Northern US Corpus from CHILDES (MacWhinney, 2000), see details below). Applying this method to two comparable adult speech corpora of the same language also led to inconsistent results with one corpus drawn from a power-law (CallFriend from CHILDES, p = 0.55) while the second corpus not (Santa Barbara, p = 0.07). In contrast, using the method we used, where the param- eters are estimated and the fit between the expected and observed frequency distribution is calculated, all three American-English corpora (the child-directed one and the two adult ones) were found to show a very good fit to a Zipfian distribution (CDS: alpha = 1.52; beta = 18.34; r = 0.965; Santa Barbara: alpha = 1.36; beta = 9.23; r = 0.99; CallFriend: alpha = 1.35, beta = 11.5, r = 0.99). Finally, to further probe the possible inconsistencies generated by this method, we applied the Clauset method to the smaller samples we used to assess alpha for each language (100,000 tokens each, nine languages, see Study 1 in page 18). In Study 1, we used these samples to investigate the stability of our alpha estimations, and to ensure that this corpus size is sufficient to obtain a reliable estimate of alpha. We now wanted to see whether different samples of the same corpus are consistently identified as being drawn (or not drawn) from a power-law dis- tribution. While the estimated alpha was relatively similar (very small SD) across different l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . / e d u o p m i / l a r t i c e - p d f / d o i / i / / . 1 0 1 1 6 2 o p m _ a _ 0 0 0 7 0 2 0 6 8 1 1 0 o p m _ a _ 0 0 0 7 0 p d / . i Table B1. Summary of application of Clauset’s method using the poweRlaw package on the full corpora for each language. Language British English Dutch French German Hebrew Japanese Portuguese Spanish Swedish Number of sub-samples receiving p-value > 0.05
37

Number of sub-samples
receiving p-value =< 0.05 32 Measures of alpha across sub-samples (mean (SD)) 1.769 (0.064) 3 15 6 3 7 2 5 3 8 4 15 0 7 1 1 0 1.703 (0.039) 1.748 (0.052) 1.69 (0.067) 1.937 (0.04) 1.769 (0.084) 1.717 (0.059) 1.797 (0.075) 1.793 (0.012) f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 OPEN MIND: Discoveries in Cognitive Science 29 Zipfian Distribution in CDS Lavi-Rotbain and Arnon samples of the same language, the p-value using the Clauset method was not (it resulted in different results for different samples from the same language). For example, in British-English 32 samples out of 69 received p <= 0.05, and 37 samples received p > 0.05. A similar pattern
was found for the other languages we looked at (see Table B1).

Future work is needed to understand why the method generates such inconclusive results
for linguistic data (which is beyond the scope of the current paper). Since our results are not
dependent on the distribution being a ‘true’ power law, we did not pursue this method of
estimation further.

je

D
o
w
n
o
un
d
e
d

F
r
o
m
h

t
t

p

:
/
/

d
je
r
e
c
t
.

m

je
t
.

/

e
d
toi
o
p
m

je
/

je

un
r
t
je
c
e

p
d

F
/

d
o

je
/

je

/

/

.

1
0
1
1
6
2
o
p
m
_
un
_
0
0
0
7
0
2
0
6
8
1
1
0
o
p
m
_
un
_
0
0
0
7
0
p
d

.

/

je

F

b
oui
g
toi
e
s
t

t

o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3

OPEN MIND: Discoveries in Cognitive Science

30REPORT image
REPORT image
REPORT image
REPORT image
REPORT image
REPORT image
REPORT image
REPORT image
REPORT image
REPORT image
REPORT image
REPORT image

Télécharger le PDF