REPORT
Most People Keep Their Word
Rather Than Their Money
Jan K. Woike
1 and Patricia Kanngiesser
2
1Center for Adaptive Rationality (ARC), Max Planck Institute for Human Development, Berlin, Allemagne
2Faculty of Education and Psychology, Freie UniversitÄt Berlin, Berlin, Allemagne
un accès ouvert
journal
Mots clés: cooperation, social norms, promises, agreements, contracts
je
D
o
w
n
o
un
d
e
d
F
r
o
m
h
t
t
p
:
/
/
d
je
r
e
c
t
.
m
je
t
.
/
e
d
toi
o
p
m
je
/
je
un
r
t
je
c
e
–
p
d
F
/
d
o
je
/
je
/
/
.
1
0
1
1
6
2
o
p
m
_
un
_
0
0
0
2
7
2
0
6
6
9
7
6
o
p
m
_
un
_
0
0
0
2
7
p
d
/
.
je
ABSTRAIT
Promises are crucial for human cooperation because they allow people to enter into
voluntary commitments about future behavior. Here we present a novel, fully incentivized
paradigm to measure voluntary and costly promise-keeping in the absence of external
sanctions. We found across three studies (N = 4, 453) that the majority of participants
(61%–98%) kept their promises to pay back a specified amount of a monetary endowment,
and most justified their decisions by referring to obligations and norms. Varying promise
elicitation methods (Study 1a) and manipulating stake sizes (Study 2a) had negligible effects.
Simultaneously, when others estimated promise-keeping rates (using two different estimation
méthodes), they systematically underestimated promise-keeping by up to 40% (Studies 1b and
2b). Additional robustness checks to reduce potential reputational concerns and possible
demand effects revealed that the majority of people still kept their word (Étude 3). Promises
have a strong normative power and binding effect on behavior. Nevertheless, people appear
to pessimistically underestimate the power of others’ promises. This behavior–estimation gap
may prevent efficient coordination and cooperation.
F
b
oui
g
toi
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
INTRODUCTION
Large-scale human cooperation depends on commitments, made in the present and to be
honored in the future. Promises create such commitments: By simply saying “I promise to
do X” (or an equivalent phrase), under the appropriate conditions one is obligated to do X
(Austin, 1975; Searle, 1969). Some have argued that promises without threats of sanctions are
just empty words with no binding power (Hobbes, 2016). Yet research has shown that promises
and verbal agreements increase cooperation rates in social dilemma games (Bicchieri, 2002;
Charness & Dufwenberg, 2006; Ellingsen & Johannesson, 2004; Orbell, van de Kragt, &
Dawes, 1988; Ostrom, Walker, & Gardner, 1992; Sally, 1995; van den Assem, van Dolder,
& Thaler, 2012) and are more effective than punishment or threats (Ellingsen & Johannesson,
2004; Ostrom et al., 1992).
While the effect of communication on cooperation is well established, ce n'est pas clair
whether promises in social dilemmas are effective because of the act of promising or be-
cause people use promises to signal their intentions to other players when given the oppor-
tunity to communicate (Ismayilov & Potters, 2016; van den Assem et al., 2012). C'est, dans
social dilemmas, promises may be used strategically to gain a partner’s trust and cooperation
and to reduce uncertainty about future behavior. There have been some recent attempts to
Citation: Woike, J.. K., & Kanngiesser,
P.. (2019). Most People Keep Their Word
Rather Than Their Money. Open Mind:
Discoveries in Cognitive Science,
3, 68–88. https://est ce que je.org/10.1162/
opmi_a_00027
EST CE QUE JE:
https://doi.org/10.1162/opmi_a_00027
Supplemental Materials:
https://doi.org/10.1162/opmi_a_00027;
https://doi.org/10.17605/OSF.IO/DXEPU
Reçu: 3 Octobre 2018
Accepté: 14 May 2019
Intérêts concurrents: Les auteurs ont
a déclaré qu'aucun intérêt concurrent
exists.
Auteur correspondant:
Jan K. Woike
woike@mpib-berlin.mpg.de
droits d'auteur: © 2019
Massachusetts Institute of Technology
Publié sous Creative Commons
Attribution 4.0 International
(CC PAR 4.0) Licence
La presse du MIT
Most People Keep Their Word Woike and Kanngiesser
measure promise-keeping with paradigms that did not involve strategic interactions: For ex-
ample, Mischkowski, Stone, and Stremitzer (2016) used vignettes to ask participants about the
likelihood of buying a product from a seller after a promise and Conrads and Reggiani (2017)
elicited nonincentivized promises to take part in a survey. Cependant, neither promise-keeping
nor promise-breaking in these studies resulted in a cost for participants; it remains unclear
how people will react when keeping a promise comes at a monetary cost. Ici, we propose a
novel, fully incentivized minimal promising paradigm, in which promises are clearly defined
and consequential and can be studied outside of strategic interactions.
In a series of experiments (Studies 1a, 2un, et 3), workers in an online labor market
could choose between (je) a small sum of money (par exemple., $0.50; henceforth, all amounts in USD) ou (ii) a larger sum of money (par exemple., $2) under the condition that they make a commitment to pay
back a fixed sum (par exemple., $1) at a later point. Choosing the commitment option (with or without paying back the fixed amount) thus always yielded a higher payoff than the no-commitment option. These incentives were set so that payoffs for cooperators dominated payoffs for non- commitment, giving even potentially uncooperative and untrustworthy participants a reason to make a promise (and an incentive to break it). We ensured that participants could decide and were incentivized to enter the commitment but were not forced to do so (see Kataria & Hiver, 2013; Shu, Mazar, Gino, Ariely, & Bazerman, 2012, for forced oaths) for two reasons: (je) because promising is by definition a voluntary act (Searle, 1969), et (ii) to respect indi- vidual autonomy and decision making (Hertwig & Grüne-Yanoff, 2017). This paradigm has several unique features: In contrast to the free-form communication used in previous studies, our promises were clearly marked as speech acts of commitment (par exemple., they contained the word “promise”) and it was unambiguous for both promisor and promisee that a promise had been made. En outre, our paradigm did not involve any uncertainty about the consequence of the promise and its payoffs. To get insight into people’s motives for keeping their promises (Bhattacharya & Sengupta, 2016; Charness & Dufwenberg, 2006, 2010; Di Bartolomeo, Dufwenberg, Papa, & Passarelli, 2018; Ederer & Stremitzer, 2017; Ellingsen & Johannesson, 2004; Mischkowski et al., 2016; Vanberg, 2008) and their experience of the process, we asked different groups of participants in Studies 1a and 2a to justify their decisions and assessed their positive and negative affect using the Positive and Negative Affect Schedule (PANAS) scale (Crawford & Henry, 2004). De plus, to find out whether promise-keeping behavior would be accurately predicted (Belot, Bhaskar, & van de Ven, 2012; Chen & Houser, 2017), we asked new groups of participants to retrodict promise-keeping rates in our paradigm using two different incentivized methods (Studies 1b and 2b). STUDY 1A Methods All participants were located in the United States and recruited through Participants Amazon’s Mechanical Turk (MTurk). Qualifications and exclusion criteria (double IP addresses, failed attention checks, etc.) for all studies are described in the Supplemental Material (Woike & Kanngiesser, 2019b). Double participation was prevented by a combination of using the Unique Turker script, checking against exclusion lists at the beginning of the survey, and car- rying out manual checks of MTurk IDs and IP addresses. After these steps, there were n1a = 568 participants in Study 1a (Mage = 36.1, 55.6% female, 44.4% male). All participants gave their informed consent and the study was approved by the Institutional Review Board (CISR) at the Max Planck Institute for Human Development. OPEN MIND: Discoveries in Cognitive Science 69 l Téléchargé à partir du site Web : / / direct . m je t . / e d u o p m i / l a r t i c e – pdf / d o i / je . / / 1 0 1 1 6 2 o p m _ a _ 0 0 0 2 7 2 0 6 6 9 7 6 o p m _ a _ 0 0 0 2 7 pd / . i f b y g u e s t t o n 0 7 Septembre 2 0 2 3 Most People Keep Their Word Woike and Kanngiesser Procedure Workers were offered a choice between (un) $0.05 et (b) $0.15 under the con- dition that they commit to paying back $0.05 (henceforth 15–5). Since there have been few
systematic studies on the effect of different commitment formats on promise-keeping (Charness
& Dufwenberg, 2010; Turmunkh, van den Assem, & van Dolder, 2019), we decided to address
this question by implementing the following commitment variants in a between-subjects de-
sign: (1) promise (write): clicking on a radio button with the promise spelled out and typing “I
promise” into a text box next to it (n = 147), (2) promise (click): only clicking on a radio button
to promise (n = 146), et (3) ask: being asked to pay back a suggested amount of money with
no mention of the word “promise” (n = 145). We also included (4) a noncommitment control
condition, in which participants had the option to select (un) $0.05 ou (b) $0.15 with the choice
to pay back $0.05 or a different amount (n = 130). Workers received the final amount (c'est à dire., the chosen amount minus any money they decided to return) as part of their bonus payment a few days after the study had finished. To illustrate the wording we used when asking participants to choose between the lower and the higher payoff, we take the promise (click) condition as an example (for the wording in the other conditions, see the Supplemental Materials, Woike & Kanngiesser, 2019b): “You have two options that have consequences for your bonus payment: 1. You can receive 5 cents without any further consequences. 2. You can receive 15 cents. In this case we would ask you to promise that you will give back 5 cents at the end of the survey. (We would ask you to pay back this money at the end).” Participants were then offered two options and had to choose one using a click button: “I take 5 cents” or “I take 15 cents and I promise to pay back 5 cents when asked to at the end of the survey.” Participants then completed a series of unrelated tasks. After the tasks, those who had accepted $0.15 were reminded of their choice (in two conditions, also of their promise) et
given a chance to pay back $0.00–$0.10. We decided to remind participants of their choice or
promise to ensure that a decision against paying back money was not due to memory errors.
De plus, participants were informed that there would be no negative consequences no matter
which amount they chose. In the promise conditions, we used the following wording (le
wording in the other conditions was similar; for details see the Supplemental Materials, Woike
& Kanngiesser, 2019b): “We gave you a choice at the beginning of the survey between taking 5
cents or 15 cents. You decided to take 15 cents and promised to pay back 5 cents. We now ask
you whether you want to pay back 5 cents. We will not force you to pay any money back. Comment
much money (in cents) do you pay back?” Participants could then enter the payback amount
in a text box.
Subsequently, participants completed the PANAS scale (Crawford & Henry, 2004) avec
items in randomized order. Enfin, they were asked to provide reasons for their decision in an
open-ended format. Open-ended answers were categorized by two independent coders based
on a coding scheme developed by the authors (for details, see the Supplemental Materials,
Woike & Kanngiesser, 2019b). As the dependent variables included proportions, we aimed
for sample sizes well beyond 100 participants for each condition in order to obtain measures
with reasonable precision.
The survey was implemented in Qualtrics. Allocation to condi-
Survey and Analysis Software
tions used the Qualtrics randomizer with samples sizes kept constant. Statistical analyses were
conducted with IBM SPSS 24 and the Exploratory Software for Confidence Intervals (ESCI),
downloaded from Cumming (2014; see Cumming, 2012).
OPEN MIND: Discoveries in Cognitive Science
70
je
D
o
w
n
o
un
d
e
d
F
r
o
m
h
t
t
p
:
/
/
d
je
r
e
c
t
.
m
je
t
.
/
e
d
toi
o
p
m
je
/
je
un
r
t
je
c
e
–
p
d
F
/
d
o
je
/
je
/
.
/
1
0
1
1
6
2
o
p
m
_
un
_
0
0
0
2
7
2
0
6
6
9
7
6
o
p
m
_
un
_
0
0
0
2
7
p
d
.
/
je
F
b
oui
g
toi
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
Most People Keep Their Word Woike and Kanngiesser
Results
Across conditions the majority of participants accepted the higher
Acceptance Decision
amount (see Figure 1a). Specifically, the acceptance rate was apw = 0.84 (99% CI [confidence
interval] = [0.75, 0.91]) in the promise (write) condition, apc = 0.88 ( 99% CI = [0.80, 0.94])
in the promise (click) condition, and aa = 0.92 (99% CI [0.85, 0.96]) in the ask condition.
The highest acceptance rate was observed in the control condition with ac = 0.94 (99% CI =
[0.86, 0.97]).
Payback Decision Next, we analyzed payback rates among those who had accepted the
higher payment of $0.15, with payback defined as returning at least the promised or suggested amount of $0.05 (see Figure 1b). The vast majority of participants repaid, as promised, dans le
je
D
o
w
n
o
un
d
e
d
F
r
o
m
h
t
t
p
:
/
/
d
je
r
e
c
t
.
m
je
t
.
/
e
d
toi
o
p
m
je
/
je
un
r
t
je
c
e
–
p
d
F
/
d
o
je
/
je
/
.
/
1
0
1
1
6
2
o
p
m
_
un
_
0
0
0
2
7
2
0
6
6
9
7
6
o
p
m
_
un
_
0
0
0
2
7
p
d
/
.
je
F
b
oui
g
toi
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
Chiffre 1. Proportion of participants in Study 1a (un) accepting the higher amount of $0.15 in the beginning and (b) repaying ≥$0.05 at the end. Proportions are shown as markers for each condition
with bars corresponding to the 99% CI (confidence interval). The payback proportion is relative to
the number of participants accepting the higher amount.
OPEN MIND: Discoveries in Cognitive Science
71
Most People Keep Their Word Woike and Kanngiesser
promise (write) condition (bpw = 0.95, 99% CI =[0.87, 0.98], n = 116) and the promise (click)
condition (bpc = 0.94, 99% CI = [0.86, 0.98], n = 118). Payback rates were lower in the ask
condition (ba = 0.74, 99% CI = [0.63, 0.83], n = 126) and very low in the control condition
(bc = 0.33, 99% CI = [0.23, 0.45], n = 114). Details of the distribution of payback amounts
can be found in the Supplemental Materials (Woike & Kanngiesser, 2019b; see Figure S33).
Affect We analyzed the positive and negative PANAS subscales with two separate ANOVAs,
entering the payback above threshold (yes/no) as one factor and condition as a second factor.
For positive affect, there were no significant main effects of condition [F(3, 465) = 2.18, p =
.09, partial η2 = .01] and payback [F(1, 465) = 0.07, p = .80, partial η2 = .00], or their
two-way interaction [F(3, 465) = 1.50, p = .22, partial η2 = .01]. For negative affect, nous
found a significant main effect of payback [F(1, 465) = 19.60, p < .001, partial η2 = .04],
but no significant main effect of condition [F(3, 465) = 0.24, p = .87, partial η2 = .00] nor a
significant two-way-interaction [F(3, 465) = 0.88, p = .45, partial η2 = .01]. Participants who
repaid below threshold experienced significantly more negative affect (M = 3.89, 99% CI =
[2.40, 5.38]) than those who repaid above threshold (M = 1.15, 99% CI = [0.56, 1.74]).
Justifications were scored separately for participants who (1) took the lower pay-
Reasons
ment (i.e., rejected $0.15), (2) paid back less than $0.05, and (3) paid back $0.05 or more (see Table 1 and the Supplemental Materials, Woike & Kanngiesser, 2019b). In the first group (64 comments), participants viewed the lower payment as the simpler option and as being free of potentially negative consequences, and in the promise conditions, some wanted to avoid a commitment. In the second group (120 comments), participants in all conditions referred to self-regarding preferences and free choice when justifying payback of less than $0.05; participants in the control condition saw no reason to repay money. In the third group
(381 comments), the majority of participants in the promise or ask conditions referred to norms
or the entered covenant when justifying paying back ≥$0.05; some also referred to moral considerations or personal values. Participants in the control condition referred to monetary reasons, norms, and morals. Table 1. Percentages of justifications for decisions in Study 1a. Group Reason (write) (click) Ask Control Promise Rejected 15 Paid back <$0.05
Avoid consequences
Simpler option
No commitment
Other
Self-regard
Free choice
No reason
Other
Paid back ≥$0.05 Norms Morals Monetary reason Personal values Other 48 39 4 13 33 50 0 0 70 14 8 20 5 47 35 29 6 56 33 0 0 74 14 9 13 4 82 9 0 9 36 45 9 9 85 12 7 9 7 71 43 0 14 44 8 35 9 18 24 32 6 21 Note. Justifications could be scored in more than one category. Percentages are relative to the number of valid comments. OPEN MIND: Discoveries in Cognitive Science 72 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . / e d u o p m i / l a r t i c e - p d f / d o i / i . / / 1 0 1 1 6 2 o p m _ a _ 0 0 0 2 7 2 0 6 6 9 7 6 o p m _ a _ 0 0 0 2 7 p d . / i f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 Most People Keep Their Word Woike and Kanngiesser STUDY 1B In this study, participants were asked to retrodict the acceptance and the payback decisions in Study 1a in an incentivized paradigm. Methods All participants were located in the United States and recruited through MTurk. Participants Double participation was prevented by a combination of using the Unique Turker script, check- ing against exclusion lists at the beginning of the survey, and carrying out manual checks of MTurk IDs and IP addresses. After these steps, there were n1b = 397 participants in Study 1b (Mage = 33.9, 53.9% female, 45.3% male, 0.8% other). All participants gave their in- formed consent and the study was approved by the IRB at the Max Planck Institute for Human Development. Procedure We asked participants to retrodict the observed rates for the acceptance and the pay- back decisions in Study 1a. In a between-subjects design, participants estimated decisions for either the promise (write) condition (n = 203) or the control condition (n = 193). Participants were incentivized: They were told that estimates close to the observed values would lead to an increase in bonus money (see the Supplemental Materials, Woike & Kanngiesser, 2019b). Results Estimations of the acceptance rates are summarized in Estimation: Acceptance Decision Figure 2a. In the promise (write) condition, participants estimated, on average, acceptance rates of ˆapw = 0.81 (99% CI = [0.76, 0.85]), n = 203) and in the control condition of ˆac = 0.84 (99% CI = [0.80, 0.88]), n = 193). Average expectations thus aligned well with the observed behavioral data in Study 1a. l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . / e d u o p m i / l a r t i c e - p d f / d o i / i / / . 1 0 1 1 6 2 o p m _ a _ 0 0 0 2 7 2 0 6 6 9 7 6 o p m _ a _ 0 0 0 2 7 p d / . i f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 Figure 2. Distribution, means, and confidence intervals (CIs) for the (a) estimated acceptance and (b) estimated payback rates in Study 1b. The histograms show the distributions of individual estimates, the markers represent mean estimates, and whiskers correspond to the 99% CI for the mean. The nonshaded areas mark the CIs for the observed rates in Study 1a. OPEN MIND: Discoveries in Cognitive Science 73 Most People Keep Their Word Woike and Kanngiesser Estimation: Payback Decision On average, participants estimated a payback rate of ˆbc = 0.30 (99% CI = [0.24, 0.35], n = 192) in the control condition, which aligned well with our results in Study 1a (see Figure 2b, top). In contrast, participants underestimated the observed payback rates in the promise (write) condition by more than 30%, with ˆbpw = 0.58 (99% CI = [0.53, 0.64], n = 192); many participants gave low estimates for promise-keeping rates (see Figure 2b, bottom). STUDY 2A We conducted a study with a new group of participants to replicate and extend the findings of Study 1a. Specifically, we manipulated the stakes by a factor of 10 for both conditions: Partic- ipants were offered a (low-stake) choice between $0.05 and $0.20 with a possible payback of $0.10 (20–10) or a (high-stake) choice between $0.50 and $2.00 with a possible payback of
$1.00 (200–100). Previous work has shown that stake sizes had no effect on promise-keeping rates in a social dilemma (van den Assem et al., 2012); here we investigated whether this would also be the case for promises in our novel, minimal promising paradigm. We used the same general setup as in Study 1a, but included only the promise (write) condition (henceforth: promise condition) and the control condition. Methods All participants were located in the United States and recruited through MTurk. Participants Double participation was prevented by a combination of using the Unique Turker script, check- ing against exclusion lists at the beginning of the survey, and carrying out manual checks of MTurk IDs and IP addresses. After these steps, there were n2a = 846 participants in Study 2a (Mage = 33.1, 52.6% female, 47.4% male). All participants gave their informed consent and the study was approved by the IRB at the Max Planck Institute for Human Development. To test for effects of stake size, we manipulated the stakes by a factor of 10 for Procedure both conditions: Participants were offered a (low-stake) choice between $0.05 and $0.20 with a possible payback of $0.10 (20–10) or a (high-stake) choice between $0.50 and $2.00 with a
possible payback of $1.00 (200–100). This resulted in a 2 (control vs. promise) x 2 (low-stake vs. high-stake) between-subject design. Sample sizes were again set for reasonably accurate estimates; they ranged from 209 to 214. We added a short postquestionnaire to test compre- hension of the task and elicited justifications through a set of questions. Participants responded to the PANAS scale (Crawford & Henry, 2004) with items in randomized order. Results In the promise condition, acceptance rates were apl = 0.83 (99% CI = Acceptance Decision [0.75, 0.88], n = 214) for low stakes and aph = 0.89 (99% CI = [0.82, 0.93], n = 211) for high stakes (see Figure 3a). In the control condition, acceptance rates were acl = 0.90 (99% CI = [0.84, 0.95], n = 209) for low stakes and ach = 0.93 (99% CI = [0.86, 0.96], n = 212) for high stakes. Overall, these rates were comparable to the respective rates in Study 1a. Participants who had accepted the higher amount were asked three Comprehension Checks comprehension questions to assess whether they (1) correctly recalled if they had promised or not, (2) correctly calculated the consequences of their decisions for their bonus payment, and (3) understood that real money was at stake and paid back to the requester, not taken from or given to another participant. The first three rows of Table 2 show the percentage of correct OPEN MIND: Discoveries in Cognitive Science 74 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . / e d u o p m i / l a r t i c e - p d f / d o i / i / / . 1 0 1 1 6 2 o p m _ a _ 0 0 0 2 7 2 0 6 6 9 7 6 o p m _ a _ 0 0 0 2 7 p d . / i f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 Most People Keep Their Word Woike and Kanngiesser l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . / e d u o p m i / l a r t i c e - p d f / d o i / i / / . 1 0 1 1 6 2 o p m _ a _ 0 0 0 2 7 2 0 6 6 9 7 6 o p m _ a _ 0 0 0 2 7 p d . / i f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 Figure 3. Proportion of participants (a) accepting the higher payment and (b) paying back at the end in Study 2a. Proportions are shown as markers for each condition with bars corresponding to the 99% CI (confidence interval). The payback proportion is relative to the number of participants accepting the higher amount. Payback rates are presented with different filters applied based on comprehension check performance: no filter (black), lenient filter (blue), and strict filter (red). responses per condition; the most common error was to regard the money that was paid back as imaginary. As there is some genuine ambiguity as to whether virtual transactions involve “real” money (Houser & Xiao, 2015), we report three variants of the payback results: (a) using a strict filter excluding all participants with false answers, (b) using a lenient filter allowing participants who viewed money as imaginary (but had answered questions 1–2 correctly), and (c) unfiltered results. OPEN MIND: Discoveries in Cognitive Science 75 Most People Keep Their Word Woike and Kanngiesser Table 2. Percentage of correct answers to the comprehension questions in Study 2a Control Promise Check $0.20
$2.00 $0.20
$2.00 (1) promise (2) own bonus (3) real money (1–2) correct (1–3) correct 71.1 94.7 62.6 64.7 47.1 68.6 82.0 64.4 55.7 41.8 100.0 97.2 84.2 94.9 81.9 100.0 89.7 80.4 87.0 74.5 Results for the three filter methods were similar (note that filtering led to Payback Decision somewhat lower payback rates for controls and to somewhat higher payback rates for promises; see Figure 3b). Here, we report the results for the strict filter. In the promise conditions, the majority of participants paid back the money; payback rates were 9% higher in the low-stake condition (bpl = 0.98, 99% CI = [0.92, 0.99], n = 145) than in the high-stake condition (bph = 0.89, 99% CI = [0.80, 0.94], n = 137) with δ hl = 0.09 (99% CI = [0.01, 0.18]). In the control conditions, participants’ payback rates were low; payback rates were 13% higher in the low-stake condition (bcl = 0.24, 99% CI = [0.14, 0.37], n = 88) than in the high-stake hl = 0.13 (99% CI = [−0.03, condition (bch = 0.11, 99% CI = [0.05, 0.23], n = 81) with δ 0.27]). Details of the distribution of paid back amounts can be found in the Supplemental Materials (Woike & Kanngiesser, 2019b; Figure S33). To preserve power, we report the analysis with the lenient filter applied (see the Sup- Affect plemental Materials, Woike & Kanngiesser, 2019b, for further analyses). For negative affect, we observed significant main effects of condition [F(1, 549) = 12.16, p < .001, partial η2 = .02] and payback (yes/no) [F(1, 549) = 33.96, p < .001, partial η2 = .06], but not of stakes [F(1, 549) = 0.10, p = .76, partial η2 = .00]. The only significant interaction was between condition and payback [F(1, 549) = 3.97, p = .047, partial η2 = .01]. Average scale values are about 2 points higher in the promise versus control condition, and 3.3 points higher for participants who failed to pay money back. This difference was higher in the promise (δ = 4.5) than in the control condition (δ = 2.2). For positive affect, we observed a significant main effect of payback [F(1, 549) = 10.35, p = .001, partial η2 = .02], but no significant effects of condition [F(1, 549) = 0.05, p = .83, partial η2 = .00] or stakes [F(1, 549) = 0.76, p = .38, partial η2 = .00]. There was no significant interaction, although the interaction between con- dition and payback was similar to the one for negative affect [F(1, 549) = 3.55, p = .06, partial η2 = .01]. Participant who paid money back had a scale value about 4.4 points higher, which was more pronounced in the promise (δ = 6.9) than in the control condition (δ = 1.8). Participants answered a set of quantitative questions on the reasons for their de- Reasons cisions, based on condition and decision type (see the Supplemental Materials, Woike & Kanngiesser, 2019b, for details). (1) Participants who rejected the higher payment, affirmed, on average, that they preferred a certain and simpler option, and to some degree that they feared negative consequences of claiming more. (2) Participants who did not pay money back in the control conditions indicated that they saw no reason for doing so, and felt to some degree enti- tled because they had worked on the survey. (3) Participants who broke their promise similarly referred to a lack of a reason for paying money back, as well as to the lack of enforcement. They also felt their hard work entitled them to the money and affirmed a selfish motivation. (4) OPEN MIND: Discoveries in Cognitive Science 76 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . / e d u o p m i / l a r t i c e - p d f / d o i / i / / . 1 0 1 1 6 2 o p m _ a _ 0 0 0 2 7 2 0 6 6 9 7 6 o p m _ a _ 0 0 0 2 7 p d / . i f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 Most People Keep Their Word Woike and Kanngiesser Participants who kept their promise affirmed, on average, that there is an obligation to keep promises, that they saw themselves as someone who keeps promises, and that they wanted to do the right thing. They rejected, on average, that they kept their promise because the amounts were too small. (5) Participants who paid money back in the control conditions affirmed that they wanted to be nice to the requester or considered themselves to be generous. They did not expect to earn more money as a consequence. STUDY 2B We asked a separate group of participants to estimate payback rates for Study 2a by specify- ing their subjective probability distribution over possible intervals. Responses were incen- tivized using the quadratic scoring rule (Harrison, Martínez-Correa, Swarthout, & Ulm, 2017; Matheson & Winkler, 1976). Methods All participants were located in the United States and recruited through MTurk. Participants Double participation was prevented by a combination of using the Unique Turker script, checking against exclusion lists at the beginning of the survey, and carrying out manual checks of MTurk IDs and IP addresses. After these steps, there were n2b = 1, 384 participants in Study 2b (Mage = 35.6, 56.9% female, 43.0% male, 0.1% other). All participants gave their in- formed consent and the study was approved by the IRB at the Max Planck Institute for Human Development. Participants were asked to retrodict the proportion of participants paying back at Procedure least the requested amount in the four conditions of Study 2a, that is, the low-stake (20–10) and high-stake (200–100) conditions in both the promise and control variants. In contrast to Study 1b, participants were incentivized using the quadratic scoring rule: Participants allocated 100 points across 22 intervals (all 5% intervals between 5% and 95% and a split of the smallest and highest intervals into 4.99% and the extreme 0.01%), resulting in relative allocation of r1 to r22 with ∑i ri = 1. Participants received bonus money based on their allocations to · $0.50.
the correct interval rc and false intervals ri(cid:3)=c, with a payoff of o =
Procedure and payment consequences were explained to participants in a multistep tutorial
(see the Supplemental Materials, Woike & Kanngiesser, 2019b, also for demonstrations of the
consequences).
2rc − ∑i(cid:3)=c r2
i
(cid:2)
(cid:3)
Data Analyses We considered participants with intermittent gaps in their allocation to inter-
vals as inconsistent and analyzed the data with and without this group.
Results
Expectations Most participants’ allocations were classified as consistent (π
all = 0.78, with
0.76 < π
i < 0.81 across conditions). Distributions and average allocations to intervals are
summarized in Figure 4. Individual estimates were calculated by summing the products of
relative allocations and interval midpoints, resulting in the distribution and means summarized
in Figure 5. Filtered results were similar to unfiltered results (see the Supplemental Materials,
Woike & Kanngiesser, 2019b). Participants estimated payback rates in the control conditions
as ˆbcl = 0.25 (99% CI =[0.21, 0.30], n = 278) for low stakes and ˆbch = 0.22 (99% CI =[0.17,
0.26], n = 265) for high stakes. The average estimates for the control conditions corresponded
well to the observed rates in Study 2a, similar to Study 1. In the promise conditions, participants
OPEN MIND: Discoveries in Cognitive Science
77
l
D
o
w
n
o
a
d
e
d
f
r
o
m
h
t
t
p
:
/
/
d
i
r
e
c
t
.
m
i
t
.
/
e
d
u
o
p
m
i
/
l
a
r
t
i
c
e
-
p
d
f
/
d
o
i
/
i
/
/
.
1
0
1
1
6
2
o
p
m
_
a
_
0
0
0
2
7
2
0
6
6
9
7
6
o
p
m
_
a
_
0
0
0
2
7
p
d
/
.
i
f
b
y
g
u
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
Most People Keep Their Word Woike and Kanngiesser
l
D
o
w
n
o
a
d
e
d
f
r
o
m
h
t
t
p
:
/
/
d
i
r
e
c
t
.
m
i
t
.
/
e
d
u
o
p
m
i
/
l
a
r
t
i
c
e
-
p
d
f
/
d
o
i
/
i
/
/
.
1
0
1
1
6
2
o
p
m
_
a
_
0
0
0
2
7
2
0
6
6
9
7
6
o
p
m
_
a
_
0
0
0
2
7
p
d
/
.
i
Figure 4. Distribution and means for individual subjective probability distributions in Study 2b for low stakes (left) and high stakes (right)
crossed with control (top) and promise (bottom) conditions. Markers correspond to the average allocations to each interval; blue shading
corresponds to the number of participants who committed at least the shaded amount of points to the respective interval.
again underestimated payback rates by about 40% for low stakes, ˆbpl = 0.53 (99% CI = [0.48,
0.57], n = 267), and by about 20% for high stakes, ˆbph = 0.59 (99% CI = [0.55, 0.64],
n = 267). The estimated rate was higher for promise high stakes than for promise low stakes
(δ = 0.06, 99% CI = [0.02, 0.11]), with the observed rates differing in the opposed direction
(see Figure 3).
STUDY 3
We conducted a further study with a new group of participants to test the robustness of the
promising-keeping rates in Studies 1a and 2a. In this study, we manipulated two new factors
and included additional robustness checks. First, we manipulated the visibility of promise-
keeping or promise-breaking. Participants were divided evenly between a clear condition and
an obfuscation condition. In the obfuscation condition, participants played a guessing game
similar to the task employed in Jiang (2013): They were asked to guess a number between
f
b
y
g
u
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
OPEN MIND: Discoveries in Cognitive Science
78
Most People Keep Their Word Woike and Kanngiesser
Figure 5. Distribution, means, and confidence intervals (CIs) for estimated payback rates in Study 2b. The histogram shows the distribution
of individual estimates for both promise conditions (top and bottom) for (a) low stakes and (b) high stakes. Histograms represent the distributions
of calculated individual point estimates. The markers show the mean estimate and whiskers correspond to the 99% CI for the mean (the filled
marker for consistent participants, the unfilled marker for all participants). Different degrees of shading delimit the CIs for the corresponding
rates observed in Study 2a, with the interval marked by orange lines for the strict filter, black for the lenient filter, and gray for no filter.
0 and 9 (without indicating their guess) and were then shown a randomly generated num-
ber. Participants whose guess matched the number were instructed to pay back zero cents
in their final decision. Participants in the obfuscation condition thus knew that a payback of
zero was ambiguous for the experimenter: There was no way of knowing whether they had
guessed correctly. The manipulation made it possible to assess the impact of potential con-
cerns about reputation and negative consequences for future work opportunities on MTurk on
payback decisions. Second, we varied whether the target of the payback was the experimenter
(as in Studies 1a and 2a) or another participant (another MTurk worker). Specifically, in the
experimenter condition participants were told that the money would be paid back to the ex-
perimenter; in the peer condition, they were told the money they paid back would go to a
future participant. This allowed us to test whether the experimenter’s authority had any influ-
ence on participants’ payback decisions. All manipulations were run in a promise and a control
condition (without a promise). This resulted in a 2 (obfuscation vs. clear) x 2 (experimenter vs.
peer) x 2 (promise vs. control) between-subject design.
We used the same payment amounts as in the low-stake conditions of Study 2a (20–10)
and presented the promise option in the (weaker) click-only variant from Study 1a. In addition,
we introduced the following changes: (i) To reduce possible demand effects, we eliminated the
reminder of the choice and promised/suggested amounts of payback before participants made
a decision about paying money back. We also clearly stated that there would be no further
changes to the bonus payment (e.g., surprise payments) after their decision, and explained in
all conditions that our analysis would focus on the group level, not on particular individuals.
(ii) To increase the sense of endowment for the chosen amount of money, we displayed how
OPEN MIND: Discoveries in Cognitive Science
79
l
D
o
w
n
o
a
d
e
d
f
r
o
m
h
t
t
p
:
/
/
d
i
r
e
c
t
.
m
i
t
.
/
e
d
u
o
p
m
i
/
l
a
r
t
i
c
e
-
p
d
f
/
d
o
i
/
i
.
/
/
1
0
1
1
6
2
o
p
m
_
a
_
0
0
0
2
7
2
0
6
6
9
7
6
o
p
m
_
a
_
0
0
0
2
7
p
d
/
.
i
f
b
y
g
u
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
Most People Keep Their Word Woike and Kanngiesser
much bonus money participants currently owned on each page of the survey. We preregistered
the study before data collection on the OSF platform (Woike & Kanngiesser, 2019a).
Methods
All participants were located in the United States and recruited through Amazon’s
Participants
Mechanical Turk. In reaction to current data quality concerns (Kennedy, Clifford, Burleigh,
Waggoner, & Jewell, 2018), we followed the procedures preventing the circumvention of lo-
cation requirements suggested by Burleigh, Kennedy, and Clifford (2018) and also included
specific tests (Moss & Litman, 2018) of language competence (see the Supplemental Materials,
Woike & Kanngiesser, 2019b). Double participation was prevented by a combination of using
the Unique Turker service, adding qualifications to exclude previous participants, and carrying
out manual checks to ensure the uniqueness of MTurk IDs and IP addresses. After these steps,
there were n3 = 1, 001 participants in Study 3 (Mage = 36.6, 46.6% female, 53.3% male, 0.1%
other). A separate group of n3p = 257 participants were passive recipients of bonus transfers in
the peer conditions. All participants gave their informed consent and the study was approved
by the IRB at the Max Planck Institute for Human development.
All participants were offered a choice between $0.05 and $0.20 with a sug-
Procedure
gested/promised payback of $0.10 (20–10). We varied three factors in this study (see the Supplemental Materials, Woike & Kanngiesser, 2019b). First, half of the participants played a guessing game (obfuscated condition), the other half did not (clear condition). Participants in the obfuscated condition mentally guessed a number between 0 and 9 and were then shown a randomly drawn number. In case of a match (labeled “winning the game”), participants were instructed to not pay back any money at the end of the survey. The procedure ensured that it was public knowledge that only the participants themselves could know whether a match had occurred. Second, the payback target was either the experimenter (as in all previous studies) or a peer. In the peer condition, participants were paired with future participants on MTurk (mem- bers of a separate group) and their decisions had consequences for both participants’ payoffs. Choosing the lower payoff resulted in $0.05 for both participants. Choosing the higher payoff
resulted in a total bonus amount of $0.20 for the current participant with a chance to pay back any amount at the end the study. Participants were informed that the future participants would receive a protocol of their decisions (including promises made and, in the obfuscated condi- tion, the existence and possible consequences of the guessing game). Third, participants were assigned to a control condition (the same as in Study 2a) or a promise condition (the promise [click] condition in Study 1a). This resulted in a 2 (obfuscation vs. clear) x 2 (experimenter vs. peer) x 2 (promise vs. control) between-subject design. There were 123 to 126 participants per cell of the design. In contrast to the previous studies, there was no reminder of promises or suggested/promised payback amounts at the end of the study, and a participant’s current bonus amount was displayed at the top of their screen throughout the study. We added a short postquestionnaire to test the success of the promise and target manipulations (see the Supplemental Materials, Woike & Kanngiesser, 2019b). Results Since the guessing game occurred after the acceptance decision, ac- Acceptance Decision ceptance rates were collapsed for the obfuscation/clear condition. In the promise condition, acceptance rates were ape = 0.88 (99% CI = [0.82, 0.93], n = 251) for the experimenter target and app = 0.85 (99% CI = [0.78, 0.90], n = 250) for the peer target. In the control condition, OPEN MIND: Discoveries in Cognitive Science 80 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . / e d u o p m i / l a r t i c e - p d f / d o i / i . / / 1 0 1 1 6 2 o p m _ a _ 0 0 0 2 7 2 0 6 6 9 7 6 o p m _ a _ 0 0 0 2 7 p d . / i f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 Most People Keep Their Word Woike and Kanngiesser acceptance rates were ace = 0.92 (99% CI = [0.87, 0.96], n = 251) for the experimenter target and acp = 0.88 (99% CI = [0.82, 0.93], n = 249) for the peer target (see Figure 6). Participants who had accepted the higher amount were asked two ma- Manipulation Checks nipulation questions to assess whether they (1) correctly recalled if they had promised or not, and (2) understood that real money was at stake and remembered who the target of the payback decision was. Among the 886 participants who had accepted the higher amount, 741 par- ticipants (74%) correctly answered the first question and 757 participants (75.6%) correctly answered the second question; 639 participants (63.8%) correctly answered both questions. Results for the payback decision were calculated for the unfiltered and filtered sample (re- stricted to the group with two correct answers), respectively. Results for the filtered and unfiltered sample were similar (note that filtering Payback Decision resulted again in somewhat lower payback rates in the control and somewhat higher payback rates in the promise conditions; see Figure 7 and Figure 8). Here we report the results for the filtered sample split by factor levels (for details on the distribution of paid-back amounts, see the Supplemental Materials (Woike & Kanngiesser, 2019b; Figure S33)), and higher in the promise conditions (bp = 0.73, 99% CI = [0.67, 0.79], n = 346) than in the control conditions (bc = 0.34, 99% CI = [0.27, 0.42], n = 293), with δpc = 0.39 (99% CI = [0.29, 0.48]). Payback rates were higher in the peer conditions (bpe = 0.66, 99% CI = [0.59, 0.73], n = 296) than in the experimenter conditions (bex = 0.46, 99% CI = [0.39, 0.53], n = 343), with δpeex = 0.20 (99% CI = [0.10, 0.29]), with a larger difference in the control conditions. Payback rates in the control conditions were higher than average donation rates in dictator games (about 29.8% in the meta-analysis conducted by Engel (2011), which might be due to our framing of taking bonus money from other participants (Oxoby & Spraggon, 2008). Payback rates in the clear conditions (bcl = 0.58, 99% CI = [0.51, 0.65], n = 313) were slightly higher than those in the obfuscated conditions (bob = 0.52, 99% CI = [0.45, 0.59], n = 326), with δ clob = 0.06 (99% CI = [-0.04, 0.16]). However, an analysis of individual conditions showed that the difference between clear and obfuscated conditions deviated only marginally from the expected difference—that is, if those participants who had won the guessing game returned zero and everyone else returned at the average rate observed in the corresponding clear conditions (see Figure 7). Interestingly, some participants voiced tensions in reconciling their promise to pay back with the instructions not to pay back after winning the guessing game in their comments at the end of the survey (see the Supplemental Materials, Woike & Kanngiesser, 2019b). GENERAL DISCUSSION We tested a minimal, fully incentivized method to measure promise-keeping and found across three studies that the majority of participants (61%–98%) in an online labor market kept their promises at a financial cost to themselves. Promises thus exert a strong normative force, even in the face of monetary temptations. Surprisingly, we found that different groups of participants, asked to estimate promise-keeping rates in Studies 1b and 2b, systematically underestimated promise-keeping by 20%–40%. This behavior–estimation gap could mean that people lose out on chances to cooperate and coordinate with others because they are overly pessimistic about the power of others’ words. The effect of promises on behavior has been studied in a wide range of situations and paradigms (Bicchieri, 2002; Carlsson et al., 2013; Carlsson & Kataria, 2018; Charness & OPEN MIND: Discoveries in Cognitive Science 81 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . / e d u o p m i / l a r t i c e - p d f / d o i / i / / . 1 0 1 1 6 2 o p m _ a _ 0 0 0 2 7 2 0 6 6 9 7 6 o p m _ a _ 0 0 0 2 7 p d . / i f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 O P E N M N D I : i D s c o v e r i e s i n C o g n i t i v e S c e n c e i M o s t P e o p l e K e e p T h e i r W o r d W o i k e a n d K a n n g i e s s e r l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . / e d u o p m i / l a r t i c e - p d f / d o i / i / . / 1 0 1 1 6 2 o p m _ a _ 0 0 0 2 7 2 0 6 6 9 7 6 o p m _ a _ 0 0 0 2 7 p d . / i Figure 6. Proportion of participants in Study 3 accepting the higher amount of $0.20 in the beginning, split by individual conditions and factor levels. Proportions
are shown as markers for each condition with bars corresponding to the 99% CI (confidence interval).
8
2
f
b
y
g
u
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
O
P
E
N
M
N
D
I
:
i
D
s
c
o
v
e
r
i
e
s
i
n
C
o
g
n
i
t
i
v
e
S
c
e
n
c
e
i
M
o
s
t
P
e
o
p
l
e
K
e
e
p
T
h
e
i
r
W
o
r
d
W
o
i
k
e
a
n
d
K
a
n
n
g
i
e
s
s
e
r
l
D
o
w
n
o
a
d
e
d
f
r
o
m
h
t
t
p
:
/
/
d
i
r
e
c
t
.
m
i
t
.
/
e
d
u
o
p
m
i
/
l
a
r
t
i
c
e
-
p
d
f
/
d
o
i
/
i
.
/
/
1
0
1
1
6
2
o
p
m
_
a
_
0
0
0
2
7
2
0
6
6
9
7
6
o
p
m
_
a
_
0
0
0
2
7
p
d
.
/
i
Figure 7. Proportion of participants paying back at least $0.10 in Study 3 for all conditions. Proportions are shown as markers for each condition with bars corre- sponding to the 99% CI (confidence interval). The payback proportion is relative to the number of participants accepting the higher amount. Payback rates are presented unfiltered (larger markers) and filtered (smaller markers). Arrowheads mark the estimated rates of payback when replacing expected zero payments (10% of participants) with observed rates from the corresponding clear conditions. 8 3 f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 Most People Keep Their Word Woike and Kanngiesser Figure 8. Proportion of participants paying back at least $0.10 in Study 3 for factor levels. Propor-
tions are shown as markers for each condition with bars corresponding to the 99% CI (confidence
interval). The payback proportion is relative to the number of participants accepting the higher
amount. Payback rates are presented unfiltered (larger markers) and filtered (smaller markers).
Dufwenberg, 2006; Ellingsen & Johannesson, 2004; Heyman, Fu, Lin, Qian, & Lee, 2015;
Jacquemet, Luchini, Rosaz, & Shogren, 2019; Kanngiesser, Köymen, & Tomasello, 2017;
Kataria & Winter, 2013; Orbell et al., 1988; Ostrom et al., 1992; Sally, 1995; van den Assem
et al., 2012; Wang & Katzev, 1990). Previous studies have found that the context of promis-
ing matters: Promises are more effective in dyadic than in group settings (Lev-On, Chavez, &
Bicchieri, 2010) and more effective in face-to-face communication than in computer-mediated
interactions (Bicchieri & Lev-On, 2007; but see Conrads & Reggiani, 2017), and are interpreted
differently in competitive contexts (Casella, Kartik, Sanchez, & Turban, 2018). We found in
Study 3 that the type of target of a promise had an impact as well: Promise-keeping rates were
higher for an unidentified peer (another MTurk participant) than for a presumed authority figure
(the experimenter)—though it should be noted that participants in the control conditions also
shifted toward more generous paybacks for the peer. Importantly, we can consistently show
across three studies that promises were highly effective in a minimal setting where neither the
promisor nor the promisee knew each other, had face-to-face contact, or exchanged personal
messages. Even in situations in which participants were not reminded of their promise and in
which breaking a promise was obfuscated through a guessing game (Study 3), the majority of
participants (61%–74%) kept their word. In light of this, our minimal promise situations can
be considered a conservative test of people’s promise-keeping.
We found in Study 1a that the commitment format did not affect promise-keeping:
Promise-keeping rates were equally high for participants who simply clicked on a prewritten
promise (promise [click]: 94%) and for participants who wrote a minimal promise themselves
(writing “I promise”: 95%). It was only when the word “promise” was omitted altogether (ask con-
dition) that payback rates decreased. Previous research has found that prewritten promises had
less of an effect on increasing cooperation than free-form messages (Charness & Dufwenberg,
2006, 2010). However, sample sizes in these two studies were considerably smaller than in our
OPEN MIND: Discoveries in Cognitive Science
84
l
D
o
w
n
o
a
d
e
d
f
r
o
m
h
t
t
p
:
/
/
d
i
r
e
c
t
.
m
i
t
.
/
e
d
u
o
p
m
i
/
l
a
r
t
i
c
e
-
p
d
f
/
d
o
i
/
i
.
/
/
1
0
1
1
6
2
o
p
m
_
a
_
0
0
0
2
7
2
0
6
6
9
7
6
o
p
m
_
a
_
0
0
0
2
7
p
d
/
.
i
f
b
y
g
u
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
Most People Keep Their Word Woike and Kanngiesser
study.1 Also, the two studies were not run simultaneously and rates of promise-making differed
substantially between these groups,2 making it likely that different pools of participants were
compared. Another study found that unconditional and explicit statements were associated
with higher cooperation rates in a TV show than conditional or implicit statements (Turmunkh
et al., 2019). Yet the types of statements varied considerably and included wishes and desires
(e.g., “I want us to do X”), which we would not count as promises. It is, in our view, perfectly
intelligible to say: ”I intend to do X, but I cannot promise it.” Future research could use our
novel paradigm to measure promise-keeping for different promise formats such as conditional
or unconditional promises, or to compare promises and mere statements of intent.
Varying stake sizes in Study 2a ($0.20 vs. $2.00) did not result in noticeable differences
in promise-keeping rates. A recent meta-analysis on stake sizes in economic games found
that these effects were game-specific and somewhat inconsistent across studies—for example,
small average effects were found for dictator games, but not for ultimatum games (Larney,
Rotella, & Barclay, 2019). Research with participant samples similar to the ones in our studies
(U.S. participants on MTurk) showed clear differences between small stakes and no stakes
in dictator games (Amir, Rand, & Gal, 2012), but virtually no differences when raising stakes
above $1.00 (Raihani, Mace, & Lamba, 2013). Our choice of stake sizes of up to $2.00 appears
appropriate to produce generalizable findings with MTurk participants.
Participants justified their promise-keeping by referring to promise norms and the obli-
gation to keep one’s word. As one participant in Study 1a stated: “Mutual agreements are the
basis of social contracts. A social agreement, once agreed to, ought to be binding regardless
of the circumstances—with very rare exceptions indeed. Once I had agreed, it’s true that there
was no force upon me to bind me, except my own moral values–but because I believe in the
necessity of living up to one’s word, and I wanted to remain consistent within myself, I chose
to honor the agreement that I had made.” These types of justifications appear more consistent
with accounts of promise-keeping that assume a preference for promise-keeping (Ellingsen &
Johannesson, 2004; Vanberg, 2008) than with accounts that assume an aversion to disappoint
others’ expectations (Charness & Dufwenberg, 2006; Scanlon, 1990). While participants’ jus-
tifications indicate that promise-keeping mattered to them, we cannot rule out that an aversion
to (or guilt about) disappointing others’ expectations may have also played a role in partici-
pants’ decisions to honor their promise. In fact, recent work has shown that promise-keeping
could be driven by both a preference to keep one’s promise and expectation-based guilt aver-
sion (Bhattacharya & Sengupta, 2016; Di Bartolomeo et al., 2018; Ederer & Stremitzer, 2017;
Kolodny & Wallace, 2003; Mischkowski et al., 2016). In line with both explanations, pay-
ing money back was on average accompanied by more positive and less negative affect than
keeping it.
Participants in our study systematically underestimated payback rates in the promise con-
ditions, but accurately estimated them in the control conditions, indicating that participants
were, in principle, able to accurately predict others’ behavior. Estimations of promise-keeping
rates were characterized by a high degree of variability, demonstrating that a vast majority of
1 Based on the reported sample sizes, the 99% CI for cooperative choices after bare promises in Charness
and Dufwenberg (2010) was [0.41, 0.78] (compared to [0.27, 0.63] in the control condition). Messages contain-
ing statements of intent in Charness and Dufwenberg (2006) resulted in a cooperation rate with a 99% CI of
bm = 0.18, 99% CI = [−0.41, 0.07]), which could be considered limited evidence for the
[0.61, 0.90], such that Δ
claim that “bare promises had substantially less effect on behavior” (Charness & Dufwenberg, 2010, p. 283).
2 More than 30% of participants self-selected to make a promise in the bare messages condition as compared
to participants classified as stating an intent in the free-form condition.
OPEN MIND: Discoveries in Cognitive Science
85
l
D
o
w
n
o
a
d
e
d
f
r
o
m
h
t
t
p
:
/
/
d
i
r
e
c
t
.
m
i
t
.
/
e
d
u
o
p
m
i
/
l
a
r
t
i
c
e
-
p
d
f
/
d
o
i
/
i
.
/
/
1
0
1
1
6
2
o
p
m
_
a
_
0
0
0
2
7
2
0
6
6
9
7
6
o
p
m
_
a
_
0
0
0
2
7
p
d
/
.
i
f
b
y
g
u
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
Most People Keep Their Word Woike and Kanngiesser
participants did not have an accurate expectation of promise-keeping; this is surprising given
the observed self-binding power of promises in Studies 1a and 2a and the fact that estima-
tors had access to the full description of conditions. A positive bias toward one’s own versus
others’ morality (Pronin, 2008) may only partially explain why people were pessimistic about
others’ promise-keeping. Moreover, our findings cannot be explained by the “false consensus
bias” that has been observed for beliefs about dictator game contributions (Iriberri & Rey-Biel,
2013), as this bias would lead only the small minority that breaks promises to underestimate
promise-keeping rates. Other studies have produced inconsistent results regarding participants’
estimations of others’ trustworthiness or promise-keeping rates: Participants underestimated the
general trustworthiness of trustees in trust games without promises (Fetchenhauer & Dunning,
2010) and underestimated the effect of voluntary promises on cooperation in game show
videos by a magnitude (Belot et al., 2012), but overestimated the trustworthiness of written
promises or statements of intent in a trust game (Chen & Houser, 2017). However, estimations
in all these studies were based on small samples. Moreover, participants were asked to judge
specific interactions or messages and it is thus possible that situational cues (e.g., nonverbal
cues, wording) affected estimation rates.
In Study 3 we tested the robustness of promise-keeping rates. We found that promise-
keeping rates were lower overall than in Studies 1a and 2a, which included a reminder of the
initial promise. It is possible that the reminder, which was included to rule out memory errors,
increased promise-keeping rates—either because it reminded people who would have other-
wise forgotten or because it highlighted the commitment, or both. We further found relatively
high amounts of payback in the control conditions with peer targets in Study 3. Participants on
MTurk have been observed to show an in-group bias toward other members of crowdsourced
platforms (Almaatouq, Krafft, Dunham, Rand, & Pentland, 2018) that is not extended to re-
searchers on the platforms (Semuels, 2018). In addition, our framing of taking bonus money
from peers may have increased payback rates in the control conditions (see e.g., Oxoby &
Spraggon, 2008). Interestingly, promise-keeping rates were similar in the obfuscation and the
clear conditions in Study 3, even though potential recipients in the obfuscation condition were
aware of the possibility of a chance-based elimination of payments and neither the recipient nor
the experimenter could know the result of the guessing game. This strengthens the argument
that promises have a self-binding function that goes beyond conforming to the expectations of
others (Ellingsen & Johannesson, 2004; Ismayilov & Potters, 2017; Vanberg, 2008).
Across studies, we incentivized the choice of the promise options in order to measure
promise-keeping rates for a large proportion of participants and reduce self-selection of, for
example, more cooperative participants. It should be noted that participants who opted out of
the promise incurred a loss of money, which is an indication of taking promises seriously rather
than a propensity for promise-breaking. Restricting the analysis to participants who did choose
to make a promise is therefore likely to yield a conservative measure of promise-keeping. In
addition, our paradigm is straightforward and simple to implement and opens many avenues for
future research on promise-keeping, such as using promises to induce more ethical behaviors.
In conclusion, we found that most people chose the integrity of their word over easy per-
sonal financial gain and did so more often than others would expect. Thus, there may well exist
a yet-untapped potential for promises to establish and facilitate coordination and cooperation.
ACKNOWLEDGMENTS
We thank two anonymous reviewers, Yaakov Kareev, Ryan Murphy, Zoe Rahwan, and members
of the Center of Adaptive Rationality at the Max Planck Institute for Human Development
OPEN MIND: Discoveries in Cognitive Science
86
l
D
o
w
n
o
a
d
e
d
f
r
o
m
h
t
t
p
:
/
/
d
i
r
e
c
t
.
m
i
t
.
/
e
d
u
o
p
m
i
/
l
a
r
t
i
c
e
-
p
d
f
/
d
o
i
/
i
/
/
.
1
0
1
1
6
2
o
p
m
_
a
_
0
0
0
2
7
2
0
6
6
9
7
6
o
p
m
_
a
_
0
0
0
2
7
p
d
/
.
i
f
b
y
g
u
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
Most People Keep Their Word Woike and Kanngiesser
for constructive criticism and feedback on the project, and we thank Anna Dania Esch and
Manuel Canella for coding the open answers in Study 1a. We thank Deborah Ain for editing
this manuscript.
FUNDING INFORMATION
Patricia Kanngiesser, Volkswagen Foundation, Award ID: 89611.
AUTHOR CONTRIBUTIONS
JKW: Conceptualization: Equal; Data curation: Lead; Formal analysis: Lead; Methodology:
Equal; Software: Lead; Visualization: Lead; Writing – Original Draft: Equal. PK: Conceptu-
alization: Equal; Methodology: Equal; Writing – Original Draft: Equal.
REFERENCES
Almaatouq, A., Krafft, P., Dunham, Y., Rand, D. G., & Pentland,
A. (2018). Turkers of the world unite: Multilevel in-group bias
amongst crowdworkers on Amazon Mechanical Turk. SSRN.
doi:10.2139/ssrn.3122520
Amir, O., Rand, D. G., & Gal, Y. K. (2012). Economic games on the
internet: The effect of $1 stakes. PloS ONE, 7, e31461. doi:10. 1371/journal.pone.0031461 Austin, J. L. (1975). How to do things with words. New York, NY: Oxford University Press. Belot, M., Bhaskar, V., & van de Ven, J. (2012). Can observers pre- dict trustworthiness? Review of Economics and Statistics, 94, 246–259. doi:10.1162/REST_a_00146 Bhattacharya, P., & Sengupta, A. (2016). Promises and guilt. SSRN. doi:10.2139/ssrn.2904957 Bicchieri, C. (2002). Covenants without swords: Group identity, norms, and communication in social dilemmas. Rationality and Society, 14, 192–228. doi:10.1177/1043463102014002003 Bicchieri, C., & Lev-On, A. (2007). Computer-mediated commu- nication and cooperation in social dilemmas: An experimental analysis. Politics, Philosophy & Economics, 6(2), 139–168. doi:10. 1177/1470594X07077267 Burleigh, T., Kennedy, R., & Clifford, S. (2018). How to screen out VPS and international respondents using Qualtrics: A protocol. SSRN. doi:10.2139/ssrn.3265459 Carlsson, F., & Kataria, M. (2018). Do people exaggerate how happy they are? Using a promise to induce truth-telling. Oxford Eco- nomic Papers, 70, 784–798. doi:10.1093/oep/gpy003 Carlsson, F., Kataria, M., Krupnick, A., Lampi, E., Löfgren, Å., Qin, P., & Sterner, T. (2013). The truth, the whole truth, and nothing but the truth—a multiple country test of an oath script. Journal of Economic Behavior & Organization, 89, 105–121. doi:10.1016/ j.jebo.2013.02.003 Casella, A., Kartik, N., Sanchez, L., & Turban, S. (2018). Commu- nication in context: Interpreting promises in an experiment on competition and trust. Proceedings of the National Academy of Sciences of the United States of America, 115, 933–938. doi:10. 1073/pnas.1714171115 Charness, G., & Dufwenberg, M. (2006). Promises and partnership. Econometrica, 74, 1579–1601. doi:10.1111/j.1468-0262.2006. 00719.x Charness, G., & Dufwenberg, M. (2010). Bare promises: An experi- ment. Economics Letters, 107, 281–283. doi:10.1016/j.econlet. 2010.02.009 Chen, J., & Houser, D. (2017). Promises and lies: Can observers de- tect deception in written messages. Experimental Economics, 20, 396–419. doi:10.1007/s10683-016-9488-x Conrads, J., & Reggiani, T. (2017). The effect of communication channels on promise-making and promise-keeping: Experimen- tal evidence. Journal of Economic Interaction and Coordination, 12, 595–611. doi:10.1007/s11403-016-0177-9 Crawford, J. R., & Henry, J. D. (2004). The Positive and Nega- tive Affect Schedule (PANAS): Construct validity, measurement properties and normative data in a large non-clinical sample. British Journal of Clinical Psychology, 43, 245–265. doi:10.1348/ 0144665031752934 Cumming, G. (2012). Understanding the new statistics: Effect sizes, confidence intervals, and meta-analysis. New York, NY: Routledge. Cumming, G. (2014). The new statistics: Estimation for better research. Retrieved from http://www.thenewstatistics.com Di Bartolomeo, G., Dufwenberg, M., Papa, S., & Passarelli, F. (2018). Promises, expectations & causation. Games and Economic Behavior, 113, 137–146. doi:10.1016/j.geb.2018.07.009 Ederer, F., & Stremitzer, A. (2017). Promises and expectations. Games and Economic Behavior, 106, 161–178. doi:10.1016/j. geb.2017.09.012 Ellingsen, T., & Johannesson, M. (2004). Promises, threats and fair- ness. The Economic Journal, 114, 397–420. doi:10.1111/j.1468- 0297.2004.00214.x Engel, C. (2011). Dictator games: A meta study. Experimental Eco- nomics, 14, 583–610. doi:10.1007/s10683-011-9283-7 Fetchenhauer, D., & Dunning, D. (2010). Why so cynical? Asym- metric feedback underlies misguided skepticism regarding the trustworthiness of others. Psychological Science, 21, 189–193. doi:10.1177/0956797609358586 Harrison, G. W., Martínez-Correa, J., Swarthout, J. T., & Ulm, E. R. (2017). Scoring rules for subjective probability distributions. Jour- nal of Economic Behavior & Organization, 134, 430–448. doi:10. 1016/j.jebo.2016.12.001 Hertwig, R., & Grüne-Yanoff, T. (2017). Nudging and boosting: Steer- ing or empowering good decisions. Perspectives on Psychologi- cal Science, 12, 973–986. doi:10.1177/1745691617702496 OPEN MIND: Discoveries in Cognitive Science 87 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . / e d u o p m i / l a r t i c e - p d f / d o i / i / / . 1 0 1 1 6 2 o p m _ a _ 0 0 0 2 7 2 0 6 6 9 7 6 o p m _ a _ 0 0 0 2 7 p d . / i f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 Most People Keep Their Word Woike and Kanngiesser Heyman, G. D., Fu, G., Lin, J., Qian, M. K., & Lee, K. (2015). Eliciting promises from children reduces cheating. Journal of Experimen- tal Child Psychology, 139, 242–248. doi:10.1016/j.jecp.2015. 04.013 Hobbes, T. (2016). Leviathan (M. Missner, Ed.). New York, NY: Routledge (Original work published 1651). Houser, D., & Xiao, E. (2015). House money effects on trust and reciprocity. Public Choice, 163, 187–199. doi:10.1007/s11127- 014-0218-7 Iriberri, N., & Rey-Biel, P. (2013). Elicited beliefs and social informa- tion in modified dictator games: What do dictators believe other dictators do? Quantitative Economics, 4, 515–547. doi:10.3982/ QE135 Ismayilov, H., & Potters, J. (2016). Why do promises affect trustwor- thiness, or do they? Experimental Economics, 19, 382–393. doi:10. 1007/s10683-015-9444-1 Ismayilov, H., & Potters, J. (2017). Elicited vs. voluntary promises. Journal of Economic Psychology, 62, 295–312. doi:10.1016/j. joep.2017.07.005 Jacquemet, N., Luchini, S., Rosaz, J., & Shogren, J. F. (2019). Truth telling under oath. Management Science, 65, 426–438. doi:10. 1287/mnsc.2017.2892 Jiang, T. (2013). Cheating in mind games: The subtlety of rules mat- ters. Journal of Economic Behavior & Organization, 93, 328–336. doi:10.1016/j.jebo.2013.04.003 Kanngiesser, P., Köymen, B., & Tomasello, M. (2017). Young children mostly keep, and expect others to keep, their promises. Journal of Experimental Child Psychology, 159, 140–158. doi:10.1016/j. jecp.2017.02.004 Kataria, M., & Winter, F. (2013). Third party assessments in trust problems with conflict of interest: An experiment on the effects of promises. Economics Letters, 120, 53–56. doi:10.1016/j.econlet. 2013.03.028 Kennedy, R., Clifford, S., Burleigh, T., Waggoner, P., & Jewell, R. (2018). How Venezuela’s economic crisis is undermining social science research—about everything. Washington Post. Retrieved from https://www.washingtonpost.com/news/monkey-cage/wp/2018/ 11/07/how-the-venezuelan-economic-crisis-is-undermining-social– science-research-about-everything-not-just-venezuela Kolodny, N., & Wallace, R. J. (2003). Promises and practices revis- ited. Philosophy & Public Affairs, 31(2), 119–154. doi:10.1111/ j.1088-4963.2003.00119.x Larney, A., Rotella, A., & Barclay, P. (2019). Stake size effects in ultimatum game and dictator game offers: A meta-analysis. Orga- nizational Behavior and Human Decision Processes, 151, 61–72. doi:10.1016/j.obhdp.2019.01.002 Lev-On, A., Chavez, A., & Bicchieri, C. (2010). Group and dyadic communication in trust games. Rationality and Society, 22(1), 37–54. doi:10.1177/1043463109337100 Matheson, J. E., & Winkler, R. L. (1976). Scoring rules for continuous probability distributions. Management Science, 22, 1087–1096. doi:10.1287/mnsc.22.10.1087 Mischkowski, D., Stone, R., & Stremitzer, A. (2016). Promises, ex- pectations, and social cooperation. Harvard Law School John M. Olin Center Discussion Paper, 887. Moss, A. J., & Litman, L. (2018). After the bot scare: Understanding what’s been happening with data collection on MTurk and how to stop it [blog post]. Retrieved from https://blog.turkprime.com/ after-the-bot-scare-understanding-whats-been-happening-with- data-collection-on-mturk-and-how-to-stop-it Orbell, J. M., van de Kragt, A. J., & Dawes, R. M. (1988). Explain- ing discussion-induced cooperation. Journal of Personality and Social Psychology, 54, 811–819. doi:10.1037/0022-3514.54.5. 811 Ostrom, E., Walker, J., & Gardner, R. (1992). Covenants with and without a sword: Self-governance is possible. American Political Science Review, 86, 404–417. doi:10.2307/1964229 Oxoby, R. J., & Spraggon, J. (2008). Mine and yours: Property rights in dictator games. Journal of Economic Behavior & Organization, 65, 703–713. doi:10.1016/j.jebo.2005.12.006 Pronin, E. (2008). How we see ourselves and how we see others. Science, 320, 1177–1180. doi:10.1126/science.1154199 Raihani, N. J., Mace, R., & Lamba, S. (2013). The effect of $1, $5 and $10 stakes in an online dictator game. PloS ONE, 8, e73131.
doi:10.1371/journal.pone.0073131
Sally, D. (1995). Conversation and cooperation in social dilemmas:
A meta-analysis of experiments from 1958 to 1992. Rationality
and Society, 7, 58–92. doi:10.1177/1043463195007001004
Scanlon, T. (1990). Promises and practices. Philosophy & Public Af-
fairs, 19, 199–226.
Searle, J. R. (1969). Speech acts: An essay in the philosophy of lan-
guage. Cambridge, England: Cambridge University Press.
Semuels, A. (2018). The internet is enabling a new kind of poorly
paid hell. The Atlantic. Retrieved from https://www.theatlantic.
com/business/archive/2018/01/amazon-mechanical-turk/551192/
Shu, L. L., Mazar, N., Gino, F., Ariely, D., & Bazerman, M. H. (2012).
Signing at the beginning makes ethics salient and decreases dis-
honest self-reports in comparison to signing at the end. Proceed-
ings of the National Academy of Sciences of the United States of
America, 109, 15197–15200. doi:10.1073/pnas.1209746109
Turmunkh, U., van den Assem, M. J., & van Dolder, D. (2019). Mal-
leable lies: Communication and cooperation in a high stakes TV
game show. Management Science. Advance online publication.
doi:10.1287/mnsc.2018.3159
Vanberg, C. (2008). Why do people keep their promises? An exper-
imental test of two explanations. Econometrica, 76, 1467–1480.
doi:10.3982/ECTA7673
van den Assem, M. J., van Dolder, D., & Thaler, R. H. (2012). Split
or steal? Cooperative behavior when the stakes are large. Man-
agement Science, 58, 2–20. doi:10.1287/mnsc.1110.1413
Wang, T. H., & Katzev, R. D. (1990). Group commitment and
resource conservation: Two field experiments on promoting
recycling. Journal of Applied Social Psychology, 20, 265–275.
doi:10.1111/j.1559-1816.1990.tb00411.x
Woike, J. K., & Kanngiesser, P. (2019a, May 2). People keep their
word–2019. doi:10.17605/OSF.IO/DXEPU
Woike, J. K., & Kanngiesser, P. (2019b). Supplemental material for
“Most people keep their word rather than their money.” Open
Mind: Discoveries in Cognitive Science, 3. doi:10.1162/opmi_a_
00027
OPEN MIND: Discoveries in Cognitive Science
88
l
D
o
w
n
o
a
d
e
d
f
r
o
m
h
t
t
p
:
/
/
d
i
r
e
c
t
.
m
i
t
.
/
e
d
u
o
p
m
i
/
l
a
r
t
i
c
e
-
p
d
f
/
d
o
i
/
i
/
/
.
1
0
1
1
6
2
o
p
m
_
a
_
0
0
0
2
7
2
0
6
6
9
7
6
o
p
m
_
a
_
0
0
0
2
7
p
d
/
.
i
f
b
y
g
u
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3