Measuring Online Debaters’ Persuasive Skill from Text over Time

Measuring Online Debaters’ Persuasive Skill from Text over Time

Kelvin Luu1 Chenhao Tan2 Noah A. Smith1,3
1Paul G. Allen School of Computer Science & Engineering, University of Washington
2Department of Computer Science, University of Colorado Boulder
3Allen Institute for Artificial Intelligence
{kellu,nasmith}@cs.washington.edu chenhao.tan@colorado.edu

Abstract

Online debates allow people to express their
persuasive abilities and provide exciting oppor-
tunities for understanding persuasion. Prior
studies have focused on studying persuasion
in debate content, but without accounting
for each debater’s history or exploring the
progression of a debater’s persuasive ability.
We study debater skill by modeling how
participants progress over time in a collection
of debates from Debate.org. We build on
a widely used model of skill in two-player
games and augment it with linguistic features
of a debater’s content. We show that online
debaters’ skill
levels do tend to improve
over time. Incorporating linguistic profiles
leads to more robust skill estimation than
winning records alone. Notably, we find that
an interaction feature combining uncertainty
cues (hedging) with terms strongly associated
with either side of a particular debate (fightin’
words) is more predictive than either feature
on its own, indicating the importance of fine-
grained linguistic features.

1

Introduction

Persuasion is an important skill with prevalent
use. Nearly every kind of social encounter, from
formal political debates to casual conversations,
can include attempts to convince others. What
linguistic phenomena are associated with higher
levels of persuasive skill? How do people develop
this skill over time? Online debate communities
offer an opportunity to investigate these questions.
These communities feature users who participate
in multiple debates over their period of engage-
ment. Such debates involve two parties who
willingly and formally present divergent opinions
before an audience. Unlike other media of
persuasion, such as letters to politicians, there is a

537

clear signal, a win or loss, indicating whether or
not a debater was successful against the adversary.
This work aims to quantify the skill level of
each debater in an online community and also
investigates what factors contribute to expertise.

Although persuasion has generated interest
in the natural language processing community,
most researchers have not tried to quantify the
persuasiveness of a particular speaker. Instead,
they estimate how persuasive a text
is, using
linguistic features such as the author’s choice
of wording or how they interact with the audience
(Tan et al., 2014, 2016; Althoff et al., 2014;
Danescu-Niculescu-Mizil et al., 2012). Previous
research has also established that debaters’ content
and interactions both contribute to the success of
the persuader (Tan et al., 2016; Zhang et al.,
2016; Wang et al., 2017). These works have
found textual factors that contribute to a debater’s
success, but they do not emphasize the role of the
individual debater.

There has been some recent work on studying
individual debaters. Durmus and Cardie (2019)
analyze users and find that a user’s success
and improvement depend on their social network
features. We take another approach by estimating
each user’s skill level in each debate they par-
ticipate in by considering their debate history.
These estimates reveal features correlated with
skill and the importance of particular debates over
time. Our study is based on debates from an
online debate forum, Debate.org, introduced
by Durmus and Cardie (2018) and discussed in §2.
This Web site is composed of primarily text-
based debates and attracts a large number of users
to debate regularly.

Our model of skill builds on the Elo (1978)
rating system, designed for rating players in
two-player games (§3). Our preliminary analysis
using Elo scores suggests that user skill is not
static; debaters in the Debate.org forum tend

Transactions of the Association for Computational Linguistics, vol. 7, pp. 537–550, 2019. https://doi.org/10.1162/TACL a 00281
Action Editor: Jordan Boyd-Graber. Submission batch: 11/2018; Revision batch: 4/2019; Published 9/2019.
c(cid:2) 2019 Association for Computational Linguistics. Distributed under a CC-BY 4.0 license.

l

D
o
w
n
o
a
d
e
d

f
r
o
m
h

t
t

p

:
/
/

d
i
r
e
c
t
.

m

i
t
.

e
d
u

/
t

a
c
l
/

l

a
r
t
i
c
e

p
d

f
/

d
o

i
/

.

1
0
1
1
6
2

/
t

l

a
c
_
a
_
0
0
2
8
1
1
9
2
3
1
9
9

/

/
t

l

a
c
_
a
_
0
0
2
8
1
p
d

.

f

b
y
g
u
e
s
t

t

o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3

to improve with practice. We extend Elo for
the debate setting using linguistic features. We
decompose our family of models into two design
choices that align with the questions we hope to
answer: 1) the features we use and 2) how we
might choose to aggregate those features from
past debates.

To validate our skill estimates, we introduce a
forecasting task (§4). Previous work predicted
the winner of a debate using the text of the
current debate. In contrast, we aim to predict
winners using our skill estimates before the
debate (ignoring the current debate’s content).
This design ensures that we are modeling skill of
the debater, as inferred from past performance,
not the idiosyncrasies of a particular debate.

We also investigate the predictive power of
our estimates through an analysis of the results
(§6). We show that our full model outperforms
the baseline Elo model, approaching the accuracy
of an oracle that does use the text of the current
debate. Our ablation studies reveal that the co-
occurrence of phrases that indicate uncertainty
or doubt (hedges) with words that are strongly
associated with one debater or the other (fightin’
words), is an effective predictor. Moreover, we
find that not all past debates are equally useful
for prediction: more recent debates are more
indicative of the user’s current level of expertise.
This adds support to our conjecture that individual
debaters tend to improve through the course of
their time debating.

Finally, we track the linguistic tendencies of
each debater over the course of their debating
history. We show that several features such as
the length of their turns and the co-occurrence of
hedges and fightin’ words increase over time for
the best debaters, but stay static for those with less
skill. These findings give further evidence that
debaters improve over time.

2 Data

We use both debate and user skill data from the
Debate.org data set introduced by Durmus and
Cardie (2018). Any registered user on the Web site
can initiate debates with others or vote on debates
conducted by others.

2.1 Mechanism of Debate.org

Registered users can create a debate under a topic
of their choosing. The person initiating the debate,

Figure 1: An example of the Debate.org voting
system.

called the instigator, fixes the debate’s number
of rounds (2–10) and chooses the category (e.g.,
politics, economics, or music) at the start of the
debate. The instigator then presents an opening
statement in the first round and waits for another
user, the contender, to accept the debate and
write another opening statement to complete the
first round.1 We define a debater’s role as being
either the instigator or contender.

To determine

the debate winner, other
Debate.org users vote after the debate ends.
In this phase, voters mark who they thought
performed better in each of seven categories (see
Figure 1).2 This phase can last between 3 days and
6 months, depending on the instigator’s choice at
the debate’s creation. After this period, the debater
who received the most points wins the debate. We
record in our data set the textual information,
participants, and voting records for each debate.

2.2 Definition of Winning

The Debate.org voting system lets us define
a ‘‘win’’ in various ways (see Figure 1). In this
study, we would like to model how convincing
each debater is. One approach might be based on
Oxford-style debates like the IQ2 data set from
Zhang et al. (2016), in which scores are based on
the number of audience members who changed
their minds as a result of the debate. We found
here that the majority of voters do not deviate
from their stances before the debate, and most
voters tend to vote for who they already agree

1Although many debaters use their first round to make an
opening statement, some use it only to propose and accept
debates. If the first turn in an n-round debate is under 250
words, we merge each debater’s first two turns and treat the
debate as an (n − 1)-round debate.

2Another system of voting lets voters choose who they
thought performed better over the entire debate. Although
this appears in the data set, we do not use it in this paper.

538

l

D
o
w
n
o
a
d
e
d

f
r
o
m
h

t
t

p

:
/
/

d
i
r
e
c
t
.

m

i
t
.

e
d
u

/
t

a
c
l
/

l

a
r
t
i
c
e

p
d

f
/

d
o

i
/

.

1
0
1
1
6
2

/
t

l

a
c
_
a
_
0
0
2
8
1
1
9
2
3
1
9
9

/

/
t

l

a
c
_
a
_
0
0
2
8
1
p
d

.

f

b
y
g
u
e
s
t

t

o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3

Completed Debates
Completed & Convincing
Full Filtered

#Users
42,424
21,753
1,284

#Debates
77,595
29,209
4,486

Table 1: Description of the debate.org data
set from Durmus and Cardie (2018) and the
filtered data sets. Full Filtered is a subset of
Completed & Convincing that requires that
participants of each debate have engaged in
five or more debates. We use Full Filtered for
the remainder of our analysis.

system, which serves as both a baseline and the
basis for our final model. In our initial data
analysis, we use Elo scores to define an upset
as a debate where a weaker debater wins over
a stronger one. By measuring the rate at which
upsets occur over a debater’s lifetime, we make an
initial observation that debaters seem to improve
with experience.

3.1 Elo Model

Elo originated as a ranking system for chess
players; it has been adapted to other domains, such
as video games. It is one of the standard methods
to rate players of a two-player, zero-sum game
(Elo, 1978).3 Elo assigns positive integer-valued
scores, typically below 3,000, with higher values
interpreted as ‘‘more skill.’’ The difference in
the scores between two debaters under a logistic
model is used as an estimate of the probability
each debater will win. For example, consider a
debate between A and B. A has an Elo rating
of RA = 1900, and B has an Elo rating of
RB = 2000. Using the Elo rating system, pA,
the probability that A wins is4

pA =

1
1 + 100.0025(RB −RA)

=

1
1 + 100.25

≈ 0.36

(1)

Ratings are updated after every debate, with
the winner (equal to A or B) gaining (and the
loser losing) Δ = 32(1 − pW ) points. (32 is an
arbitrary scalar; we follow non-master chess in
selecting this value.)

Note that the magnitude of the change cor-
responds to how unlikely the outcome was.

Figure 2: Complementary cumulative distribution func-
tions (1 − CDF) for the total number of debates a
user finished. The blue line tracks only debaters who
engaged in and successfully concluded at least one
debate, and the red line tracks users who have finished
at least five debates in the filtered data set. The right
plot similarly shows the complementary cumulative
distribution for the number of votes given per debate
for all debates and the filtered data set.

with. Therefore, we count the number of times
each debater was rated as more convincing to a
voter despite presenting a viewpoint that the voter
disagreed with before the debate. The debater with
the higher count of such votes is considered the
‘‘winner’’ in the remainder of this paper.

2.3 Data Set Statistics

From an unfiltered set of 77,595 debates, we
remove all debates where a user forfeits, that lack
a winner (§2.2), or that do not have a typical
voting style with seven categories, leaving 29,209
completed debates.

We record the number of debates that each user
completes (see Figure 2, left). We find that the
quantity of debates per user follows a heavy-tailed
distribution where most users do not participate
in more than one debate. For the remainder of the
work, we focus on the 1,284 (out of 42,424 total
users) who have completed five or more of the
debates described above. This leaves us with 4,486
debates where the participants have completed at
least five debates (see Table 1).

We also record the number of votes that each
debate attracted (see Figure 2, right). This number
also follows a heavy-tailed distribution.

3 Expertise Estimation

In order to explore debater expertise and discover
what contributes to a user’s expertise over their
time on Debate.org, we begin with a conven-
tional approach to skill estimation, the Elo rating

3The Elo model is a special case of the Bradley-Terry

model (Bradley and Terry, 1952).

4The base of the exponent, 10, and the multiplicative
factor on the difference RA − RB, 0.0025, are typically used
in chess.

539

l

D
o
w
n
o
a
d
e
d

f
r
o
m
h

t
t

p

:
/
/

d
i
r
e
c
t
.

m

i
t
.

e
d
u

/
t

a
c
l
/

l

a
r
t
i
c
e

p
d

f
/

d
o

i
/

.

1
0
1
1
6
2

/
t

l

a
c
_
a
_
0
0
2
8
1
1
9
2
3
1
9
9

/

/
t

l

a
c
_
a
_
0
0
2
8
1
p
d

.

f

b
y
g
u
e
s
t

t

o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3

Although the Elo ratings traditionally take only a
win or loss as input, there have been adjustments
to account for the magnitude of victory. One
such method would be to use the score difference
between the two players to adjust the Elo gain
(Silver, 2015). If we let SA and SB be scores for
A and B, respectively, the modified gain Δ(cid:4) is

Δ(cid:4) = log(|SA − SB| + 1) × Δ

(2)

Under this model, we represent a user’s history
and skill level as a single scalar, that is, their Elo
rating. The Elo system ignores all other features,
which include the style a debater uses in the
debates and the content of their argument. We
therefore view this model as a baseline and extend
it.

3.2 Do Debaters Get Better Over Time?

Our initial data analysis uses the Elo model to
investigate whether users improve over time at all
in the first place. An increase in Elo score can be
seen as the rating system merely becoming more
accurate with another sample or an actual increase
in the player’s skill. We wish to show that there
is no fixed rating for a user, but rather that the
rating is a moving target. A counterhypothesis
is then that the users’ skill levels do not change
despite their activity on Debate.org. If true,
then we would expect their final Elo rating to be
also indicative of their skill level at the beginning
of their Debate.org activity.

To test this hypothesis, we first define an upset
as a debate where pW < τ , that is, where the winner of the debate was estimated (under Elo) to win with low probability (set using threshold τ ). If users tend to have static skill levels, then participating in many debates does not affect debaters’ skill, but simply provides more samples for measuring their skill. It follows that we would expect upsets to occur at the same rate early and late in each one’s career given debaters’ static skill. In this analysis, we calculate pA and pB for each debate using A’s and B’s final Elo ratings, which we take to be the most accurate estimate of their (presumed static) skill levels.5 To operationalize ‘‘early’’ and ‘‘late,’’ we divide a user’s debates into quintiles by time, 5In a forecasting analysis like the one in §4, this would be inappropriate, as it uses ‘‘future’’ information to define upsets. This notion of upset is not used in the forecasting task. Figure 3: Upset rates (aggregated across users) across history quintiles, τ = 0.45. The error bars represent the 95% confidence intervals. comparing the upset rate in different quintiles. In this analysis, we only consider users with at least ten debates (N = 4, 420).6 We see a downward trend in the upset rate (Figure 3). In particular, the first and last quintiles show a statistically significant difference under a paired t-test (p < .001), meaning that a user’s final Elo score is not a good measure of skill at earlier times. We take this finding as suggestive that users of Debate.org adapt as they participate in more debates. 4 Predicting Expertise using Earlier Debates 1 , . . . , dA Our aim is to estimate a debater A’s persuasive ability after observing a series of debates they participated in (denoted dA t−1 if we are estimating ability just before the tth debate). We wish to take into account the content of those debates (not merely their outcomes, as in Elo), so as to understand what factor reveals a debater’s skill levels. Drawing inspiration from Elo’s interpretation as a score that can be used to predict each debater’s probability of winning (Eq. 1), we formulate a prediction task: estimate pA for A’s tth debate, given dA t−1 and also the opponent’s debate history (which might be of a different length). 1 , . . . , dA 6In addition to requiring 10 debates instead of 5 per user, as in the rest of our analysis, we also do not subject the opponents to the same requirements. 540 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . e d u / t a c l / l a r t i c e - p d f / d o i / . 1 0 1 1 6 2 / t l a c _ a _ 0 0 2 8 1 1 9 2 3 1 9 9 / / t l a c _ a _ 0 0 2 8 1 p d . f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 By observing debate outcomes alongside the two participants’ histories, we can estimate the parameters of such a probability model. Elo provides a baseline; rather than opaque scores associated with individual users at different times, we seek to explain the probability of winning through linguistic features of past debate content. Unlike previous work (Zhang et al., 2016; Potash and Rumshisky, 2017; Tan et al., 2018), we do not use the content of the current debate (dA t ) to predict its outcome; rather, we forecast the outcome of the debate as if the debate has not yet occurred. To our knowledge, this is the first work to derive scores for current skill levels based on observing participants’ behavior over time. In the remainder of this section, we discuss features of past debates and ways of aggregating them. 4.1 Incorporating a Linguistic Profile into Elo Elo scores are based entirely on wins and losses; they ignore debate content. We seek to incorporate content into expertise estimation by using linguistic features. If we modify the exponential base in Equation 1 from 10 to e, we can view Elo probabilities (e.g., pA) as the output of a logistic regression model with one feature (the score difference, RA − RB) whose weight is 0.0025; that is, pA = σ(0.0025 · (RA − RB)). It is straightforward to incorporate more features, letting the contender rounds of that debate were written by his opponent). Table 2 shows the full list of features. Hedging with fightin’ words. We introduce one novel feature for our work: the hedging with fightin’ words. ‘‘Fightin’ words’’ refer to words found using a method, introduced by Monroe et al. (2008), which seeks to identify words (or phrases) most strongly associated with one side or another in a debate or other partisan discourse.7 We are interested in situations where debaters evoke fightin’ words (their own, or their opponents’) with an element of uncertainty or doubt. We use each debater’s top 20 fightin’ words (unigrams or bigrams) as features, following Zhang et al., 2016, who found this feature useful in predicting winners of Oxford-style televised debates. We also count cooccurences of fightin’ words with hedge phrases like ‘‘it could be possible that’’ or ‘‘it seems that.’’ An example of this conjoined feature is found in the utterance ‘‘Could you give evidence that supports the idea that married couples are more likely to be committed to [other tasks]?’’, where hedge phrases are emboldened and brackets denote fightin’ words (which are selected separately within each debate). We use a list of hedging cues curated by Tan et al. (2016) and derived from Hyland (1996) and Hanauer et al. (2012). The conjoined feature is the count of the user’s sentences in a debate where a fightin’ word cooccurs with a hedge phrase in a sentence. pA = σ (w · (RA − RB)) (3) 4.3 Aggregating Earlier Debates where w is a vector of weights and RU is user U ’s ‘‘profile,’’ a vector of features derived from past debates. In this work, the linguistic profiles are designed based on extant theory about the linguistic markers of persuasion, and the vectors are weighted averages of features derived from earlier debates. 4.2 Features We select features discussed in prior work as the basis for our linguistic profiles (Tan et al., 2016, 2018; Zhang et al., 2016). We extract these measurements from each of the user’s debates. For a given debate and user, we calculate these values over the rounds written by the user. For example, if we were interested in a debate by Alex as the instigator, we would only calculate features from the instigator rounds of that debate (since Because we consider the full history of a debater when estimating their skill level, we opt to ag- gregate the textual features over each debate. We do so by taking a weighted sum of the feature vectors of the previous debates. We consider four weighting schemes, none of which have free parameters, to preserve interpretability. Let f be any one of the features in the linguistic profile, a function from a single debate to a scalar. debates (cid:2) t−1 i=1 1. Exponential growth: The most indicative of recent are most skill, f (dA i ) 2t−i . We take this to be the most intuitive choice, experimentally comparing against the alternatives below to confirm this intuition. 7The method estimates log-odds of words given a side, with Dirichlet smoothing, and returns the words with the highest log-odds for each side. 541 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . e d u / t a c l / l a r t i c e - p d f / d o i / . 1 0 1 1 6 2 / t l a c _ a _ 0 0 2 8 1 1 9 2 3 1 9 9 / / t l a c _ a _ 0 0 2 8 1 p d . f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 Feature Description Elo Score Length Part of speech Flesch reading ease Emotional words Links Traditional Elo score calculated and updated. Updated tradition- ally, not averaged as in §4.3. Number of words uttered in the debate. each noun, verb, Count of adjective, preposition, adverb, or pronoun from the participant in the entire debate. this user Measure of readability given the number of sentences in a document and the number of words in each sentence (Kincaid et al., 1975). Cues that indicate a positive or negative emotion (Tausczik and Pennebaker, 2010). Links to external websites out- side of debate.org. This fea- ture operationalizes the number of sources a debater used. Questions The number of questions the user asked in the debate. Quotations The number of quotations the user included in the debate. (Hyland, Hedging The number of phrases that soften a statement by adding 1996; uncertainty Hanauer et al., 2012). The number of instances of words most strongly associated with either debater (Monroe et al., 2008). Fightin’ words H∧FW The number of cooccurences of hedging and fightin’ words, described in §4.2. n/a ↑↑↑ Nouns (↑↑↑) Adjec- tives (↑↑↑) ↑↑ Pos (↑↑↑) Neg (↑↑↑) ↓↓↓ ↑↑↑ ↑↑ levels. Aside from Elo, Table 2: Debate-level features used in estimating skill the features are a part of the user’s linguistic profile. The third column represents statistical significance levels in compar- ing winners and losers’ features (independently) with Bonferroni correction: ↑ is p < 0.05, ↑↑ is p < 0.01, ↑↑↑ is p < 0.001. 2. Simple average: Each earlier debate’s equally, is weighted feature (cid:2) vector t−1 i=1 f (dA i ). 1 t−1 3. Exponential decay: The first debates are (cid:2) most indicative of skill, t−1 i=1 f (dA i ) 2i . 542 4. Last only: Only the single most recent t−1) (an extreme version debate matters, f (dA of ‘‘exponential growth’’). In each variation of our method, some or all of the linguistic profile features are aggregated (using one of the four weighted averages), then applied to predict debate outcomes through logistic re- gression. We note that if our sole aim were to maximize predictive accuracy, we might explore much richer linguistic profiles, perhaps learning word embeddings for the task and combining them using neural networks and enabling interactions between a debate’s two participants’ profiles.8 In this study, we seek to estimate skill but also to understand it, so our focus remains on linear models. 5 Experimental Setup We validate our skill models with a binary classification task, specifically forecasting which of two debaters will win a debate (without looking at the content of the debate). Here we remove any debates where someone forfeits or there is no winner. We then considered debates where each debater has completed at least five of the re- maining debates. As discussed in §2.2, the winner is taken to be the debater receiving the most ‘‘more convincing argument’’ votes from observers who did not initially agree with them. We create four training/evaluation splits of the data, using debates from 2013, 2014, 2015, and 2016 as evaluation sets (i.e., development and test) and debates prior to the evaluation year for training. Figure 4 shows the number of debates in each split. We note that our training sets are cumulative. For example, if we were to test on 2015 data, we would use the 2012, 2013, and 2014 training as training data. Because we do not test on 2012 data or train on 2016 data due to the low number of debates before 2012 and after 2016, we treat the whole of 2012 as training data and 2016 as development and test. We report the accuracy for each run. We compare several predictors:9 • Full model: Our model with all features, as described in §4, and (except where otherwise 8Indeed, in preliminary experiments we did explore using a recurrent neural network instead of a fixed weighted average, but it did not show any benefit, perhaps owing to the relatively small size of our data set. 9We use (cid:2)2 regularization in our models with the linguistic profile features. l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . e d u / t a c l / l a r t i c e - p d f / d o i / . 1 0 1 1 6 2 / t l a c _ a _ 0 0 2 8 1 1 9 2 3 1 9 9 / / t l a c _ a _ 0 0 2 8 1 p d . f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 Figure 4: We split data into training/development/test based on year. This chart shows the number of debates in each subset of the data. We note that, for training, we use all the training debates from previous years (e.g., if we were to test on 2015, we would train using the training splits from 2012, 2013, and 2014) Each instance in this figure corresponds to a debater. stated) the exponential growth weighting. This model combines linguistic profiles from earlier debates with a conventional Elo score. • Full model with point difference: Our full model as described above, but we scale the Elo gain by the point difference as described in Equation 2. • Linguistic profile only: Our model with exponential growth weighting (except where otherwise stated), but ablating the Elo feature. This model is most similar to those found in prior literature (Zhang et al., 2016; Tan et al., 2018; Wang et al., 2017). • Elo: The prediction is based solely on the Elo score calculated just before the debate. This is equivalent to ablating the linguistic profiles from our model. • Final Elo oracle: The prediction is based solely on the two debaters’ final Elo scores (i.e., using all debates from the past, present, and future). • Current debate text oracle: A model that uses the linguistic profile derived just from the current debate. Although this model is most similar to previous work, is not a fair estimate of skill (because it ignores past performance). We therefore view it as another oracle. it • Majority choice: A baseline that always predicts that the contender will win.10 10In this data set, contenders win nearly 59% of the time, a fact frequently discussed in the Debate.org Figure 5: Our results for the prediction task. Our full model outperforms the Elo baseline and approaches the current debate text oracle. 6 Results In this section, we first show that the expertise of a debater can be better estimated with the linguistic profile, and then analyze the contribution of dif- ferent components. We further examine the robust- ness of our results by controlling for additional variables. 6.1 Prediction Performance We first present our results with what we consider our best model, that is, our full model (with point differences), which consists of all features and uses the exponentially growing weight. Importance of linguistic features. We see from Figure 5 that the full model outperforms the standard Elo baseline. The gap between the two models suggests that the addition of the linguis- tic profile contributes to the performance of the model and therefore plays a useful role in skill esti- mation. Moreover, the linguistic profile only model shows that the linguistic profile features are not only useful, but have at least as strong predictive power as Elo alone. By only using the linguistic features aggregated over the course of a debater’s history, without knowing winning records, we can forecast at least as well as the Elo baseline. see http://ddo.wikia.com/wiki/ community; Contender Advantage. The contender advantage is sometimes attribted to having the ‘‘final say,’’ or to the fact that contenders choose the instigators they wish to debate. 543 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . e d u / t a c l / l a r t i c e - p d f / d o i / . 1 0 1 1 6 2 / t l a c _ a _ 0 0 2 8 1 1 9 2 3 1 9 9 / / t l a c _ a _ 0 0 2 8 1 p d . f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . e d u / t a c l / l a r t i c e - p d f / d o i / . 1 0 1 1 6 2 / t l a c _ a _ 0 0 2 8 1 1 9 2 3 1 9 9 / / t l a c _ a _ 0 0 2 8 1 p d . f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 Figure 6: Our bootstrap on the feature ablations. We record the average drop in performance across 100 iterations and tested on the 2016 test set. Higher means a larger drop in performance. Importance of multiple debates. We also note that our full model only performs slightly worse than the current debate text oracle despite the current debate text model directly observing the content of the debate. This result implies that using information from only previous debates has at least similar predictive strength to information from the debate at hand. Moreover, the large gap between the final Elo and current debate text oracles implies that a user’s skill is evidenced by more than the content of a single debate. These results further demonstrate the importance of accounting for debaters’ prior history. Magnitude of victory might matter. Our full model with the point difference scaled Elo gain does roughly as well or slightly better than our normal full model. As the focus of this paper is on incorporating linguistic profiles, we use the full model without the point difference scaling for analyses in the rest of the paper. 6.2 Feature Ablations We inspect the contribution of each feature by removing each one from the model. Then, for each feature, we perform a bootstrap test over the last year of data (trained on 2012–2015 data; tested on 2016). At each iteration, we sample 1,000 training examples to train on, but fix the test set across iterations. We then train our full model alongside several other models, each with a feature ablated, on the sample. We track the drop in performance between our full model and Figure 7: Comparison in performance for the four ways we aggregate features over time. each of our other models. We record the average performance over 100 iterations for comparison. From Figure 6, we find that removing Elo re- sults in the most severe drop in performance (5.8%). Ablating part-of-speech, negative emotion, and length from our model had a moderate ef- fect on performance. Surprisingly, we find that, although the H∧FW feature is the overlap between hedge cues and fightin’ words, the latter two features contribute very little to the performance of our model compared with H∧FW. 6.3 Combining Prior Debate Features As described in §4.3, we explore several ways of combining features over the past debates. By inspecting how these different aggregation functions might differ in performance, we hope to find out whether or not some debates are more important than others, if recency matters at all, and if some history is important at all. We do so by 1) giving the last debates more weight (exponential growth), 2) giving all weights equal weight (simple average), 3) giving the first debates more weight (exponential decay), and 4) giving all the weight to the last debate (last debate only). The rest of this section discusses these results, shown in Figure 7. More debates help. When using only the last debate’s features and ignoring all previous debates, our performance is initially very good in the years 2013 and 2014 (when our training sets are 544 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . e d u / t a c l / l a r t i c e - p d f / d o i / . 1 0 1 1 6 2 / t l a c _ a _ 0 0 2 8 1 1 9 2 3 1 9 9 / / t l a c _ a _ 0 0 2 8 1 p d . f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 Figure 9: Errors of the full model broken down by the history sizes of the users. The overall distribution is to the left and the error distribution is to the right. new, 2) there is only one new user, and 3) both debaters have been seen before. We expect that we will have better performance as our debates have more complete debate histories. We complete the analysis by looking at the 2016 test. Figure 8 shows that both models vastly improve when removing debates where users have no history and only improve as we require users to have longer histories. These results are consistent with our intuition that longer debate histories allow our models to better infer a debater’s skill. Moreover, our model generally outperforms the Elo baseline across the different history lengths, implying that our model’s performance the is consistent. We do note, however, gap (between the full model and Elo baseline) is consistent when there is only one new user and no new users. This could imply that our model, in addition to using longer histories more effectively, is more capable of estimating a de- bater’s skill against an unknown user compared with the normal Elo baseline. The slightly higher performance in the one-new-user case, compared with no-new-users, is likely due to the small size of each subset. that We extend our analysis of the effect of history lengths on skill estimation by inspecting the errors that our model makes. We are particularly interested in characterizing the mistakes that our model makes by the participant with the shorter debate history. For each debate that the full model predicts incorrectly, we keep count of history length of the user with the lower history count (e.g., if a debate has a user with a debate history of length 5 against a user with a debate history Figure 8: Our results separated by the number of times we have seen each participant before in training for the 2016 test set. The parenthetical numbers show how many debates are in the subset. smaller and histories shorter). However, as debate histories become richer in the later years, last- only’s performance drops in comparison to that of the growing weight aggregation. These results match our intuition: with more experienced users, considering more debates give us a better gauge of how a debater performs and a debater’s more recent debates give a better snapshot of the user’s current skill level. Later debates matter more. As hinted in our last-only results, giving the last debates more weight does best in all four years. In light of its performance compared with the simple mean and decaying weight settings, our results imply that not all debates contribute equally to a debater’s skill under our model. Indeed, our results show that the most recent debates are also the most important for estimating a debater’s skill rating. 6.4 Impact of the Length of History We finally examine how often our models have seen a user’s history impacts performance. Com- pared with the final Elo and current debate text oracle models, both the full and the Elo models suffer when both debaters have no prior history—these examples devolve into having to guess the majority class. Therefore, we further analyze these models by inspecting the performance on subsets of the test set pertaining to instances where 1) both users are 545 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . e d u / t a c l / l a r t i c e - p d f / d o i / . 1 0 1 1 6 2 / t l a c _ a _ 0 0 2 8 1 1 9 2 3 1 9 9 / / t l a c _ a _ 0 0 2 8 1 p d . f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 Figure 10: Our results for the prediction task on a restricted subset of the test set comprising only debates that are within both participants’ first 10 debates. Although our model does suffer in accuracy under this setting, it still outperforms both baselines and approaches the text-only oracle. of length 1, we would treat the debate having a history size of 1). We find in Figure 9 that a large percentage of the errors (51 % across all debates) come from debates where one user had a history length of 0 at test time. Moreover, the debates with at least one user with no history tend to contribute proportionally more to the error rate relative to the overall distribution. Conversely, the debates with users who have deeper histories of three debates or more tend to contribute less to the error rate. 6.5 Controlling for Other Variables We also inspect how well our model generalizes by measuring our model’s accuracy on subsets of the test set. In particular, we hope to see if high activity users or the topic of a debate affect our model’s performance. Controlling for prolific users. As seen in Figure 2, debater activity follows a heavy-tailed distribution where a minority of debaters perform the majority of debates. In our data set, 18.9% of debaters who have engaged in more than 10 debates participate in 66.9% of the debates. Thus, in order to validate that our model does not simply recognize the most active users, we test on a subset of the 2016 data where we cap the number of debates a user can participate in to 10 (142 debates). Figure 11: Our results for the prediction task broken down by topic of debate. Our full model maintains a consistent performance across all categories. From Figure 10, we see that the order is roughly the same even when we restrict the test set to require that both users have participated in fewer than 10 debates at the time of the debate. This implies that our model generalizes relatively well to most debaters regardless of their experience. We do note, however, that most of the models incur a large decrease in performance whereas the final Elo oracle maintains roughly the same accuracy. Our observations are invariant to debate topic. We also break down our test set by debate topic. We use the three most popular topics (politics, religion, and philosophy) and aggregate the rest of the categories as ‘other’. We see from Figure 11 that our model consis- tently outperforms both baselines and typically below the oracle models with the exception of ‘other’. These results are consistent with those found in Figure 5 and show that our model per- forms consistently across different debate topics. 7 Language Change over Time: The Best Improve and the Worst Stagnate Our findings from §3.2 and §6.3 imply that users tend to experience some change over time and that these changes are helpful for forecasting who in a debate is more likely to win. In this section, we explore whether debaters’ linguistic tendencies 546 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . e d u / t a c l / l a r t i c e - p d f / d o i / . 1 0 1 1 6 2 / t l a c _ a _ 0 0 2 8 1 1 9 2 3 1 9 9 / / t l a c _ a _ 0 0 2 8 1 p d . f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 Figure 12: Feature measurements averaged across each history quintile. We see that there is a general trend for those who eventually become the best debaters to improve on these measures while the bottom users stagnate. The error bars represent the 95% confidence intervals. change over time by tracking their use of features over the course of their debate history. We examine how language use changes over time for the best and worst users. To do so, we divide each user’s debate history into quintiles as in §3.2. For each quintile, we average the same features we use in the linguistic profile that we use in our model (Table 2). We take the top 100 and bottom 100 users ranked by our model to see how the trajectories change over time. Figure 12 shows that best and worst users have different linguistic preferences even from the beginning of their debating activities. The best debaters have a higher feature count in every case except for the Flesch reading ease score. However, the best users do seem to improve over time in length, H∧FW, use of emotional cues, and link use, which mostly correlates with our feature ablation in Figure 6 (p < 0.05 after Bonferroni correction). In contrast, the worst users do not seem to experience any significant change over time except for the length of their rounds and negative emotional cue use. In those cases, the worst users seem to worsen over time. 8 Related Work In addition to the most relevant studies mentioned so far, our work is related to three broad areas: skill estimation, argumentation mining, and studies of online debates. Skill estimation. Ranking player strength has been studied extensively for sports and for online matchmaking in games such as Halo (Herbrich et al., 2007). The Bradley-Terry models (of which Elo is an example) serve as a basis for much of the research in learning from pairwise comparisons (Bradley and Terry, 1952; Elo, 1978). Another rating system used for online matchmaking is Microsoft’s Trueskill rating system (Herbrich et al., 2007), which assumes performance is nor- mally distributed. Neural networks have recently been explored (Chen and Joachims, 2016; Menke and Martinez, 2008; Delalleau et al., 2012), incor- porating player or other contextual game features from previous games at the cost of interpretability of those features. Argumentation and persuasion. Past studies have noted the persuasiveness of stylistic effects such as phrasing or linguistic accommodation. For example, Danescu-Niculescu-Mizil et al. (2012) showed that, in a pool of people vying to become an administrator of a Web site, those who were promoted tended to coordinate more than those who were not. Similarly, other work defines and discusses power relations over discussion threads such as emails (Prabhakaran and Rambow, 2014, 2013). Additionally, Tan et al. (2018) explored how debate quotes are selected by news media. They found that linguistic and interactive factors of an utterance are predictive of whether or not it would be quoted. Prabhakaran et al. (2014) 547 also studied political debates and found that a debater’s tendency to switch topics correlates with their public perception. Argumentation has also been studied extensively in student persuasive essays and web discourse (Persing and Ng, 2015; Ong et al., 2014; Song et al., 2014; Stab and Gurevych, 2014; Habernal and Gurevych, 2017; Lippi and Torroni, 2016). Most relevant to our work on how users improve over time, Zhang et al. (2017) study how one document may improve over time through annotated revisions. Where our work examines users’ linguistic change across multiple debates, they focus on how a user improves a single document over multiple revisions. Online debates. There has also been recent work in characterizing specific arguments in online settings in contrast to our focus on the debaters themselves. For example, Somasundaran and Wiebe (2009), Walker et al. (2012), Qiu et al. (2015), and Sridhar et al. (2015) built systems for identifying the stances users take in online debate forums. Lukin et al. (2017) studied how persuasiveness of arguments depends on person- ality factors of the audience. Other researchers have focused on annotation tasks. For example, Park and Cardie (2014) anno- tated online user comments to identify and clas- sify different propositions. Hidey et al. (2017) annotated comments from the changemyview sub- reddit, a community where participants ask the community to change a view they hold. Likewise, Anand et al. (2011) annotated online blogs with a classification of persuasive tactics. Inspired by Aristotle’s three modes of persuasion (ethos, pathos, and logos) their work annotates claims and premises within the comments. Habernal and Gurevych (2016) used crowdsourcing to study what makes an argument more convincing. They paired two similar arguments and asked annotators which one was more convincing. This framework allowed them to study the flaws in the less convincing arguments. The annotations they produced offer a rich understanding of arguments which, though costly, can be useful as future work. 9 Conclusion In this work, we introduced a method that uses a linguistic profile derived from a debater’s their skill history of past debates to model level as it changes over time. Using data from Debate.org, we formulate a forecasting task around predicting which of two debaters will be most convincing to observers predisposed to be unconvinced. We find that linguistic profiles on their own are similarly predictive to the clas- sic Elo method, which does not parameterize skill according to attributes of a participant or their behavior, but only models wins and losses. Moreover, we show that our findings are robust to topic of debate and frequency of user activity. Further, a model combining linguistic profiles with Elo achieves predictive accuracy nearly on par with an oracle based on the content of the debate itself. A particular feature combining hedging with fightin’ words is notably important in our model, and consistent with evidence that debaters improve with practice, more recent de- bates appear to provide better estimates of skill than earlier ones. To verify our hypothesis that users improve, we explicitly track the feature use of debaters over the course of their debating activity to show that the best users improve while the bottom users stagnate. Our approach sets the stage for future explorations of the role of history profile, discourse, more fine-grained sentiment features, or notions of topic in persuasion. Acknowledgments We thank Esin Durmus for her help with the DDO corpus. We also thank the anonymous reviewers and the action editor for their helpful comments and suggestions that helped improve the paper. This research was supported in part by a University of Washington Innovation Award to the third author. References Tim Althoff, Cristian Danescu-Niculescu-Mizil, and Dan Jurafsky. 2014. How to ask for a favor: A case study on the success of altruistic re- quests. In Proceedings of ICWSM. Pranav Anand, Joseph King, Jordan Boyd-Graber, Earl Wagner, Craig Martell, Doug Oard, and Philip Resnik. 2011. Believe me? We can do this! Annotating persuasive acts in blog text. In Proceedings of the Workshops at AAAI. Ralph Allan Bradley and Milton E. Terry. 1952. Rank analysis of incomplete block designs: I. the method of paired comparisons. Biometrika, 39(3/4):324–345. 548 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . e d u / t a c l / l a r t i c e - p d f / d o i / . 1 0 1 1 6 2 / t l a c _ a _ 0 0 2 8 1 1 9 2 3 1 9 9 / / t l a c _ a _ 0 0 2 8 1 p d . f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 Shuo Chen and Thorsten Joachims. 2016. Pre- dicting matchups and preferences in context. In Proceedings of SIGKDD. Ken Hyland. 1996. Writing without conviction? Hedging in science research articles. Applied Linguistics, 17(4):433–454. Cristian Danescu-Niculescu-Mizil, Lillian Lee, Bo Pang, and Jon Kleinberg. 2012. Echoes of power: Language effects and power differences in social interaction. In Proceedings of WWW. Olivier Delalleau, Emile Contal, Eric Thibodeau- Laufer, Raul Chandias Ferrari, Yoshua Bengio, and Frank Zhang. 2012. Beyond skill rating: Advanced matchmaking in ghost recon online. IEEE Transactions on Computational Intelli- gence and AI in Games, 4(3):167–177. J Peter Kincaid, Robert P. Fishburne Jr, Richard L. Rogers, and Brad S. Chissom. 1975. Deri- vation of new readability formulas (Automated Readability Index, Fog Count and Flesch Reading Ease Formula) for Navy enlisted personnel. CNTECHTRA Research Branch Report 8-75. Marco Lippi and Paolo Torroni. 2016. Argumen- tation mining: State of the art and emerging trends. ACM Transactions on Internet Tech- nology (TOIT), 16(2):10. Esin Durmus and Claire Cardie. 2018. Exploring the role of prior beliefs for argument persuasion. In Proceedings of NAACL-HLT. Esin Durmus and Claire Cardie. 2019. Modeling the factors of user success in online debate. In Proceedings of WWW. Arpad E. Elo. 1978. The Rating of Chessplayers, Past and Present. Arco Publishers. Ivan Habernal and Iryna Gurevych. 2016. What makes a convincing argument? Empirical analysis and detecting attributes of convincing- ness in Web argumentation. In Proceedings of EMNLP. Ivan Habernal and Iryna Gurevych. 2017. Argu- mentation mining in user-generated Web dis- course. Computational Linguistics, 43(1):125–179. David A. Hanauer, Yang Liu, Qiaozhu Mei, Frank J. Manion, Ulysses J. Balis, and Kai Zheng. 2012. Hedging their mets: The use of uncertainty terms in clinical documents and its potential implications when sharing the documents with patients. In Proceedings of AMIA. Ralf Herbrich, Tom Minka, and Thore Graepel. 2007. TrueskillTM: A Bayesian skill rating sys- tem. In Proceedings of NIPS. Christopher Hidey, Elena Musi, Alyssa Hwang, Smaranda Muresan, and Kathy McKeown. 2017. Analyzing the semantic types of claims and premises in an online persuasive forum. In Proceedings of the Workshop on Argument Mining. Stephanie M. Lukin, Pranav Anand, Marilyn Walker, and Steve Whittaker. 2017. Argument strength is in the eye of the beholder: Audience effects in persuasion. In Proceedings of EACL. Joshua E. Menke and Tony R. Martinez. 2008. A Bradley–Terry artificial neural network model for individual ratings in group com- petitions. Neural Computing and Applications, 17(2):175–186. Burt L. Monroe, Michael P. Colaresi, and Kevin M. Quinn. 2008. Fightin’ words: Lexical feature selection and evaluation for identify- ing the content of political conflict. Political Analysis, 16(4):372–403. Nathan Ong, Diane Litman, and Alexandra Brusilovsky. 2014. Ontology-based argument mining and automatic essay scoring. In Pro- ceedings of the Workshop on Argumentation Mining. Joonsuk Park and Claire Cardie. 2014. Identifying appropriate support for propositions in online user comments. In Proceedings of the Workshop on Argumentation Mining. Isaac Persing and Vincent Ng. 2015. Modeling argument strength in student essays. In Pro- ceedings of ACL. Peter Potash and Anna Rumshisky. 2017. To- wards debate automation: A recurrent model for predicting debate winners. In Proceedings of EMNLP. Vinodkumar Prabhakaran, Ashima Arora, and Owen Rambow. 2014. Staying on topic: An 549 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . e d u / t a c l / l a r t i c e - p d f / d o i / . 1 0 1 1 6 2 / t l a c _ a _ 0 0 2 8 1 1 9 2 3 1 9 9 / / t l a c _ a _ 0 0 2 8 1 p d . f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 indicator of power in political debates. In Pro- ceedings of EMNLP. Vinodkumar Prabhakaran and Owen Rambow. 2013. Written dialog and social power: Mani- festations of different types of power in dialog behavior. In Proceedings of IJCNLP. Vinodkumar Prabhakaran and Owen Rambow. relations between 2014. Predicting power participants in written dialog from a single thread. In Proceedings of ACL. Minghui Qiu, Yanchuan Sim, Noah A. Smith, and Jing Jiang. 2015. Modeling user arguments, interactions, and attributes for stance prediction in online debate forums. In Proceedings of SDM. Nate Silver. 2015. How our NFL predictions work. https://fivethirtyeight.com/ methodology/how-our-nfl-predictions- work/. Accessed: 2019-02-30. Swapna Somasundaran and Janyce Wiebe. 2009. Recognizing stances in online debates. In Pro- ceedings of ACL-IJCNLP. Yi Song, Michael Heilman, Beata Beigman Klebanov, and Paul Deane. 2014. Applying argumentation schemes for essay scoring. In Proceedings of the Workshop on Argumentation Mining. Dhanya Sridhar, James Foulds, Bert Huang, Lise Getoor, and Marilyn Walker. 2015. Joint models of disagreement and stance in online debate. In Proceedings of ACL. Christian Stab and Iryna Gurevych. 2014. Iden- tifying argumentative discourse structures in persuasive essays. In Proceedings of EMNLP. Chenhao Tan, Lillian Lee, and Bo Pang. 2014. The effect of wording on message propagation: Topic-and author-controlled. In Proceedings of ACL. Chenhao Tan, Vlad Niculae, Cristian Danescu- Niculescu-Mizil, and Lillian Lee. 2016. Win- Interaction dynamics and ning arguments: persuasion strategies in good-faith online discussions. In Proceedings of WWW. Chenhao Tan, Hao Peng, and Noah A. Smith. 2018. You are no Jack Kennedy: On media selection of highlights from presidential de- bates. In Proceedings of WWW. Yla R. Tausczik and James W. Pennebaker. 2010. The psychological meaning of words: LIWC and computerized text analysis methods. Journal of Language and Social Psychology, 29(1):24–54. Marilyn A. Walker, Pranav Anand, Robert Abbott, and Ricky Grant. 2012. Stance classification using dialogic properties of persuasion. In Pro- ceedings of NAACL-HLT. Lu Wang, Nick Beauchamp, Sarah Shugars, and Kechen Qin. 2017. Winning on the merits: The joint effects of content and style on debate outcomes. Transactions of the Association for Computational Linguistics, 5:219–232. Fan Zhang, Homa B. Hashemi, Rebecca Hwa, and Diane Litman. 2017. A corpus of annotated revisions for studying argumentative writing. In Proceedings of ACL. Justine Zhang, Ravi Kumar, Sujith Ravi, and Cristian Danescu-Niculescu-Mizil. 2016. Con- versational flow in Oxford-style debates. In Proceedings of NAACL-HLT. 550 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . e d u / t a c l / l a r t i c e - p d f / d o i / . 1 0 1 1 6 2 / t l a c _ a _ 0 0 2 8 1 1 9 2 3 1 9 9 / / t l a c _ a _ 0 0 2 8 1 p d . f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3Measuring Online Debaters’ Persuasive Skill from Text over Time image
Measuring Online Debaters’ Persuasive Skill from Text over Time image
Measuring Online Debaters’ Persuasive Skill from Text over Time image
Measuring Online Debaters’ Persuasive Skill from Text over Time image
Measuring Online Debaters’ Persuasive Skill from Text over Time image
Measuring Online Debaters’ Persuasive Skill from Text over Time image
Measuring Online Debaters’ Persuasive Skill from Text over Time image
Measuring Online Debaters’ Persuasive Skill from Text over Time image
Measuring Online Debaters’ Persuasive Skill from Text over Time image
Measuring Online Debaters’ Persuasive Skill from Text over Time image
Measuring Online Debaters’ Persuasive Skill from Text over Time image

Download pdf