Categorical Metadata Representation for Customized Text Classification

Categorical Metadata Representation for Customized Text Classification

Jihyeok Kim*1 Reinald Kim Amplayo*2

Kyungjae Lee1

Sua Sung1 Minji Seo1

Seung-won Hwang1

(* equal contribution)

1Yonsei University
zizi1532@yonsei.ac.kr

2University of Edinburgh

reinald.kim@ed.ac.uk

yo

D
oh
w
norte
oh
a
d
mi
d

F
r
oh
metro
h

t
t

pag

:
/
/

d
i
r
mi
C
t
.

metro

i
t
.

mi
d
tu

/
t

a
C
yo
/

yo

a
r
t
i
C
mi

pag
d

F
/

d
oh

i
/

.

1
0
1
1
6
2

/
t

yo

a
C
_
a
_
0
0
2
6
3
1
9
2
3
0
4
7

/

/
t

yo

a
C
_
a
_
0
0
2
6
3
pag
d

.

F

b
y
gramo
tu
mi
s
t

t

oh
norte
0
7
S
mi
pag
mi
metro
b
mi
r
2
0
2
3

{lkj0509,dormouse,ggatalminji,seungwonh}@yonsei.ac.kr

Abstracto

The performance of text classification has
improved tremendously using intelligently
engineered neural-based models, especially those
injecting categorical metadata as additional
información, p.ej., using user/product informa-
tion for sentiment classification. This infor-
mation has been used to modify parts of the
modelo (p.ej., word embeddings, attention mech-
anisms) such that results can be customized
according to the metadata. We observe that
current representation methods for categorical
metadata, which are devised for human con-
sumption, are not as effective as claimed in
popular classification methods, outperformed
even by simple concatenation of categorical
features in the final
layer of the sentence
encoder. We conjecture that categorical fea-
tures are harder to represent for machine use,
as available context only indirectly describes
la categoría, and even such context is often
scarce (for tail category). Para tal fin, nosotros pro-
pose using basis vectors to effectively incor-
porate categorical metadata on various parts
of a neural-based model. This additionally
decreases the number of parameters dramatic-
ally, especially when the number of categori-
cal features is large. Extensive experiments on
various data sets with different properties are
performed and show that through our method,
we can represent categorical metadata more
effectively to customize parts of the model,
including unexplored ones, and increase the
performance of the model greatly.

1

Introducción

Text classification is the backbone of most NLP
tareas: review classification in sentiment analysis

201

(Pang et al., 2002), paper classification in sci-
entific data discovery (Sebastiani, 2002), y
question classification in question answering (li
and Roth, 2002), to name a few. While prior meth-
ods require intensive feature engineering, recent
methods enjoy automatic extraction of features
from text using neural-based models (Socher et al.,
2011) by encoding texts into low-dimensional
dense feature vectors.

This paper discusses customized text clas-
sification, generalized from personalized text
clasificación (Baruzzo et al., 2009), where we
customize classifiers based on possibly multiple
different known categorical metadata information
information for sentiment
(p.ej., user/product
clasificación) instead of just the user information.
As shown in Figure 1, in addition to the text,
a customizable text classifier is given a list of
categories specific to the text to predict its class.
Existing works applied metadata information to
improve the performance of a model, como
user and product (Tang et al., 2015) información
in sentiment classification, and author (rosa-
Zvi et al., 2004) and publication (Joorabchi and
Mahdi, 2011) information in paper classification.
Towards our goal, we are inspired by the ad-
vancement in neural-based models, incorporat-
ing categorical information ‘‘as is’’ and injecting
it on various parts of the model such as in the
word embeddings (Tang et al., 2015), atención
mechanism (Chen et al., 2016; Amplayo et al.,
2018a) and memory networks (Dou, 2017).
these methods theoretically make use
En efecto,
of combined features from both textual and
categorical
características, which make them more
powerful than disconnected features. Sin embargo,
metadata is generated for human understanding,
and thus we claim that these categories need
to be carefully represented for machine use to

Transacciones de la Asociación de Lingüística Computacional, volumen. 7, páginas. 201–215, 2019. Editor de acciones: Bo Pang.
Lote de envío: 11/2018; Lote de revisión: 1/2019; Final submission: 2/2019; Publicado 4/2019.
C(cid:2) 2019 Asociación de Lingüística Computacional. Distribuido bajo CC-BY 4.0 licencia.

yo

D
oh
w
norte
oh
a
d
mi
d

F
r
oh
metro
h

t
t

pag

:
/
/

d
i
r
mi
C
t
.

metro

i
t
.

mi
d
tu

/
t

a
C
yo
/

yo

a
r
t
i
C
mi

pag
d

F
/

d
oh

i
/

.

1
0
1
1
6
2

/
t

yo

a
C
_
a
_
0
0
2
6
3
1
9
2
3
0
4
7

/

/
t

yo

a
C
_
a
_
0
0
2
6
3
pag
d

.

F

b
y
gramo
tu
mi
s
t

t

oh
norte
0
7
S
mi
pag
mi
metro
b
mi
r
2
0
2
3

Cifra 1: A high-level framework of models for the Customized Text Classification Task that inputs a text with n
tokens (p.ej., revisar) and m categories (p.ej., users, products) and outputs a class (p.ej., positive/negative). Ejemplo
tasks are shown in the left of the figure.

improve the performance of the text classifier
effectively.

Primero, we empirically invalidate the results from
previous studies by showing in our experiments
on multiple data sets that popular methods using
metadata categories ‘‘as is’’ perform worse than
a simple concatenation of textual and categorical
feature vectors. We argue that this is because of
the difficulties of the model in learning optimized
dense vector representation of the categorical
features to be used by the classification model.
The reasons are two-fold: (a) categorical features
do not have direct context and thus rely solely
on classification labels when training the feature
vectores, y (b) there are categorical information
that are sparse and thus cannot effectively learn
optimal feature vectors.

Segundo, we suggest an alternative represen-
tation, using low-dimensional basis vectors to
mitigate the optimization problems of categorical
feature vectors. Basis vectors have nice properties
that can solve the issues presented here because
ellos (a) transform multiple categories into useful
combinations, which serve as mutual context to all

categories, y (b) intelligently initialize vectors,
especially of sparse categorical information, a
a suboptimal location to efficiently train them
further. Además, our method reduces the
number of trainable parameters and thus is flex-
ible for any kinds and any number of available
categories.

We experiment on multiple classification tasks
with different properties and kinds of catego-
ries available. Our experiments show that while
customization methods using categorical infor-
mation ‘‘as is’’ do not perform as well as the
naive concatenation method, applying our pro-
posed basis-customization method makes them
much more effective than the naive method. Nuestro
method also enables the use of categorical meta-
data to customize other parts of the model, semejante
as the encoder weights, that are previously un-
explored due to their high space complexity and
weak performance. We show that this unexplored
use of customization outperform popular and con-
ventional methods such as attention mechanism
when our proposed basis-customization method
se utiliza.

202

2 Preliminaries

2.1 Problema: Customized Text Classification

The original text classification task is defined
como sigue: Given a text W = {w1, w2, …, wn},
we are tasked to train a mapping function f (W. )
to predict a correct class y ∈ {y1, y2, …, yp}
among the p classes. The customized text
classification task makes use of the categorical
a
metadata information attached on the text
customize the mapping function. en este documento,
we define categorical metadata as non-continuous
information that describes the text.1 An example
task is review sentiment classification with user
and product information as categorical metadata.
Formalmente, given a text t = {W., C}, dónde
W = {w1, w2, …, wn}, C = {c1, c2, …, cm}, wx
is the xth of the n tokens in the text, and cz is the
category label of the text on the zth category of
the m available categories, the goal of customized
text classification is to optimize a function fC(W. )
to predict a label y, where fC(W. ) is the classifier
dependent with C. In our example task, W is the
review text, and we have m = 2 categories where
c1 and c2 are the user and product information.

This is an interesting problem because of the
vast opportunities it provides. Primero, we are moti-
vated to use categorical metadata because exist-
ing work has shown that non-textual additional
información, such as POS tags (Go et al., 2009)
and latent topics (Zhao et al., 2017), can be used
as strong supplementary supervision to improve
the performance of text classification. Segundo,
while previously used additional information is
they are either domain-
found to be helpful,
dependent or very noisy (Amplayo et al., 2018b).
Por otro lado, categorical metadata are
usually factual and valid information that are
either inherent (p.ej., user/product information)
or human-labeled (p.ej., research area). Finalmente,
the customized text classification task generalizes
the personalization problem (Baruzzo et al.,
2009), where instead of personalizing based on
single user information, we customize based on

1We limit our scope to texts with categorical metadata
información (product reviews, news articles, tweets, etc.),
which covers most of the texts on the Web. Texts without
metadata can use predicted categorical information, como
topics from a topic model, which are commonly used (zhao
et al., 2017; Chou et al., 2017). Sin embargo, because the predic-
tion may be incorrect, performance gains cannot be guaran-
teed. We leave the investigation of this area in future work.

possibly multiple categories, which may or may
not include user information. This consequently
creates an opportunity to develop customizable
virtual assistants (Papacharissi, 2002).

2.2 Base Classifier: BiLSTM

We use a Bidirectional Long Short Term Memory
(BiLSTM) network (Hochreiter and Schmidhuber,
1997) as our base text classifier as it is proven to
work well on classifying text sequences (zhou
et al., 2016). Although the methods that are
described here apply to other effective classifiers
también, such as convolutional neural networks
(CNNs) (kim, 2014) and hierarchical models
(Yang et al., 2016), we limit our experiments
to BiLSTM to cover more important findings.

Our BiLSTM classifier starts by encoding the
word embeddings using a forward and a back-
ward LSTM. The resulting pairs of vectors are
concatenated to get the final encoded word vec-
tores, as shown here:

wi ∈ W
−→
h i = LST Mf (Wisconsin,
←−
h i = LST Mb(Wisconsin,
−→
h i;

←−
h i]

hi = [

−→
h i−1)
←−
h i+1)

(1)

(2)

(3)

(4)

Próximo, we pool the encoded word vectors hi into
a text vector d using an attention mechanism
(Bahdanau et al., 2015; Luong et al., 2015), cual
calculates importance scores using a latent context
vector x for all words, normalizes the scores using
softmax, and uses them to do weighted sum on
encoded word vectors, as shown:

ei = x(cid:6)hi

ai =

re =

(cid:2)

exp.(ei)
j exp(ej)

(cid:3)

hi ∗ ai

(5)

(6)

(7)

i

Finalmente, we use a logistic regression classifier to
classify labels using learned weight matrix W (C)
and bias vector b(C):

y(cid:8) = W (C)d + b(C)

(8)

We can then train our classifier using any gradient
descent algorithm by minimizing the negative log
likelihood of the log softmax of predicted labels
y(cid:8) with respect to the actual labels y.

203

yo

D
oh
w
norte
oh
a
d
mi
d

F
r
oh
metro
h

t
t

pag

:
/
/

d
i
r
mi
C
t
.

metro

i
t
.

mi
d
tu

/
t

a
C
yo
/

yo

a
r
t
i
C
mi

pag
d

F
/

d
oh

i
/

.

1
0
1
1
6
2

/
t

yo

a
C
_
a
_
0
0
2
6
3
1
9
2
3
0
4
7

/

/
t

yo

a
C
_
a
_
0
0
2
6
3
pag
d

.

F

b
y
gramo
tu
mi
s
t

t

oh
norte
0
7
S
mi
pag
mi
metro
b
mi
r
2
0
2
3

2.3 Base 1: Concatenated BiLSTM

To incorporate the categories into the classifier,
a simple and naive method is to concatenate the
categorical features with the text vector d. Hacer
este, we create embedding spaces for the dif-
ferent categories and get the category vectors
c1, c2, …, cm based on the category labels of text
d. We then use the concatenated vector as features
for the logistic regression classifier:

y(cid:8) = W (C)[d; c1; c2; …; cm] + b(C)

(9)

2.4 Base 2: Customized BiLSTM

Although the Concatenated BiLSTM easily makes
use of the categories as additional features for
the classifier, it is not able to leverage on the
possible low-level dependencies between textual
and categorical features.
There are different

levels of dependencies
between texts and categories. Por ejemplo, cuando
predicting the sentiment of a review ‘‘The food is
very sweet,’’ given the user who wrote the review,
the classifier should give a positive label if the user
likes sweet foods and a negative label otherwise.
En este caso, the dependency between the review
and the user is on the higher level, where we
look at relationships between the full text and the
categories. Another example is when predicting
the acceptance of a research paper given that the
research area is NLP, the classifier should focus
more on NLP words (p.ej., idioma, texto) bastante
than less-related words (p.ej., biology, chemistry).
En este caso, the dependency between the research
paper and the research area is on the lower level,
where we look at relationships between segments
of text and the categories.

We present five levels of Customized BiLSTM,
which differ on the location where we inject the
categorical features, listed here from the highest
level to the lowest level of dependencies between
text and categories. The main idea is to impose
category-specific weights, rather than a single
weight at each level of the model:

1. Customize on the bias vector: At this level
of customization, we look at the general
biases the categories have towards the prob-
lem. As a concrete example, when classify-
ing the type of message a politician wrote,
he/she can be biased towards writing personal
messages than policy messages. Instead of
using a single bias vector b(C) in the logistic

regression classifier (Ecuación 8), we use
additional multiple bias vectors for each
categoría, as shown below. De hecho, this is
in spirit essentially equivalent to concate-
nated BiLSTM (Ecuación 9), donde el
derivation is:

y(cid:8) = Wdd + bc1 + … + bcm + b(C)

= Wdd + Wc1c1 + … + Wcmcm + b(C)
= W (C)[d; c1; c2; …; cm] + b(C)

2. Customize on the linear transformation:
At this level of customization, we look at
the text-level semantic biases the categories
tener. As a concrete example, in the sentiment
classification task, the review ‘‘The food is
very sweet’’ can have a negative sentiment
if the user who wrote the review does
not like sweets. Instead of using a single
weight matrix W (C) in the logistic regres-
sion classifier (Ecuación 8), we use different
weight matrices for each category:

y(cid:8) = W (C)
c1

d + W. (C)
c2

d + … + W. (C)

cm d + b(C)

3. Customize on the attention pooling: En
this level of customization, we look at the
word importance biases the categories have.
A concrete example is, when classifying a
research paper, NLP words should be focused
more when the research area is NLP. En cambio
of using a single context vector x when calcu-
lating the attention scores e (Ecuación 5),
we use different context vectors for each
categoría:

hi + … + X(cid:6)

cmhi

hi + X(cid:6)
ei = x(cid:6)
c2
c1
a = sof tmax(mi)

(cid:3)

re =

i

hi ∗ ai

4. Customize on the encoder weights: At this
level of customization, we look at the word
contextualization biases the categories need.
A concrete example is, given the text ‘‘deep
learning for political message classifica-
tion’’, when encoding the word classifica-
ción, the BiLSTM should retain the semantics
of words political message more and forget
the semantics of other words more when
the research area is about politics. Instead of

204

yo

D
oh
w
norte
oh
a
d
mi
d

F
r
oh
metro
h

t
t

pag

:
/
/

d
i
r
mi
C
t
.

metro

i
t
.

mi
d
tu

/
t

a
C
yo
/

yo

a
r
t
i
C
mi

pag
d

F
/

d
oh

i
/

.

1
0
1
1
6
2

/
t

yo

a
C
_
a
_
0
0
2
6
3
1
9
2
3
0
4
7

/

/
t

yo

a
C
_
a
_
0
0
2
6
3
pag
d

.

F

b
y
gramo
tu
mi
s
t

t

oh
norte
0
7
S
mi
pag
mi
metro
b
mi
r
2
0
2
3

using a single set of input, forget, producción, y
memory cell weights for each LSTM (Equa-
ciones 2 y 3), we use multiple sets of the
weights, one for each category:




gt
él
ft
ot




⎦ =




tanh
pag
pag
pag




(cid:10)

(cid:3)

(cid:11)

W. (mi)
ck

[peso; ht−1] + b

0

cs.CR (Cryptography and
Security)
Reject

Political Bias
Classification

Neutral
Personal

Partisan
Support

Mesa 5: Example texts from the AAPR data set (superior) and Political Media data set (más bajo) with a variable
category label (research field and political bias) that changes the classification label.

yo

D
oh
w
norte
oh
a
d
mi
d

F
r
oh
metro
h

t
t

pag

:
/
/

d
i
r
mi
C
t
.

metro

i
t
.

mi
d
tu

/
t

a
C
yo
/

yo

a
r
t
i
C
mi

pag
d

F
/

d
oh

i
/

.

1
0
1
1
6
2

/
t

yo

a
C
_
a
_
0
0
2
6
3
1
9
2
3
0
4
7

/

/
t

yo

a
C
_
a
_
0
0
2
6
3
pag
d

.

F

b
y
gramo
tu
mi
s
t

t

oh
norte
0
7
S
mi
pag
mi
metro
b
mi
r
2
0
2
3

Cifra 4: TSNE Visualization of the category vectors of Customized BiLSTM (primera fila) and Basis-Customized
BiLSTM (fila del medio), and the γ coefficients of the latter model (last row), when epoch is equal to 1, 2, 4, y
when training has finished (left to right).

We finally examine the performance of our
models when data contain cold-start entities (es decir.,
users/products may have zero or very few reviews)
using the Sparse80, subset of the Yelp 2013 datos
set provided in Amplayo et al. (2018a). We com-
pare our models with three competing models:

NSC (Chen et al., 2016), which uses a hierarchi-
cal LSTM encoder coupled with customization
on the attention mechanism, BiLSTM+CSAA
(Amplayo et al., 2018a), which uses a BiLSTM
encoder with customization on a CSAA mecha-
nism, and HCSC (Amplayo et al., 2018a), cual es

211

yo

D
oh
w
norte
oh
a
d
mi
d

F
r
oh
metro
h

t
t

pag

:
/
/

d
i
r
mi
C
t
.

metro

i
t
.

mi
d
tu

/
t

a
C
yo
/

yo

a
r
t
i
C
mi

pag
d

F
/

d
oh

i
/

.

1
0
1
1
6
2

/
t

yo

a
C
_
a
_
0
0
2
6
3
1
9
2
3
0
4
7

/

/
t

yo

a
C
_
a
_
0
0
2
6
3
pag
d

.

F

b
y
gramo
tu
mi
s
t

t

oh
norte
0
7
S
mi
pag
mi
metro
b
mi
r
2
0
2
3

Cifra 5: Accuracy per user/product review frequency on Yelp 2013 data set. The review frequency value f
represents the frequencies in the range [F, F + 10), except when f = 100, where it represents the frequencies in
the range [F, inf).

Modelos
NSC
BiLSTM+CSAA
HCSC
BiLSTM+encoder-basis-cust
BiLSTM+linear-basis-cust
BiLSTM+bias-basis-cust
BiLSTM+word-basis-cust
BiLSTM+attention-basis-cust

Accuracy
51.1
52.7
53.8
50.4
50.8
51.9
51.9
53.1

Mesa 6: Performance comparison of competing models
in the Yelp 2013 Sparse80 data set.

a combination of CNN and the BiLSTM encoder
with customization on CSAA.

Results are reported in Table 6, which provide
us two observations. Primero, the BiLSTM model
customized on the linear transformation matrix,
which performs the best on the original Yelp 2013
data set (ver tabla 3), obtains a very sharp decrease
en desempeño. We posit that this is because basis
customization is not able to handle zero-shot cold-
start entities, which are amplified in the Yelp 2013
Sparse80 data set. We leave extensions of basis for
zero-shot or cold-start, studied actively in machine
aprendiendo (Wang y cols., 2019) and recommendation
dominios (Sun et al., 2012), respectivamente. Inspired
by CSAA (Amplayo et al., 2018a), using similar
review texts for inferring the cold-start user (o
product), we expect to infer meta context, similarmente
based on similar meta context, which may mitigate
the zero-shot cold-start problem. Segundo, a pesar de
having no zero-shot learning capabilities, Basis-
Customized BiLSTM on the attention mechanism
performs competitively with HCSC and performs

better than BiLSTM+CSAA, which is Custom-
ized BiLSTM on attention mechanism with cold-
start awareness.

6 Conclusión

We presented a new study on customized text
clasificación, a task where we are given, aside
from the text, its categorical metadata informa-
ción, to predict the label of the text, customized by
the categories available. The issue at hand is that
these categorical metadata information are hardly
understandable and thus difficult to use by neural
máquinas. Este,
por lo tanto, makes neural-based
models hard to train and optimize to find a proper
categorical metadata representation. This issue is
very critical, in such a way that a simple concate-
nation of these categorical information provides
better performance than existing popular neural-
based methods. We propose solving this problem
by using basis vectors to customize parts of a clas-
sification model such as the attention mechanism
and the weight matrices in the hidden layers. Nuestro
results show that customizing the weights using
the basis vectors boosts the performance of a basic
BiLSTM model, and also effectively outperforms
the simple yet robust concatenation methods. Nosotros
share the code and data sets used in our experi-
ments here: https://github.com/zizi1532/
BasisCustomize.

Expresiones de gratitud

This work was supported by Microsoft Research
Asia and IITP/MSIT research grant (No. 2017-0-
01779).

212

Referencias

Reinald Kim Amplayo, Jihyeok Kim, Sua Sung,
and Seung-won Hwang. 2018a. Cold-start
aware user and product attention for sentiment
clasificación. In Proceedings of the 56th An-
nual Meeting of the Association for Compu-
lingüística nacional (Volumen 1: Artículos largos),
pages 2535–2544. Asociación de Computación-
lingüística nacional.

Reinald Kim Amplayo, Kyungjae Lee, Jinyoung
yo, and Seung-won Hwang. 2018b. Trans-
lations as additional contexts for sentence clas-
sification. In Proceedings of the Twenty-Seventh
International Joint Conference on Artificial
Inteligencia, IJCAI 2018, pages 3955–3961.

Dzmitry Bahdanau, Kyunghyun Cho, y yoshua
bengio. 2015. Traducción automática neuronal por
aprender juntos a alinear y traducir. En profesional-
ceedings of the 3rd International Conference
on Learning Representations, ICLR’15.

Andrea Baruzzo, Antonina Dattolo, Nirmala
Pudota, and Carlo Tasso. 2009. A general
framework for personalized text classification
and annotation. In Proceedings of the Workshop
on Adaptation and Personalization for Web 2.0,
AP WEB 2.0@UMAP.

Huimin Chen, Maosong Sun, Cunchao Tu, Yankai
lin, and Zhiyuan Liu. 2016. Neural sentiment
classification with user and product atten-
ción. En Actas de la 2016 Conferencia
sobre métodos empíricos en lenguaje natural
Procesando, pages 1650–1659. Asociación para
Ligüística computacional.

Po-Hao Chou, Richard Tzong-Han Tsai, and Jane
Yung-jen Hsu. 2017. Context-aware sentiment
propagation using LDA topic modeling on
chinese conceptnet. Soft Computing, 21(11):
2911–2921.

Zi-Yi Dou. 2017. Capturing user and product
information for document level sentiment anal-
ysis with deep memory network. En curso-
el 2017 Conferencia sobre Empirismo
cosas de
Métodos en el procesamiento del lenguaje natural,
pages 521–526. Asociación de Computación
Lingüística.

Alec Go, Richa Bhayani, and Lei Huang. 2009.
Twitter sentiment classification using distant

supervision. CS224N Project Report, stanford,
1(12).

Sepp Hochreiter y Jürgen Schmidhuber. 1997.
Memoria larga a corto plazo. Computación neuronal,
9(8):1735–1780.

Arash Joorabchi and Abdulhussain E. Mahdi.
2011. An unsupervised approach to automatic
classification of scientific literature utilizing
bibliographic metadata. Journal of Information
Ciencia, 37(5):499–514.

Yoon Kim. 2014. Convolutional neural networks
for sentence classification. En procedimientos de
el 2014 Conferencia sobre métodos empíricos
en procesamiento del lenguaje natural (EMNLP),
pages 1746–1751. Asociación de Computación-
lingüística nacional.

Xuan Nhat Lam, Thuc Vu, Trong Duc Le,
and Anh Duc Duong. 2008. Addressing cold-
start problem in recommendation systems.
En procedimientos de
the 2nd International
Conference on Ubiquitous Information Man-
agement and Communication, ICUIMC 2008,
pages 208–211.

Fei-Fei Li, Robert Fergus, and Pietro Perona.
2006. One-shot learning of object categories.
IEEE Transactions on Pattern Analysis and
Machine Intelligence, 28(4):594–611.

Xin Li and Dan Roth. 2002. Learning question
In COLING 2002: The 19th
classifiers.
International Conference on Computational
Lingüística.

Rui Lin, Shujie Liu, Muyun Yang, Mu Li,
Ming Zhou, and Sheng Li. 2015. Hierarchical
recurrent neural network for document model-
En g. En Actas de la 2015 Conferencia
sobre métodos empíricos en lenguaje natural
Procesando, pages 899–907. Asociación para
Ligüística computacional.

Yunfei Long, Mingyu Ma, Qin Lu, Rong
Xiang, and Chu-Ren Huang. 2018. Dual mem-
ory network model for biased product review
clasificación. In Proceedings of the 9th Work-
shop on Computational Approaches to Subjec-
actividad, Sentiment and Social Media Analysis,
pages 140–148. Asociación de Computación
Lingüística.

213

yo

D
oh
w
norte
oh
a
d
mi
d

F
r
oh
metro
h

t
t

pag

:
/
/

d
i
r
mi
C
t
.

metro

i
t
.

mi
d
tu

/
t

a
C
yo
/

yo

a
r
t
i
C
mi

pag
d

F
/

d
oh

i
/

.

1
0
1
1
6
2

/
t

yo

a
C
_
a
_
0
0
2
6
3
1
9
2
3
0
4
7

/

/
t

yo

a
C
_
a
_
0
0
2
6
3
pag
d

.

F

b
y
gramo
tu
mi
s
t

t

oh
norte
0
7
S
mi
pag
mi
metro
b
mi
r
2
0
2
3

approaches

Thang Luong, Hieu Pham, and Christopher D.
a
Manning. 2015. Effective
attention-based neural machine translation. En
Actas de la 2015 Conference on Em-
pirical Methods in Natural Language Pro-
cesando, pages 1412–1421. Asociación para
Ligüística computacional.

Dehong Ma, Sujian Li, Xiaodong Zhang,
Houfeng Wang, and Xu Sun. 2017. Cas-
cading multiway attentions for document-level
sentiment classification. En Actas de la
Eighth International Joint Conference on Nat-
ural Language Processing (Volumen 1: Largo
Documentos), pages 634–643. Asian Federation of
Natural Language Processing.

Tomas Mikolov, Kai Chen, Greg Corrado, y
Jeffrey Dean. 2013. Efficient estimation of
word representations in vector space. CORR,
abs/1301.3781.

Bo

y

Angustia, Liliana Lee,

Shivakumar
Vaithyanathan. 2002. Thumbs up? sentiment
classification using machine learning tech-
niques. En Actas de la 2002 Conferencia
sobre métodos empíricos en lenguaje natural
Procesando (EMNLP 2002).

Zizi Papacharissi. 2002. The presentation of self
in virtual life: Characteristics of personal home
paginas. Journalism & Mass Communication
Quarterly, 79(3):643–660.

jeffrey

Socher,

Pennington, Ricardo

y
Christopher Manning. 2014. GloVe: Global
vectors for word representation. En procedimientos
del 2014 Conferencia sobre métodos empíricos
en procesamiento del lenguaje natural (EMNLP),
pages 1532–1543. Asociación de Computación-
lingüística nacional.

Matthew Peters, Mark Neumann, Mohit Iyyer,
Matt Gardner, Christopher Clark, Kenton Lee,
and Luke Zettlemoyer. 2018. Deep contex-
tualized word representations. En procedimientos
del 2018 Conference of the North American
Chapter of the Association for Computational
Lingüística: Tecnologías del lenguaje humano,
Volumen 1 (Artículos largos), pages 2227–2237.
Asociación de Lingüística Computacional.

Proceedings of the 4th International Confer-
ence on Learning Representations, ICLR’16.

Michal Rosen-Zvi, Thomas L. Griffiths, Marca
Steyvers, and Padhraic Smyth. 2004. El
author-topic model for authors and documents.
In UAI ’04, Proceedings of the 20th Confer-
ence in Uncertainty in Artificial Intelligence,
pages 487–494.

Fabrizio Sebastiani. 2002. Machine learning in
automated text categorization. ACM Computing
Surveys, 34(1):1–47.

Richard Socher, Jeffrey Pennington, Eric H.
Huang, Andrew Y. Ng, and Christopher D.
Manning. 2011. Semi-supervised recursive
autoencoders for predicting sentiment distri-
butions. En Actas de la 2011 Conferencia
sobre métodos empíricos en lenguaje natural
Procesando, pages 151–161. Asociación para
Ligüística computacional.

Dong-ting Sun, Tao He, and Fu-hai Zhang. 2012.
Survey of cold-start problem in collaborative
filtering recommender system. Computer and
Modernization, 5:59–63.

Duyu Tang, Bing Qin, and Ting Liu. 2015.
Learning semantic representations of users and
products for document level sentiment classi-
fication. In Proceedings of the 53rd Annual
Meeting of the Association for Computational
Linguistics and the 7th International Joint
Conferencia sobre procesamiento del lenguaje natural
(Volumen 1: Artículos largos), pages 1014–1023.
Asociación de Lingüística Computacional.

Wei Wang, Vincent W. Zheng, Han Yu, y
Chunyan Miao. 2019. A survey of zero-shot
aprendiendo: Settings, methods, and applications.
ACM Transactions on Intelligent Systems and
Tecnología, 10(2):13:1–13:37.

Pengcheng Yang, Xu SUN, wei li, and Shuming
Mamá. 2018. Automatic academic paper rating
based on modularized hierarchical convolu-
tional neural network. En Actas de la
56ª Reunión Anual de la Asociación de
Ligüística computacional (Volumen 2: Short
Documentos), pages 496–502. Asociación para
Ligüística computacional.

Sachin Ravi and Hugo Larochelle. 2016. Optimi-
zation as a model for few-shot learning. En

Zichao Yang, Diyi Yang, Chris Dyer, Xiao Dong
Él, Alex Smola, and Eduard Hovy. 2016.

214

yo

D
oh
w
norte
oh
a
d
mi
d

F
r
oh
metro
h

t
t

pag

:
/
/

d
i
r
mi
C
t
.

metro

i
t
.

mi
d
tu

/
t

a
C
yo
/

yo

a
r
t
i
C
mi

pag
d

F
/

d
oh

i
/

.

1
0
1
1
6
2

/
t

yo

a
C
_
a
_
0
0
2
6
3
1
9
2
3
0
4
7

/

/
t

yo

a
C
_
a
_
0
0
2
6
3
pag
d

.

F

b
y
gramo
tu
mi
s
t

t

oh
norte
0
7
S
mi
pag
mi
metro
b
mi
r
2
0
2
3

Hierarchical attention networks for document
clasificación. En Actas de la 2016 Estafa-
ference of the North American Chapter of the
Asociación de Lingüística Computacional: Hu-
man Language Technologies, pages 1480–1489.
Asociación de Lingüística Computacional.

Mateo D.. Zeiler. 2012. ADADELTA: An adap-
tive learning rate method. CORR, abs/1212.5701.

Peng Zhou, Zhenyu Qi, Suncong Zheng, Jiaming
Xu, Hongyun Bao, and Bo Xu. 2016. Texto
classification improved by integrating bidi-
rectional LSTM with two-dimensional max
pooling. In Proceedings of COLING 2016,
the 26th International Conference on Com-
Lingüística putacional: Technical Papers,
pages 3485–3495.

Rui Zhao, Kezhi Mao, Rui Zhao, and Kezhi
Mao. 2017. Topic-aware deep compositional
models for sentence classification. IEEE/ACM
Transactions on Audio, Speech and Languange
Procesando, 25(2):248–260.

Pengcheng Zhu and Yujiu Yang. 2017. Parallel
multi-feature attention on neural sentiment
clasificación. En procedimientos de
the Eighth
International Symposium on Information and
Communication Technology, pages 181–188.

yo

D
oh
w
norte
oh
a
d
mi
d

F
r
oh
metro
h

t
t

pag

:
/
/

d
i
r
mi
C
t
.

metro

i
t
.

mi
d
tu

/
t

a
C
yo
/

yo

a
r
t
i
C
mi

pag
d

F
/

d
oh

i
/

.

1
0
1
1
6
2

/
t

yo

a
C
_
a
_
0
0
2
6
3
1
9
2
3
0
4
7

/

/
t

yo

a
C
_
a
_
0
0
2
6
3
pag
d

.

F

b
y
gramo
tu
mi
s
t

t

oh
norte
0
7
S
mi
pag
mi
metro
b
mi
r
2
0
2
3

215Categorical Metadata Representation for Customized Text Classification image
Categorical Metadata Representation for Customized Text Classification image
Categorical Metadata Representation for Customized Text Classification image
Categorical Metadata Representation for Customized Text Classification image
Categorical Metadata Representation for Customized Text Classification image
Categorical Metadata Representation for Customized Text Classification image
Categorical Metadata Representation for Customized Text Classification image
Categorical Metadata Representation for Customized Text Classification image
Categorical Metadata Representation for Customized Text Classification image
Categorical Metadata Representation for Customized Text Classification image
Categorical Metadata Representation for Customized Text Classification image
Categorical Metadata Representation for Customized Text Classification image
Categorical Metadata Representation for Customized Text Classification image
Categorical Metadata Representation for Customized Text Classification image
Categorical Metadata Representation for Customized Text Classification image
Categorical Metadata Representation for Customized Text Classification image
Categorical Metadata Representation for Customized Text Classification image
Categorical Metadata Representation for Customized Text Classification image
Categorical Metadata Representation for Customized Text Classification image
Categorical Metadata Representation for Customized Text Classification image
Categorical Metadata Representation for Customized Text Classification image
Categorical Metadata Representation for Customized Text Classification image
Categorical Metadata Representation for Customized Text Classification image
Categorical Metadata Representation for Customized Text Classification image
Categorical Metadata Representation for Customized Text Classification image
Categorical Metadata Representation for Customized Text Classification image
Categorical Metadata Representation for Customized Text Classification image
Categorical Metadata Representation for Customized Text Classification image
Categorical Metadata Representation for Customized Text Classification image
Categorical Metadata Representation for Customized Text Classification image

Descargar PDF