Sentiment Analysis of Events in Social Media
Alexandru Petrescu1 , Ciprian-Octavian Truică2 , Elena-Simona Apostol3
Computer Science and Engineering Department, Faculty of Automatic Control and Computers
University Politehnica of Bucharest, Bucharest, Romania
Email: 1 apetrescu0506@stud.acs.upb.ro, 2 ciprian.truica@cs.pub.ro, 3 elena.apostol@cs.pub.ro
Abstract—The growing popularity of Online Social Networks
has open new research directions and perspectives for content
analysis, i.e., Network Analysis and Natural Language Processing.
From the perspective of information spread, the Network
Analysis community propose Event Detection. This approach
focuses on the network features, without an in-depth analysis
of the textual content, summarization being a preferred method.
Natural Language Processing analyses only the textual content,
not integrating the graph-based structure of the network. To
address these limitations, we propose a method that bridges the
two directions and integrates content-awareness into networkawareness. Our method uses event detection to extract topics of
interest and then applies sentiment analysis on each event. The
obtained results have high accuracy, proving that our method
determines with high precision the overall sentiment of the
detected events.
Index Terms—Event Detection, Sentiment Analysis, Social
Networks Analysis, MABED, Online LDA, SVM, Logistic
Regression
I. I NTRODUCTION
With the increasing use of Online Social Networks and
the many challenges related to analyzing and mining the
textual content generated daily by this new type of media,
new methods that extract discussions and topics of interest and
the opinion that users have about the need to be developed.
Multiple research communities are working on analyzing
and making sense of the contact generated daily by these
networks, namely the Network Analysis and Natural Language
Processing communities. The Network Analysis community
develop approaches dealing with information spread and
mitigation of harmful content using Event Detection, while
the Natural Language Processing community develop methods
to analyze the user opinion using Sentiment Analysis. Event
Detection (ED) is used to detect the impact and spread of
topics on Social Network. Sentiment Analysis (SA) uses
supervised and unsupervised learning to determine the polarity
of textual data. The dimensions or the actual classification
of a text is different for each method, ranging from binary
(positive/negative or neutral if we cannot tell) to multidimensional approaches (happiness, contempt, surprise and so
on).
Our method brings together the research of these two
isolated communities, i.e., Network Analysis and Natural
Language Processing, by combining Network Analysis with
Machine Learning for detecting event sentiments in Online
Social Networks. We aim to put the solutions developed in
these two areas at work in combination with each other,
978-1-7281-4914-1/19/$31.00 © 2019 IEEE
facilitating the detection and spread of events in a contentaware and network-aware manner.
This paper presents a new approach that combines event
detection and sentiment analysis methods on social media
networks. Although for our experiments we consider only
the Twitter platform, our solution can be easily adapted to
other types of social media data. Twitter is a micro-blogging
platform that supports only relatively small posts with a
maxim size of 140 characters. These posts can also contain
images, videos, mentions or tags. Twitter is a widely used
platform in textual processing related to social interaction and
human behavior patterns. As the relational system of Twitter is
modeled like a directional graph, it is helpful for information
diffusion or anomaly detection models.
This paper is structured as follows. Section II presents
a survey of the current state of the art methods for Event
Detection and Sentiment Analysis. The proposed solution and
its main functionalities are described in Section III. Likewise,
in this section, we briefly describe how each component
interacts with the overall platform and how each chosen
algorithm works. Section IV showcases the obtained results,
analyzes them and gives some directions for future trials, using
the chosen Event Detection and Sentiment Analysis methods.
In the final section (Section V) of this article, we conclude
and we present several new directions and improvements for
the proposed solution.
II. R ELATED W ORK
In this section, we will present an overview of the current
state of the art related to Event Detection (ED) and Sentiment
Analysis (SA).
A. Event Detection
Information diffusion is the field that analyses how the states
of individual nodes in a graph evolve over time [9]. It is used
to detect user behavior and information spreading across social
networks through monitor the magnitude (the number of users
it reaches) and the lifespan of a topic. In the case of Online
Social Networks, these topics are named bursty topics [6], i.e.,
topics that spread information better than the average.
In [13], the authors survey the state-of-the-art methods
for ED and build and taxonomy. Their analysis focuses on
detecting bursty topics on the Social Media platform Twitter.
The analysis includes multiple algorithms and methods,
e.g. PT (Peaky Topics) [19] that uses normalized term
frequency, TSTE (Temporal and Social Terms Evaluation) [4]
143
that uses a five-step approach considering both temporal
and social properties, MACD (Moving Average Convergence
Divergence) [15] that uses the trend momentum of a topic,
SDNML (Sequentially Discounting Normalized Maximum
Likelihood) [21] specialized in tweets with media and URLs,
and OLDA (Online Latent Dirichlet Allocation) [2] which
incrementally update the topic model at each time slice using
the previously generated model as a prior. Experimental results
prove that the information is diffused by users that appear as
the node for which the graph partitions [11].
Another approach uses machine learning techniques and
the inference of time-dependent diffusion probabilities from a
multidimensional analysis of individual behaviors and builds
communities (sub-graphs) [12]. To build these sub-graphs and
to have a better view of the information propagation, three
dimensions are considered: semantics, social, and time. The
experimental results that using machine learning to infer the
information diffusion has high accuracy on the tested datasets.
MABED (Mention-Anomaly-Based Event Detection) [10]
is a method that uses user mentions to detect events and their
impact on Social Networks. The approach considers for each
user at most 2 levels of followers (i.e., the direct followers of
the user and the direct followers of the followers) to build
communities. The experimental results prove that MABED
has higher topic readability and temporal precision than the
algorithm used as the baseline, i.e., PS (Peakiness Score) [19]
and EDCoW ((Event Detection with Clustering of Waveletbased Signals) [23].
B. Sentiment Analysis
SA is a field in Natural Language Processing that analyzes
user opinions and emotions from written language [16].
There are two main approaches for SA: i) Lexicon Based
(unsupervised methods that use word polarity to classify
textual data), and ii) Machine Learning (supervised methods
that use polarity labeled dataset to build a model).
In [20], the authors present a comprehensive study of
the two main SA approaches. The experiments show that
the approach that achieves the best results uses Multivariate
Analysis of Variance (MANOVA) to exact features. This
method of feature extraction manages to highlight the relation
between dependent and independent variables for the task of
SA.
SA was used to detect the opinions presented in tweets [22].
In this study, the authors used a scale from 0 to 5 to represent
the sentiment intensity for both positive and negative tweets,
where 0 is the lack of intensity and 5 means a strong intensity.
The approach also uses the concept of Peak Hours (i.e., 5
hours before and after the maximum volume hour) to detect
events. The results show that events with negative sentiments
are more frequent, whereas positive predictions were accurate
only during peak hours.
In [14], the authors proposed the use of metadata extracted
from tweets (i.e., hashtags, emoticons, and abbreviations) to
enhance the feature space for supervised methods. This method
uses a hybrid approach combining multiple lexicons containing
the metadata with classification algorithms. The experimental
results show that this hybrid approach increases the accuracy
of detecting positive, negative, and neutral sentiments.
The overuse of adjectives in a phrase influence the accuracy
of sentiment detection. To overcome this limitation, one
method proposes to classify sentences with multiple adjectives
in the 2 classes (i.e., positive and negative) [18]. This method
shows that with the increase of adjectives the accuracy
of detecting neutral tweets decreases. For the experimental
validation, the authors used SVM and Naı̈ve Bayes on two
small datasets, between 100 and 2917 sentences.
III. M ETHODOLOGY
In this section, we present our proposed solution for
accurate event and sentiment detection in social media. This
solution offers two classes of tasks: Event Detection (ED) and
Sentiment Analysis (SA) tasks. Both ED and SA uses a raw
frequency-based embedding as a method to word vectors. This
approach simply counts the appearance of each word in a
document using NLTK Twitter Tokenizer.
A. Proposed Architecture
The architecture of the solution is presented in Figure 1.Its
main functionality and modules are as follows:
• Training Engine is used for the training step of the ML
models. It can be run either locally if the machine is
powerful enough or on a stand-alone remote machine that
can be accessed through the internet to trigger an ED task.
• TwitterMining Module is constantly listening for twitter
data, in order to fetch and pass it on to the CosmosDB
managed the environment.
• Storage Module uses multiple data sources depending on
the task at hand: Azure Storage for data persistence, CSV
for training modules and JSON for twitter mining.
• AzureML Module uses the models trained by the Training
Engine and offers Web API interfaces for them. It also
grabs data from the Storage Module and exposes several
APIs to the WEB Manager Module.
• WEB Manager Module offers both server and clientside functionalities. It is managing all the other modules
and also the user requests. Based on the nature of the
user request, it can either submit a task to the AzureML
Module, look in the Metadata Module for cached results
or present the data with the help of the PowerBI Module.
• Metadata Module is used as a data source for the PowerBI
Module and as a cache server by the WEB Manager
Module.
• PowerBI Module uses the data from the tasks submitted
by the WEB Manager Module and feeds reports to it.
B. Event Detection Tasks
For ED, we use the Mention-Anomaly-Based Event
Detection (MABED) and Online LDA (OLDA) algorithms.
Both approaches detect bursty topics either on cleaned or
raw text. The cleaned text results by removing the stop
words and any non-character from the raw text and applying
144
3) Nti and Nt′q ti the time-series with the correlation
coefficient proposed in [8]
OnlineLDA [2], is an update on Latent Dirichlet
Allocation [3] which uses a non-Markov on-line LDA Gibbs
sampler topic model. This model is capable of detecting bursty
topics by capturing thematic patterns and identifying emerging
topics and their changes over time. This approach is sometimes
better than the original LDA approach, as it updates the
previous model at each time frame.
OLDA is a hierarchical Bayesian network that relates words
and documents through latent topics using a probabilistic
formula for a given term to be assigned to a given topic in the
(current) context Equation ((5)).
VK
Cw
+ βwi ,j
¬ j
P (zi = j|z¬i , wdi , α, β) ∝ V i
×
VK
v=1 Cv¬ j + βv,j
i
CdDK
+ αwi ,j
¬ j
V
i
DK
v=1 Cd¬ j + αd,j
Figure 1: Architecture diagram
lemmatization. After this preprocessing step, we can see some
relevant results, as shown in Figure 2, otherwise, all the stop
words would have been seen as important.
Mention-Anomaly-Based Event Detection [10] is an
efficient statistical method used for detecting events in social
networks since it is immune to social media bias consisting of
unrelated texts on a given topic. Although social networks can
get spammy, meaning that messages with no actual intent are
posted around certain hours, this method has proved good in
filtering the irrelevant content. MABED is not always viable
since, without external context, it can sometimes distort some
type of events.
MABED [10] uses Equation (1) to detect bursty topics,
with the components from Equation (2), Equation (3) and
Equation (4).
ρOt,t′ + 1
q
wq =
(1)
2
ρOt,t
A2t
A2t′q
q′
=
=
=
b
i=a+1
·At,tq′
(b − a − 1) · At · Atq′
b
·(Nti − Nti−1 )2
(b − a − 1) · At · Atq′
i=a+1
b
i=a+1
·(Nt′q − Nti−1
)2
′
q
(b − a − 1) · At · Atq′
Where:
1) t and tq are therms
2) I = [a; b] is the time interval
(2)
Where:
• D is the total number of documents; K is the number of
topics
• V is the total number of unique words
VK
• Cw j is the number of times word w is assigned to topic
¬i
j, not including the current token instance i
DK
• Cd j is the number of times topic j is assigned to some
¬i
word token in document d, not including the current
instance i
• z¬i are all other word tokens; wdi are the unique word
associated with the i-th token in document d at the current
time
• βk is the V vector of priors for topic k at the current time
• αk is the K vector of priors for document k at the current
time
C. Sentiment Analysis Tasks
Support vector machines (SVM) [5] is a supervised learning
algorithm used in ML for classifications. Given a tagged
dataset, SVM builds a function that separates the space of
this given dataset in two classes In SA, this type of model is
used for binary classification, where the classes are positive
and negative representing the tweets’ sentiments. For the SA
tasks we employ a linear SVM that uses the function from
Equation (6) that tries to separate the values of the training
set using Equation (7). We consider that −1 and 1 are the
numerical equivalent of the two classes.
(3)
(4)
(5)
i
f (x) = wT ∗ x + b
(6)
yi ∗ f (wiT ∗ xi + b) = 1
(7)
Where:
• xi is the ”i”-th example
• yi is the actual class of the the example xi
• w is the weight matrix that we need to compute
• b is the bias
145
Logistic regression [7] is a statistical model, widely used,
that in its basic form uses a logistic function to model a
binary classification for the given dataset. It will try to build
a function, a linear one for this case that will summarize the
input data. Logistic regression is a supervised method that
behaves better than Linear Regression [17], when applied on
a sparse dataset, and builds classification models within a
probabilistic context [1]. We consider that a dataset is sparse
when the matrix of its representation has a lot of zeros and a
few one values, like ours, will be. If we represent the words
depending on their appearance into a certain block of text
(tweet) or not we will get a sparse matrix since the tweets
have at most 140 characters.
The probability for this last presented model is given by
Equation (8), which has the same notations as to the ones
from Equation 6. As can be observed, this is a Sigmoid-type
function.
Pw (y = ±1|x) ≡
1
1+
e−y·wT ·x
(8)
D. Sentiment Analysis of Events
The ED step produces a sentence that will be the topic of an
event. This result contains the key-words of all the tweets that
are included in that topic. This implies that we can apply SA
over the results, in order to guess if the event has a positive
or negative impact over the network.
Combining the magnitude and lifespan of the events with
the sentiment label, we can create a better view of the tweet
dataset and understanding of the user behavior and the topics
of interest. For example, we can determine if negative topics
have a higher impact or longer duration. Also, if we consider
the individual tweets belonging to an event in a topic we can
construct communities based on their opinions.
Algorithm 1 Sentiment Analysis of Events Algorithms
Require: a document set D
Ensure: a event-sentiment set ES
1: ES = ∅
2: DC = EDT P (D)
3: E = ED(DC )
4: for each ei ∈ E do
5:
D(i) = extract(ei , D)
(i)
6:
DC = SAT P (D(i) )
7:
S=∅
(i)
8:
for d ∈ DC do
9:
S = S ∪ s|S = SA(d)
(i)
10:
sj = M axSA(S)
(i)
11:
ES = ES ∪ (ei , sj )
12:
return ES
The proposed method is implemented through Algorithm 1.
The input is a set of documents D whereas the output is a
(i)
(i)
event-sentiment set ES = {(ei , sj )|ei ∈ E ∧ sj ∈ S}
(i)
that contains the sentiment sj ∈ S for each event ei ∈ E,
where E are the events and S are the sentiments and. The
fist step initializes the event-sentiment set (Line 1), then uses
text prepossessing for ED (EDT P ) (Line 2), and extract
the list of events (Line 3). During the iteration of the list
of events E (Line 4) we extract the list of documents for
the event D(i) = extract(ei , D) (Line 5), apply the text
preprocessing steps for SA and extract the clean document set
(i)
DC = SAT P (D(i) ) (Line 6). An empty list S = ∅ (Line 7)
for storing the sentiment for each document is initialized. For
(i)
each document d in set DC detects the sentiment (Lines 8
and 9) and update the list S = S ∪ s|S = SA(d). After these
steps, we label the events with the sentiment with the most
number of occurrences (Lines 10 and 11).
IV. E XPERIMENTAL R ESULTS
In this section, first, we present the used dataset for Event
Detection (ED) and Sentiment Analysis (SA) tasks. Secondly,
we analyze the results obtained for the two types of tasks
individually. Finally, we present the result of detecting the
sentiments for each individual detected event. The source code
is publicly available on GitHub 1
A. Dataset
As a dataset, we used Sentiment140 2 . This dataset has
the following fields: i) the text, the date, and the sentiment
label. The text field is used in both tasks, after preprocessing
is applied. The date is used in the ED task and the label is
use only in the SA task.
To improve the accuracy for detecting the magnitude and
lifespan of an event, we use text preprocessing. Thus, for the
ED task, we apply the following text preprocessing steps: i)
lowercase the text, ii) split the text into tokens, iii) remove
stop words, iv) remove punctuation, v) lemmatize each word
and extract the lemma, vi) weighed each term using the raw
frequency (ft,d ), i.e., the number of times a term appears
in a document. For testing, we constructed three different
prepossessed texts: a) MT (Minimal Text) uses only steps i),
ii), iii), and vi) for prepossessing, b) PT (Pure Text) uses
only steps i), ii), iii), iv), and vi) for preprocessing; and
c) CT (Clean Text) includes all the preprocessing steps. By
employing these steps, we manage to minimize the vocabulary
and improve the accuracy of the ED methods. Figure 2
presents the word cloud after text preprocessing.
For the SA task, we apply the following text preprocessing
steps: i) split the text into tokens, ii) remove punctuation, ii)
weighed each term using the raw frequency (ft,d ). Using these
steps we obtain the SCT (Sentiment Clean Text). In order to
achieve better accuracy we further preprocess the text to obtain
SFE (Sentiment Clean Text with Feature Engineering): i) keep
the case of words, duplicate letter in words, and stop words to
preserve the sentiment polarity, ii) expand contractions (e.g.,
”don’t” becomes ”do not”), iii) enhance the text using feature
engineering (FE) by combining the negation with the words
in front.
146
1 Repository
https://github.com/xander96/EventDetection-SentimentAnalysis
http://sentiment140.com/
2 Sentiment140
Figure 2: Event Detection Text Preprocessing Word Cloud
We applied Event Detection on the entire dataset, while we
split the corpus into three subsets to see how each Sentiment
Analysis algorithm scales. Thus, for the Sentiment Analysis
tasks we have the following subsets: i) C1 contains 200̇00
random tweets with the labels equally distributed so it would
not over-fit, ii) C2 contains 5000̇00 random tweets with the
labels equally distributed so it would not over-fit and iii)
C3 containing the entire dataset of 1 600 233 tweets. Table I
presents the number of features without and with FE.
Features
30 651
307 890
684 492
3) Results Comparison: We can observe that MABED
and OLDA manage to detect different emerging events when
analyzing the most representative topic keywords using the
text preprocessing CT, although some are the same (Figure II).
MABED provides better results, as it also uses fewer words.
C. Sentiment Analysis
Table I: Number of features for Sentiment Analysis
Dataset
C1
C2
C3
Figure 4: OLDA - Most used words
FE Features
31 018
311 240
691 646
B. Event Detection
We set both algorithms to return 50 events, each event
summarized by the top-10 most relevant terms. The lifespan
of the events is between April and July 2009. For this set of
experiments, we used the three text prepossessing strategies
for ED, i.e, MT, PT, and CT.
1) MABED: MABED manages to extract accurate events
and bursty topics (Figure 3), but the topic of readability is
influenced by the preprocessing strategy we employed. The
text processing is done with MT and PT give more humanreadable results, while when using CT, MABED manages to
summarize the topics, even more, providing higher magnitudes
and life span.
Figure 3: MABED - Event Impact
2) OLDA: OLDA produces topics that are more humanreadable than the ones of MABED (Figure 4). To summarize
a topic, OLDA uses the entire vocabulary. As in the case of
MABED, OLDA with CT manages to better generalize the
topics and cover more tweets.
For the Sentiment Analysis tasks, we use Logistic
Regression (LR) and Support Vector Machine (SVM) together
with SCT and SFE text preprocessing strategies. We use the
Area Under the Curve (AUC), Accuracy, Precision, and Recall
for evaluating the results, and K-Folds for the evaluation
scores.
1) LR Results: The first set of experiments for LR uses
SCT (Table III). The evaluation scores improve with the size
of the text.
When using LR with SFE (Table IV), the evaluation scores
worsen for the small subset of documents C1 , but they improve
with the scale.
Comparing the two experiments, we can observe that the
improvement in the evaluation scores is minimal w.r.t. the text
preprocessing strategy we employed, i.e., overall for all the
measures ∼ 0.004. Furthermore, the scores improvement w.r.t.
the scale is also minimal, i.e., overall for all the measures
∼ 0.05. Thus we can conclude that the number of features
does not impact any of the measures.
2) SVM: For the SVM approach, we have used a linear
kernel as the classes are balanced. We set the penalty
parameter of the error term C = 0.1, value obtained after
hyperparameter tuning.
The first set of experiments for SVM uses SCT (Table III).
The evaluation scores improve with the increasing number of
used features, with an overall increase of almost ∼ 0.5.
When using SVM with SFE (Table IV) we observe the same
ascending trend of the score improvement w.r.t. the scale of
the dataset.
As in the case of LR, we can observe that the improvement
in the evaluation scores is minimal w.r.t. the text preprocessing
strategy we employed, i.e., overall for all the measures ∼
0.01. Furthermore, the overall score improvement for all the
147
Table II: Similar topics example for MABED and OLDA
MABED Topic
nt sleep wa na morning gon early class getting school
movie look read wa song watching getting
phone wa miss ah finally quot doe lt facebook amp
nt night feel show wa hour fun doe
OLDA Topic
work sleep hour tired done early get go still woke
sorry movie hear watch goodnight exam season mileycyrus love watched
twitter friend best leaving wont bye facebook know lol let
fun bit show tonight much little bbq going passed wine
Table III: LR with SCT
Dataset
C1
C2
C3
AUC
0.756
0.790
0.800
Accuracy
0.756
0.790
0.800
Precision
0.703
0.727
0.739
discussed in subsection IV-A. After this step is completed,
we determine the subsets of tweets belonging to each event,
on which we apply the SFE text preprocessing strategy. For
each subset of processed tweets, we use the proposed SA
algorithms and compute the evaluation scores. The results are
presented in Table VIII. The overall best scores for Sentiment
Analysis of events are obtained when combining MABED with
LR. Although, there is only a small gap between the resulted
evaluation scores.
Recall
0.766
0.802
0.810
Table IV: LR with SFE
Dataset
C1
C2
C3
AUC
0.752
0.794
0.804
Accuracy
0.752
0.794
0.804
Precision
0.690
0.730
0.742
Recall
0.752
0.807
0.815
Table VIII: Average scores for Sentiment Analysis of Events
Table V: SVM with SCT
Dataset
C1
C2
C3
AUC
0.803
0.852
0.876
Accuracy
0.729
0.778
0.803
Precision
0.707
0.799
0.795
ED
MABED
Recall
0.772
0.742
0.820
OLDA
Table VI: SVM with SFE
Dataset
C1
C2
C3
AUC
0.816
0.857
0.877
Accuracy
0.748
0.781
0.806
Precision
0.725
0.763
0.795
Recall
0.796
0.815
0.825
measures w.r.t. the scale is also minimal, i.e. ∼ 0.05. Thus we
can conclude once more than the number of features does not
impact any of the measures.
3) Results Comparison: SVM provides better evaluation
scores than LR w.r.t. the number of features, but the
performance improvement is minimal, i.e., ∼ 0.05. Using
SFE improves the evaluation scores for both algorithms. In
conclusion, we affirm that both algorithms present comparable
evaluation scores results. From a runtime perspective, LR is
much faster than SVM (Table VII), as the model construction
takes minutes (m) for LR, while for SVM it takes days (D).
We can conclude the LR algorithm is the best choice for
constructing the SA model. Although the SVM achieves a
very small increase in evaluation score over LR, the waiting
time for building the model is not feasible.
Table VII: SVM vs. LR runtime comparison
Algorithm
LR
SVM
C1
2m
2D
SCT
C2
5m
32 D
C3
20 m
65 D
C1
1.33 m
1D
SFE
C2
4m
4D
C3
11 m
62 D
D. Sentiment in Tweets
To analyze the accuracy of the proposed method, we
run the Event Detection tasks on the entire dataset. For
these experiments, we use the CT text preprocessing strategy,
SA
LR
SVM
LR
SVM
AUC
0.768
0.743
0.764
0.747
Accuracy
0.793
0.781
0.794
0.781
Precision
0.570
0.545
0.661
0.644
Recall
0.720
0.666
0.803
0.777
To better understand this gap between the scores, we analyze
the results individually. Thus, for each event detected in the
previous experiment, we extract the topics and labeled them
with the most recurring sentiment. We also extract the most
recurring sentiment detected by the SA algorithms for each
event. Tables IX and X present a subset of the sentiment
detected for each event. The analysis of the results shows
that our method (Algorithm 1) is accurate and is correctly
identifying the sentiment for each event.
V. C ONCLUSIONS
In this paper, we present a new approach that combines
Event Detection and Sentiment Analysis for extracting the
sentiments of events that appear in Social Networks. By
combining Network Analysis with Machine Learning, our
solution brings together two otherwise isolated communities,
i.e., Network Analysis and Natural Language Processing.
Event Detection algorithms manage to extract different
bursty topics. The experiment results show that OLDA
increases topic readability and coherence, while MABED
detects a wider range of topics.
The evaluation of Sentiment Analysis algorithms provides
better inside in the construction of the model. Although, with
the increase of the dataset size the gap between the evaluation
scores obtained by algorithms increases, this increase does not
prove to be a real gain when taking into account the runtime.
The method we proposed for extracting the sentiments of the
social events shows interesting results. The overall best results
are obtained when combining MABED with LR, although the
increase in the evaluation scores is minimal in comparison
with the other tested approaches. Furthermore, the experiments
prove that our method manages to correctly determine the
overall sentiment for an event.
148
Table IX: Sentiment Analysis of Events detect with MABED
MABED Topic
nt sleep wa na morning gon early class getting school
movie look read wa song watching getting
phone wa miss ah finally quot doe lt facebook amp
nt night feel show wa hour fun doe
course saw party quot lt dude look em hey
True Label
NEGATIVE
POSITIVE
NEGATIVE
NEGATIVE
POSITIVE
SVM Prediction
NEGATIVE
POSITIVE
NEGATIVE
NEGATIVE
POSITIVE
LR Prediction
NEGATIVE
POSITIVE
NEGATIVE
NEGATIVE
POSITIVE
Table X: Sentiment Analysis of Events detect with OLDA
Topic OLDA
work sleep hour tired done early get go still woke
sorry movie hear watch goodnight exam season mileycyrus love watched
twitter friend best leaving wont bye facebook know lol let
fun bit show tonight much little bbq going passed wine
happy looking birthday party forward bday http today great american
As future work, we aim to improve accuracy by using Deep
Learning methods for Sentiment Analysis. We also want to
implement unsupervised methods for detecting the sentiment
of the event. Furthermore, we plan to use better feature
engineering and weighting schemes for the text preprocessing
strategies.
ACKNOWLEDGEMENT
This research was funded by grant No. PN-III-P1-1.2PCCDI-2017-0734.
R EFERENCES
[1] E. Adeli, X. Li, D. Kwon, Y. Zhang, and K. Pohl, “Logistic regression
confined by cardinality-constrained sample and feature selection,” IEEE
Transactions on Pattern Analysis and Machine Intelligence, pp. 1–1,
2019.
[2] L. AlSumait, D. Barbará, and C. Domeniconi, “On-line lda: Adaptive
topic models for mining text streams with applications to topic detection
and tracking,” in IEEE International Conference on Data Mining. IEEE,
2008, pp. 3–12.
[3] D. M. Blei, A. Y. Ng, and M. I. Jordan, “Latent dirichlet allocation,”
Journal of machine Learning research, vol. 3, no. Jan, pp. 993–1022,
2003.
[4] M. Cataldi, L. Di Caro, and C. Schifanella, “Emerging topic detection on
twitter based on temporal and social terms evaluation,” in International
Workshop on Multimedia Wata Mining. ACM, 2010, p. 4.
[5] C. Cortes and V. Vapnik, “Support-vector networks,” Machine learning,
vol. 20, no. 3, pp. 273–297, 1995.
[6] Q. Diao, J. Jiang, F. Zhu, and E.-P. Lim, “Finding bursty topics from
microblogs,” in Association for Computational Linguistic, 2012, pp.
536–544.
[7] R. S. D.W. Hosmer Jr., S. Lemeshow, Applied Logistic Regression, 3rd
Edition. John Wiley & Sons, 2013.
[8] O. Erdem, E. Ceyhan, and Y. Varli, “A new correlation coefficient for
bivariate time-series data,” Physica A: Statistical Mechanics and its
Applications, vol. 414, pp. 274–284, 2014.
[9] A. Goswami and A. Kumar, Event Detection Using Twitter Platform.
Springer International Publishing, 2019, pp. 429–480.
[10] A. Guille and C. Favre, “Mention-anomaly-based event detection and
tracking in twitter,” in IEEE/ACM International Conference on Advances
in Social Networks Analysis and Mining, 2014, pp. 375–382.
[11] A. Guille, C. Favre, H. Hacid, and D. A. Zighed, “Sondy: An open
source platform for social dynamics mining and analysis,” in ACM
SIGMOD International Conference on Management of Data, 2013, pp.
1005–1008.
[12] A. Guille and H. Hacid, “A predictive model for the temporal dynamics
of information diffusion in online social networks,” in International
Conference on World Wide Web. ACM, 2012, pp. 1145–1152.
[13] A. Guille, H. Hacid, C. Favre, and D. A. Zighed, “Information diffusion
in online social networks: A survey,” ACM Sigmod Record, vol. 42,
no. 2, pp. 17–28, 2013.
True Label
NEGATIVE
NEGATIVE
POSITIVE
POSITIVE
POSITIVE
SVM Prediction
NEGATIVE
NEGATIVE
POSITIVE
POSITIVE
POSITIVE
LR Prediction
NEGATIVE
NEGATIVE
POSITIVE
POSITIVE
POSITIVE
[14] E. Kouloumpis, T. Wilson, and J. Moore, “Twitter sentiment analysis:
The good the bad and the omg!” in AAAI International Conference on
Weblogs and Social Media, 2011.
[15] R. Lu and Q. Yang, “Trend analysis of news topics on twitter,”
International Journal of Machine Learning and Computing, vol. 2, no. 3,
p. 327, 2012.
[16] W. Medhat, A. Hassan, and H. Korashy, “Sentiment analysis algorithms
and applications: A survey,” Ain Shams Engineering Journal, vol. 5,
no. 4, pp. 1093–1113, 2014.
[17] D. C. Montgomery, E. A. Peck, and G. G. Vining, Introduction to linear
regression analysis. John Wiley & Sons, 2012, vol. 821.
[18] A. Pak and P. Paroubek, “Twitter based system: Using twitter for
disambiguating sentiment ambiguous adjectives,” in Proceedings of the
5th International Workshop on Semantic Evaluation. Association for
Computational Linguistics, 2010, pp. 436–439.
[19] D. A. Shamma, L. Kennedy, and E. F. Churchill, “Peaks and persistence:
Modeling the shape of microblog conversations,” in ACM Conference
on Computer Supported Cooperative Work, 2011, pp. 355–358.
[20] D. Sirbu, A. Secui, M. Dascalu, S. A. Crossley, S. Ruseti, and
S. Trausan-Matu, “Extracting gamers’ opinions from reviews,” in
International Symposium on Symbolic and Numeric Algorithms for
Scientific Computing, 2016, pp. 227–232.
[21] T. Takahashi, R. Tomioka, and K. Yamanishi, “Discovering emerging
topics in social streams via link anomaly detection,” in IEEE
International Conference on Data Mining. IEEE, 2011, pp. 1230–1235.
[22] M. Thelwall, K. Buckley, and G. Paltoglou, “Sentiment in twitter
events,” Journal of the American Society for Information Science and
Technology, vol. 62, no. 2, pp. 406–418, 2011.
[23] J. Weng and B.-S. Lee, “Event detection in twitter,” in AAAI
International Conference on Weblogs and Social Media, 2011.
149