Anzovino 2018
Anzovino 2018
Abstract. Hate speech may take different forms in online social media.
Most of the investigations in the literature are focused on detecting
abusive language in discussions about ethnicity, religion, gender iden-
tity and sexual orientation. In this paper, we address the problem of
automatic detection and categorization of misogynous language in online
social media. The main contribution of this paper is two-fold: (1) a cor-
pus of misogynous tweets, labelled from different perspective and (2) an
exploratory investigations on NLP features and ML models for detecting
and classifying misogynistic language.
1 Introduction
Twitter is part of ordinary life of a great amount of people1 . Users feel free to
express themselves as if no limits were imposed, although behavioural norms
are declared by the social networking sites. In these settings, different targets
of hate speech can be distinguished and recently women emerged as victims of
abusive language both from men and women. A first study [9] focused on the
women’s experience of sexual harassment in online social network, reports the
women perception on their freedom of expression through the #mencallmethings
hashtag. More recently, as the allegations against the Hollywood producers were
made public, a similar phenomenon became viral through the #metoo hashtag.
Another investigation was presented by Fulper et al. [4]. The authors studied
the usefulness of monitoring contents published in social media in foreseeing sex-
ual crimes. In particular, they confirmed that a correlation might lay between
the yearly per capita rate of rape and the misogynistic language used in Twit-
ter. Although the problem of hate speech against women is growing rapidly,
1
https://www.statista.com/statistics/282087/number-of-monthly-active-twitter-
users/.
c Springer International Publishing AG, part of Springer Nature 2018
M. Silberztein et al. (Eds.): NLDB 2018, LNCS 10859, pp. 57–64, 2018.
https://doi.org/10.1007/978-3-319-91947-8_6
58 M. Anzovino et al.
most of the computational approaches in the state of the art are detecting abu-
sive language about ethnicity, religion, gender identity, sexual orientation and
cyberpedophilia. In this paper, we investigate the problem of misogyny detection
and categorization on social media text. The rest of the paper is structured as
follows. In Sect. 2, we describe related work about hate speech and misogyny
in social media. In Sect. 3, we propose a taxonomy for modelling the misogyny
phenomenon in online social environments, together with a Twitter dataset man-
ually labelled. Finally, after discussing experiments and the obtained results in
Sect. 4, we draw some conclusions and address future work.
Recently, several studies have been carried out in the attempt of automatically
detecting hate speech. The work presented in [15] makes a survey of the main
methodologies developed in this area. Although linguistic features may differ
within the various approaches, the classification models implemented so far are
supervised. Furthermore, a great limit is that a benchmark data set still does not
exist. Hence, the authors of [11] built and made public a corpus labelled accord-
ingly to three subcategories (hate speech, derogatory, profanity), where hate
speech is considered as a kind of abusive language. In [17] the authors described
a corpus that was labelled with tags about both racial and sexist offenses. A
recent study about the distinction of hate speech and offensive language has
been presented in [8].
Misogyny is a specific case of hate speech whose targets are women. Poland,
in her book about cybermisogyny [13], remarked among the others the prob-
lem of online sexual harassment. She deepened the matter about Gamergate
occurred in 2014 primarly bursted on 4chan and then spread across different
social networking sites. Gamergate was an organized movement which seriously
threatened lives of women belonging to the video games industry. The harass-
ment took place firstly online, then it degenerated offline. This episode confirms
that the cybermisogyny does exist, thus it is necessary to put effort in trying
to prevent similar phenomena. To the best of our knowledge only a preliminary
exploratory analysis of misogynous language in online social media has been
presented in [7]. The authors collected and manually labelleld a set of tweets as
positive, negative and neutral. However, nothing has been done from a compu-
tational point of view to recognize misogynous text and to distinguish among
the variety of types of misogyny.
Automatic Identification and Classification 59
3.2 Dataset
In order to collect and label a set of text, and subsequently address the misog-
yny detection and categorization problems, we start from the set of keywords in
[7] and we enriched these keywords with new ones as well as hashtags to down-
load representative tweets in streaming. We added words useful to represent the
different misogyny categories, when they were related to a woman, as well as
when implying potential actions against women. We also monitored tweets in
which potential harassed users might have been mentioned. These users have
been chosen because of their either public effort in feminist movements or piti-
ful past episodes of harassment online, such as Gamergate. Finally, we found
Twitter profiles who declared to be misogynistic, i.e. those user mentioning hate
against women in their screen name or the biography. The streaming download
started on 20th of July 2017 and was stopped on 30th of November 2017. Next,
among all the collected tweets we selected a subset querying the database with
the co-presence of each keyword with either a phrase or a word not used to
download tweets but still reasonable in picturing misogyny online. The labeling
phase involved two steps: firstly, a gold standard was composed and labeled by
two annotators, whose cases of disagreement were solved by a third experienced
contributor; secondly, the remaining tweets were labeled through a majority vot-
ing approach by external contributors on the CrowdFlower platform. The gold
standard has been used for the quality control of the judgements throughout
the second step. As far as it concerns the gold standard, we estimated the level
of agreement among the annotators before the resolution of the cases of dis-
agreement. The kappa coefficient [3] is the most used statistic for measuring the
degree of reliability between annotators. The need for consistency among anno-
tators immediately arises due to the variability among human perceptions. This
interagreement measure can be summarized as:
observedagreement − chanceagreement
k= (1)
1 − chanceagreement
However, considering only this statistic is not appropriate when the prevalence of
a given response is very high or very low in a specific class. In this case, the value
of kappa may indicate a low level of reliability even with a high observed pro-
portion of agreement. In order to address these imbalances caused by differences
in prevalence and bias, the authors of [2] introduced a different version of the
kappa coefficient called prevalence adjusted bias-adjusted kappa (PABAK). The
estimation of PABAK depends solely on the observed proportion of agreement
between annotators:
4 Methodology
4.1 Feature Space
Misogyny detection might be considered as a special case of abusive language.
Therefore, we chose representative features taking into account the guidelines
suggested in [11]:
1. N-grams: we considered both character and token n-grams, in particular,
from 3 to 5 characters (blank spaces included), and tokens as unigrams,
bigrams and trigrams. We chose to include these features since they usually
perform well in text classification.
2. Linguistic: misogyny category classification is a kind of stylistic classifica-
tion, which might be improved by the use of quantitative features [1,5]. Hence,
for the purpose of the current study we employed the following stylistic fea-
tures:
(a) Length of the tweet in number of characters, tweets labelled as sexual
harassment are usually shorter.
(b) Presence of URL, since a link to an external source might be an hint for
a derailing type tweet.
(c) Number of adjectives, as stereotype and objectification tweets include
more describing words.
(d) Number of mentions of users, since it might be useful in distinguishing
between individual and generic target.
3. Syntactic: we considered Bag-Of-POS. Hence, unigrams, bigrams and tri-
grams of Part of Speech tags.
4. Embedding: the purpose of this type of features is to represent texts through
a vector space model in which each word associated to a similar context lays
close to each other [10]. In particular, we employed the gensim library for
Python [14] and used the pre-trained model on a Twitter dataset3 that was
made public.
2
The dataset has been made available for the IberEval-2018 (https://amiibereval2018.
wordpress.com/) and the EvalIta-2018 (https://amievalita2018.wordpress.com/)
challenges.
3
https://www.fredericgodin.com/software/.
62 M. Anzovino et al.
Results. In Table 3 we report the results that have been obtained with the
classifiers employed in the misogyny identification. Further, in Table 4 the results
that have been obtained in the misogynistic behavior classification are shown.
As far as it concerns the misogynistic language identification, the performances
reached by the classifiers chosen are close to each other. On the contrary, about
the misogyny classification they may differ substantially. Token n-grams allow
to achieve competitive results, especially for the task of misogynistic language
identification in Twitter where indeed using all the features decreases marginally
the obtained accuracy6 . Results about misogyny classification show the difficulty
of recognizing the different phenomena of misogyny. No matter that, token n-
grams obtained competitive results if compared when also linguistic features were
employed. Although the investigated features show promising results, additional
sets related to skip character [6] could be considered in the future.
Acknowledgements. The work of the third author was partially funded by the Span-
ish MINECO under the research project SomEMBED (TIN2015-71147-C2-1-P).
5
When training the considered classifiers, we didn’t apply any feature filtering or
parameter tuning.
6
Results obtained with All Features are statistically significant (Student t-test with
p-value equal to 0.05).
64 M. Anzovino et al.
References
1. Argamon, S., Whitelaw, C., Chase, P., Hota, S.R., Garg, N., Levitan, S.: Stylistic
text classification using functional lexical features: research articles. J. Am. Soc.
Inf. Sci. Technol. 58(6), 802–822 (2007)
2. Byrt, T., Bishop, J., Carlin, J.B.: Bias, prevalence and kappa. J. Clin. Epidemiol.
46(5), 423–429 (1993)
3. Cohen, J.: A coefficient of agreement for nominal scales. Educ. Psychol. Measur.
20(1), 37–46 (1960)
4. Fulper, R., Ciampaglia, G.L., Ferrara, E., Ahn, Y., Flammini, A., Menczer, F.,
Lewis, B., Rowe, K.: Misogynistic language on Twitter and sexual violence. In:
Proceedings of the ACM Web Science Workshop on Computational Approaches to
Social Modeling (ChASM) (2014)
5. HaCohen-Kerner, Y., Beck, H., Yehudai, E., Rosenstein, M., Mughaz, D.: Cuisine:
classification using stylistic feature sets and/or name-based feature sets. J. Assoc.
Inf. Sci. Technol. 61(8), 1644–1657 (2010)
6. HaCohen-kerner, Y., Ido, Z., Ya’akobov, R.: Stance classification of tweets
using skip char Ngrams. In: Altun, Y., Das, K., Mielikäinen, T., Malerba, D.,
Stefanowski, J., Read, J., Žitnik, M., Ceci, M., Džeroski, S. (eds.) ECML PKDD
2017. LNCS (LNAI), vol. 10536, pp. 266–278. Springer, Cham (2017). https://doi.
org/10.1007/978-3-319-71273-4 22
7. Hewitt, S., Tiropanis, T., Bokhove, C.: The problem of identifying misogynist
language on Twitter (and other online social spaces). In: Proceedings of the 8th
ACM Conference on Web Science, pp. 333–335. ACM, May 2016
8. Davidson, T., Warmsley, D., Macy, M., Weber, I.: Automated hate speech detection
and the problem of offensive language. In: Proceedings of the 12th International
AAAI Conference on Web and Social Media (2017)
9. Megarry, J.: Online incivility or sexual harassment? Conceptualising women’s expe-
riences in the digital age. In: Women’s Studies International Forum, vol. 47, pp.
46–55. Pergamon (2014)
10. Le, Q., Mikolov, T.: Distributed representations of sentences and documents. In:
International Conference on Machine Learning, pp. 1188–1196, January 2014
11. Nobata, C., Tetreault, J., Thomas, A., Mehdad, Y., Chang, Y.: Abusive language
detection in online user content. In: Proceedings of the 25th International Confer-
ence on World Wide Web, pp. 145–153. International World Wide Web Conferences
Steering Committee (2016)
12. Parker, R.I., Vannest, K.J., Davis, J.L.: Effect size in single-case research: a review
of nine nonoverlap techniques. Behav. Modif. 35(4), 303–322 (2011)
13. Poland, B.: Haters: Harassment, Abuse, and Violence Online. University of
Nebraska Press, Lincoln (2016)
14. Rehurek, R., Sojka, P.: Software framework for topic modelling with large cor-
pora. In: Proceedings of the LREC 2010 Workshop on New Challenges for NLP
Frameworks (2010)
15. Schmidt, A., Wiegand, M.: A survey on hate speech detection using natural lan-
guage processing. In: Proceedings of the Fifth International Workshop on Natural
Language Processing for Social Media. Association for Computational Linguistics,
Valencia, Spain, pp. 1–10 (2017)
16. Sebastiani, F.: Machine learning in automated text categorization. ACM Comput.
Surv. (CSUR) 34(1), 1–47 (2002)
17. Waseem, Z., Hovy, D.: Hateful symbols or hateful people? predictive features for
hate speech detection on Twitter. In: SRW@ HLT-NAACL, pp. 88–93 (2016)