2003 Neuro Computing
2003 Neuro Computing
Abstract
Fake news, which can be defined as intentionally and verifiably false news, has a strong
influence on critical aspects of our society. Manual fact-checking is a widely adopted ap-
proach used to counteract the negative effects of fake news spreading. However, manual
fact-checking is not sufficient when analysing the huge volume of newly created information.
Moreover, the number of labeled datasets is limited, humans are not particularly reliable
labelers and databases are mostly in English and focused on political news. To solve these
issues state-of-the-art machine learning models have been used to automatically identify fake
news. However, the high amount of models and the heterogeneity of features used in litera-
ture often represents a boundary for researchers trying to improve model performances. For
this reason, in this systematic review, a taxonomy of machine learning and deep learning
models and features adopted in Content-Based Fake News Detection is proposed and their
performance is compared over the analysed works. To our knowledge, our contribution is the
first attempt at identifying, on average, the best-performing models and features over mul-
tiple datasets/topics tested in all the reviewed works. Finally, challenges and opportunities
in this research field are described with the aim of indicating areas where further research is
needed.
Keywords: Fake News Detection, Content Based Fake News Detection, Content-Based
Features
∗
Corresponding author
Email addresses: [Link]@[Link] (Nicola Capuano), gfenza@[Link] (Giuseppe Fenza),
loia@[Link] (Vincenzo Loia), [Link]@[Link] (Francesco David Nota)
1. Introduction
In the last years increasing attention to the phenomena of fake news has grown. Fake news,
which can be defined as intentionally and verifiably false news [1], is considered to destabilize
democracies, weaken the trust that citizens have in public institutions, and have a strong
influence on critical aspects of our society such as elections, the economy, and public opinion
(e.g. on wars). To counteract the negative effects of the spreading of fake news, several
initiatives have started to appear; one widespread approach to studying and analysing fake
news is fact-checking.
Content-Based Fake News Detection (CBFND) has the purpose of assessing news inten-
tion as a set of quantifiable features, often machine learning features, extracted from news
content. CBFND is a critical tool for identifying news harmfulness. Specifically, CBFND
can be considered: 1) a fake news detector; 2) a complement to other fake news detection
techniques (such as credibility assessment and network propagation pattern analysis) [1] [2].
In this systematic review, we consider all the features that can be extrapolated from text con-
tent, excluding images or audio as they are not always present in news or social network posts.
In table 1 the differences with other reviews in the field of fake news detection are highlighted.
Specifically, compared to other reviews, this work is the only one that does an extensive eval-
uation of features and models as well as their performances on multiple datasets; moreover,
some reviews focus only on a subset of models (e.g. Natural Language Processing or Deep
Learning) or topics (e.g. covid-19), this work, instead, focuses on textual content-based fea-
tures, but does not exclude any kind of algorithms or topics. Finally, some of the existing
reviews tend to be too general and do not cover with enough detail the types of features and
models that, in most cases, provide better performances. To our knowledge, this is the first
attempt of identifying, on average, the best performing models and features over multiple
datasets/topics tested in multiple works.
Table 1: Differences between this work and other reviews in the literature. When using the symbol ”x” we
refer to a complete absence of a characteristic in the work. Written text may represent a complete or partial
presence of the characteristic.
Identifier Algorithms description features description datasets description performance analysis year
this It includes a description It describes the features It shows most of the It does an extensive 2022
work of both Machine Learn- and organizes them in available datasets in performance analysis on
ing (ML) and Deep groups literature with links all works by showing
Learning (DL) models. to them best-performing algo-
Specifically, traditional rithms and features over
ML, ensemble ML, DL, multiple datasets
pre-trained DL, and
Mixed models
[6] It shows the accuracy The study shows accu- 2022
performance of individ- × × racy of individual works
ual works over machine over machine and deep
and deep learning algo- learning algorithms,
rithms but there is no there is no average
short description related performance evaluation
to each of them over algorithms in lit-
erature and there is no
reference to features.
[7] It lists the algorithms There is a reference to 2022
used, but no description feature groups (e.g. lin- × ×
is provided for each of guistic) but no descrip-
them tion of individual fea-
tures and of the group
itself is provided
[8] It is focused only Natu- The performances are 2021
ral Language Processing × × represented by a small
models test conducted on some
traditional machine
learning algorithms on
one dataset
[9] It lists machine and It lists features used in It lists datasets for 2021
deep learning models the reviewed works re- fake news detection ×
used in the reviewed lated to covid-19, but but with no descrip-
works related to covid- they do not describe tion related to them
19, but they do not de- them
scribe them
[10] It provides a description 2021
only of traditional Ma- × × ×
chine learning (exclud-
ing also ensemble learn-
ing models)
[11] Analyses in detail a 2020
strong minority of Deep × × ×
Learning algorithms
[12] A minority of the algo- A minority of the fea- Some datasets used 2019
rithms used in literature tures used in literature in literature are listed ×
are shown, but there is listed and only some and described
is no description associ- of them are described
ated to them
[13] Some algorithms used in An extensive descrip- The individual accuracy 2018
literature are cited but × tion of the datasets is of some algorithms is
not described provided shown
Because of the heterogeneity and complexity of the models and features related to Automatic
Fake News Detection our systematic review, in contrast with existing works in literature,
focuses on content-based models and features and therefore excludes also hybrid approaches.
This work is structured similarly to other reviews of the sector, with the goal of facilitating
the comparison among them. The expected benefit is to provide guidelines to researchers
facing the issue of content-based fake news detection and to system developers dealing with
the implementation of accurate content-based fake news detectors.
The remainder of the work is structured as follows: Section 2. introduces the methodology
adopted in the systematic literature review; Section 3 answers to the research questions and
presents the results. Section 4 discusses the findings, and shows opportunities and gaps in
the area. Finally, Section 5 includes the conclusion, highlighting the scientific contributions
of the work, and the challenges to be addressed in the future.
2. Methodology
In this section, the adopted procedures, methods, and decisions taken into account in the
development of this systematic literature review are presented.
Acronym Portal
SCOPUS SCOPUS (Elsevier database)
ScienceDirect ScienceDirect (Elsevier database)
IEEE IEEE Xplore Digital Library
ACM ACM Digital Library
Table 4: Selected Studies.
Study name
Liar, Liar Pants on Fire: A New Benchmark Dataset for Fake News Detection [16]
Fake news detection using naive Bayes classifier [17]
Detection of Online Fake News Using N-Gram Analysis and Machine Learning Techniques [18]
Automatic Detection of Fake News [19]
A Tool for Fake News Detection [20]
Fake News Detection: A Deep Learning Approach [21]
A Benchmark Study on Machine Learning Methods for Fake News Detection [22]
Comparative Performance of Machine Learning Algorithms for Fake News Detection [23]
Which machine learning paradigm for fake news detection? [24]
Semantic Fake News Detection: A Machine Learning Perspective [25]
Supervised Learning for Fake News Detection [26]
Detecting Fake News using Machine Learning and Deep Learning Algorithms [27]
Behind the cues: A benchmarking study for fake news detection [28]
Deep learning methods for Fake News detection [29]
Machine Learning Methods for Fake News Classification [30]
Comparison of Various Machine Learning Models for Accurate Detection of Fake News [31]
A Closer Look at Fake News Detection: A Deep Learning Perspective [3]
Fake News Detection with Semantic Features and Text Mining [32]
dEFEND: Explainable Fake News Detection [33]
Learning Hierarchical Discourse-level Structure for Fake News Detection [34]
Sentiment Aware Fake News Detection on Online Social Networks [35]
Content Based Fake News Detection Using N-Gram Models [36]
Fake News Detection Using Machine Learning Ensemble Methods [37]
Performance Comparison of Machine Learning Classifiers for Fake News Detection [38]
Fake news detection in multiple platforms and languages [39]
FNDNet – A deep convolutional neural network for fake news detection [40]
Fake News Detection using Deep Learning [41]
Fake News Detection with Different Models [42]
Fake News Detection: An Ensemble Learning Approach [43]
Fake News Early Detection: A Theory-driven Model [44]
Fake News Detection Using Machine Learning Approaches [45]
Linguistic feature based learning model for fake news detection and classification [46]
A benchmark study of machine learning models for online fake news detection [47]
Constraint 2021: Machine Learning Models for COVID-19 Fake News Detection Shared Task [48]
Fake news detection: A hybrid CNN-RNN based deep learning approach [2]
WELFake: Word Embedding Over Linguistic Features for Fake News Detection [49]
Fake Detect: A Deep Learning Ensemble Model for Fake News Detection [5]
Transformer based Automatic COVID-19 Fake News Detection System [50]
FakeBERT: Fake news detection in social media with a BERT-based deep learning approach [51]
Evaluating Deep Learning Approaches for Covid19 Fake News Detection [52]
In Fig. 2 details related to the filtering process are presented. After the process, the 40
articles selected were used to elaborate the systematic literature review. In Table 4, the
selected studies are shown and sorted from the less recent to the most recent.
Identifier Description
QC1 Is the goal of the study in the article clear?
QC2 Does the study include the literature review or background?
QC3 Is the main contribution of the study clear?
QC4 Does the article explains a research methodology?
QC5 Does the study show research results?
QC6 Is the conclusion of the study related to the research aim?
QC7 Does the article include future works and developments?
3. Results
This section answers the proposed research questions. To this end, the results obtained from
the 40 final reviewed works are shown and discussed.
The groups of algorithms are: 1) Machine Learning (ML) 2) Ensemble Machine Learning
3) Deep Learning (DL) 4) pre-trained Deep Learning 5) Mixed models (models that use a
mix of multiple sub-models). To give a better context we include some basic information
related to machine learning in general and the task associated with CBFND; furthermore,
for each group of algorithms, we provide an additional short introductory description.
The two most important higher-level classes of machine learning algorithms are supervised
learning, which relies on labeled input and output training data, and unsupervised learning,
which processes unlabeled or raw data. In this review we focus our attention on supervised
learning algorithms; in particular, a supervised machine learning algorithm has the goal of
predicting output values from given input data. The two major tasks of supervised machine
learning algorithms are classification and regression. The most important difference between
regression and classification is that while regression helps predict a continuous quantity, clas-
sification predicts discrete class labels. CBFND is a classification task, and in its simplest
form is associated with predicting whether a text has to be considered fake news or not. The
most important metrics to measure the performance of classification algorithms are accuracy,
precision, recall, and f1score.
Accuracy is the ratio of correct predictions out of all predictions. It is not a good metric
when the data set is unbalanced. Using accuracy in such scenarios can result in a misleading
interpretation of results.
Precision is the ratio of true positives over the sum of false positives and true negatives. The
question that this metric answer is: What proportion of positive identifications was actually
correct?
Recall is the ratio of correctly predicted outcomes to all predictions. The question recall
answers is: What proportion of actual positives was identified correctly?
F1 Score combines precision and recall into one single metric that ranges from 0 to 1; being
the harmonic mean of precision and recall and it is considered a better measure than accuracy.
Traditional Machine learning uses algorithms to parse data, learn from that data, and make
informed decisions based on what it has learned. These algorithms have been widely adopted
for solving the task of fake news detection, as they are fast to train and implement, do not
require dedicated hardware, and achieve good results on small datasets. In table 6 we list
the algorithms used in the reviewed papers as well as a short description of each algorithm.
Table 6: Traditional Machine Learning algorithms.
Identifier Studies
CNN+LSTM [16, 37, 2, 50]
LSTM + Attention [35]
BI-LSTM + Attention [34, 37, 5]
GRU + Attention [25]
CNN+BI-LSTM + attention [3]
CNN+Bi-GRU [34]
CNN+HAN [37]
CNN + BERT [51]
LSTM + BERT [51]
Naive Bayes + BERT [51]
3.2. RQ2 What features are used in CBFND machine learning models?
Numerous features are used in the task of CBFND, in the analysed studies often features are
grouped into several categories such as linguistic, grammatical, readability, and sentiment
features. Although most of these groups are associated with similar features we noticed that
some works include the same feature in different groups, for this reason, we propose here
a simplified division of the features into groups intending to provide more intuitive feature
groups.
Each of the tables below represents a feature group; moreover, for each feature in the group,
we provide a brief description as well as the studies that use it. We omit the description of
the feature in the table whenever a feature identifier is self-explaining; In addition, although
we consider the list of features below exhaustive we do not exclude the possibility of having
missed other important features or variants of features that are used in literature but not in
the considered studies.
• Table 13 - Word tagging: Word tagging features are features that associate tags related
to words. The most used features are Part Of Speech tagging and Named Entities.
Both of these are usually calculated by employing a machine learning algorithm.
• Table 16 - Readability: represent the level of ease of reading the text. Readability fo-
cuses on textual content such as lexical, semantical, syntactical, and discourse cohesion
analysis. In the context of Machine Learning and Fake News detection, several indices
are used to extrapolate readability-related features, most of which can be extracted by
using the Linguistic Inquiry and Word Count (LIWC) software. 1
• Table 17 - Sentiment and emotion-based: are a group formed by all those features that
extrapolate high-level sensations that a text expresses.
• Table 18 - Other: contains all the features to which we did not assign any group but
that have been used in some of the analysed studies.
1
[Link]
Table 12: Embedding features.
Identifier Studies
Average number of words per sentence [28, 39, 44, 49]
Average number of syllables per word [28]
Average number of characters per word [44]
Average word length [37, 44, 45, 47]
Percentage of adverbs [28, 39, 49]
Percentage of adjectives [47, 28, 39, 49]
Percentage of articles [28]
Percentage of exclamation marks [39]
Percentage of negations [28]
Percentage of nouns [39]
Percentage of preposition [47, 28]
Percentage of question marks [39]
Percentage of uppercase characters [39]
Percentage of verbs [47]
Percentage of words longer than 6 letters [28]
Table 15: Absolute quantity features.
Identifier Studies
Number of characters [19, 39, 46]
Number of uppercase/special characters [46, 49]
Number of words [28, 37, 39, 45, 46, 47, 49]
Number of long words [19, 28, 44]
Number of words in the title [46]
Number of syllables [19, 28, 44, 49]
Number of numbers [37, 44, 45, 47]
Number of determinants [49]
Number of adverbs [19, 46, 49]
Number of verbs [19, 26, 44, 46, 49]
Number of nouns [19, 44, 46]
Number of pronouns [19, 46]
Number of adjectives [19, 26, 37, 44, 45, 46, 47, 49]
Number of articles [49]
Number of long/short sentences [28, 49]
Number of sentences [28, 44, 49]
Number of paragraphs [19, 44]
Number of conjunctions [28]
Number of punctuation characters [47, 26, 44]
Number of exclamation marks [37, 44, 47]
Article length [37, 45, 47]
All the readability features listed in the table 16 produce an approximate representation of
the education grade level needed to comprehend the text; however, they use different formulas
and algorithms.
Table 16: Readability features.
Identifier Description Studies
Automated read- [19, 44,
ability index characters words 46, 48, 49]
4.71 × ( ) + 0.5( ) − 21.43
words sentences
where ”characters” is the number of letters and numbers,
words the number of spaces, and sentences the number of
sentences
Coleman Liau index [44, 46,
(0.0588 × letters) − (0.296 × sentences) − 15.8 48]
where ”letters” is the average number of letters every 100
words and sentences is the average number of sentences ev-
ery 100 words
Flesch Kincaid [19, 28,
Score words syllabes 44, 46, 48]
4.71 × ( ) + 11.8( ) − 15.59
sentences words
where words, sentences, and syllables are the numbers of
them.
Flesch reading ease [19, 44,
words syllabes 46, 48]
206.835 − 1.015( ) − 84.6( )
sentences words
where words, sentences, and syllables are the numbers of
them.
Gunning fog index [19, 44,
words complexwords 46, 49]
0.4[( ) + 100( ))]
sentences words
where words, sentences, and complex words are the number
of them, and, where complex words are the one consisting
of three or more syllables
Lensear write for- It is calculated as follows: a) Count a 100-word sample; [46]
mula b) Count one point for all one-syllable words; c) Give three
points for each sentence in the 100-word sample to the near-
est period or semicolon; d) to obtain the final score, add
together the one-syllable word count and the three points
for each sentence.
Smog index Is calculated by by using a piece of text which is 30 sentences [46, 49]
or longer and doing the following: a) Counting ten sentences
near the beginning of the text, 10 in the middle, and ten near
the end, totaling 30 sentences; b) counting every word with
three or more syllables; c) square-rooting the number and
rounding it to the nearest 10; d) Adding three to this figure
Table 17: Sentiment/emotion-based features.
In table 20, we list algorithms achieving an average accuracy higher than 80% as well as the
number of reviewed studies in which the algorithm was used and the number of datasets on
which it has been tested. Considering the relevance of the number of studies and datasets on
which the algorithm has been tested, we can state that the more robust results are given by
Gradient Boosting, eXtreme Gradient Boosting, Multilayer perceptron, and Naive Bayes. It
is also important to mention that the highest accuracy is given by XLNet, ALBERT, LSTM
with BERT, and the Passive Aggressive Classifier; however, the number of analysed studies
and datasets these models are based on is only one. As soon as enough articles on CBFND
will include recall, precision, and f1score this simple but effective comparison method can be
reused to evaluate the performances of the analysed algorithms with respect to the full set
of metrics (accuracy, recall, specificity, and f1score).
Few literature works analyse these algorithms in the context of CBFND and in general in
the field of Fake News detection, indicating the need for more research to confirm the per-
formance of these models by testing them on multiple datasets. The high performance of
Gradient Boosting over multiple datasets and studies could be justified by the model’s high
generalization capability (given by several weak learners) as well as the characteristics of hav-
ing better weak learners in the process of training. Transformer pre-trained models (XLNet,
BERT, and its variants) have shown the best performance compared to other Deep Learning
algorithms; this result is expected as these models have shown high generalization capabili-
ties and have been tested and used successfully over many different research areas involving
Natural Language Processing.
In table 21 we show all the features being present in more than one study, tested on more
than one dataset, and with an average accuracy higher than 80%. Considering the number
of studies and datasets on which each feature has been tested we can state that the stronger
results are given by Bag of Words, Automated Readability Index, Number of verbs, and
TF-IDF. Particular attention should be given to readability features such as the Coleman
Liau index and Smog Index; however, more research is needed to confirm the results of these
last features.
The most performant features tend to be language and topic invariant, and, therefore, provide
higher performances on average compared to other features. Interestingly quantity-based fea-
tures such as the number of uppercase characters, verbs, adverbs, and syllables are associated
with higher performances, as well as readability scores, which are mostly formulas composed
of quantity-based features. Furthermore, it is equally important to highlight the TF-IDF and
bag of words features, which, are showing surprisingly high average performance over many
datasets. Given the analysis of features in table 21 we can state that the most performant
features will show consistently high performance in different datasets. However, we do not
exclude the possibility of improving detection results with the advancements of automatic
feature extraction techniques in Deep Learning or by introducing new features.
Table 21: Features with average accuracy higher than 80%, tested on more than 1 dataset and used in more
than 1 study
The rationale for the classification of results as promising and robust is based on the in-
tuition that as the number of studies that apply a specific feature/algorithm increases and
the number of databases on which it is tested increases as well also the robustness of the re-
sult increases; conversely a performance resulting only from one or two of the reviewed study
has lower robustness and must be further verified by future works, in particular for what
concerns features and algorithms that are not commonly used in literature at the moment of
writing.
Specifically, the algorithms with ”promising results” are the deep learning algorithms XLNet,
ALBERT, and LSTM with BERT; while the ones with ”robust results” are Gradient Boost-
ing, eXtreme Gradient Boosting, Multilayer perceptron, and Naive Bayes. In addition, the
Coleman Liau Index can be classified as ”promising results”, while Automated Readability
Index, Bag of Words, Number of verbs, and TF-IDF can be classified as ”robust results”.
Standard for representing fake and real news in a Dataset: As shown in section
3.3 a wide variety of datasets is used in the task of fake news detection, all of them providing
information to perform the CBFND task. However, there is a need for a standard in how the
process of data collection, verification, and storage. Often the source of data, the verification
process, and the lack of diversity in terms of topics related to the database (e.g. political top-
ics) are barriers to progress in this research field. Furthermore, it is critical that studies use
datasets that respect the principles of Open and FAIR (Findable, Accessible, Interoperable,
Reusable) data; in particular for what concerns interoperability and the datasets metadata
[61, 62].
Continuous learning models: The analysed studies have focused on the problem of iden-
tifying fake news by training models over existing databases. However, the continuous evolu-
tion of malicious techniques employed by fake news spreaders requires an adaptive continuous
learning model. Creating a model able to adapt to changing fake news characterization pat-
terns is a real-world necessity that is currently not discussed adequately in literature and,
therefore, an opportunity for feature research.
Association with context: The detection of fake news based on content is critically impor-
tant, however, it is often necessary to associate the news with a context in order to verify it.
Therefore, there is a need to have models that are able to relate contexts/topics to news con-
tent. One interesting research outcome of studying context and content together is verifying
if the performance of content-based features in relation to the context leads to an increase
or decrease for each feature.
Attacks on natural language learning: The use of Natural Language Processing to
identify fake news is vulnerable to attacks on the machine and deep learning models [63].
The distortion of facts, the exchange between subject and object, and the confusion of causes
are three types of attacks that could lead to poor performance. The distortion is to exaggerate
or modify some words. Textual elements can be distorted to lead to a false interpretation.
The exchange between subject and object aims to confuse the reader between those who
practice and those who suffer the reported action. The attack of confusion of cause consists
of creating non-existent causal relations between two independent events or cutting parts of
a story, leaving only the parts that the attacker wishes to present to the reader.
4.2. Limitations
In this section, we provide the limitation of this work. This systematic review is limited to
aspects related to the quality assessment of articles, and, potentially, a partial selection of
works. An additional limitation is the search string and the selected scientific databases;
both of them may be a restriction factor with respect to the analysed works. In this work we
have chosen accuracy as a metric to evaluate performance as it is the only metric available in
all the analysed studies; considering only accuracy is a limit of this work and must be taken
into consideration while reading the analysis. Furthermore, for simplicity, the accuracy has
been evaluated over an algorithm, or a feature but not on both at the same time. As more
articles are published future research will be able to analyse performance differences also
from metrics other than accuracy (e.g. recall, precision and f1score).
This work considers Machine/Deep Learning models without considering the performance
variability deriving from the choice of hyperparameters. Currently, the number of works us-
ing deep learning algorithms is limited and therefore our performance results on deep learning
can be considered only as promising rather than robust results; however, future research will
be able to cover this gap as the number of studies employing DL for CBFND increases and the
databases become bigger and more standardized. In addition, as more datasets are published,
more specialized benchmark studies, each one dealing with a single algorithm (or a single
class of algorithms) tested over multiple datasets and hyperparameters, could be performed.
However, deep learning application in this field is still in the early ages and the number of
large multi-topic datasets is still insufficient. This work is limited to content-based features
and models; in literature, there exist works dealing with all three categories defined in Figure
1, but they provide a high-level description of features and models. Our work differs in that
it provides greater details related to the content-based category. This approach allows the
comparison of performances on multiple datasets and provides more detailed knowledge to
researchers and system developers. The adopted approach is a potential alternative for future
systematic reviews related to the user and social-context-based categories.
5. Conclusions
In this review state-of-the-art algorithms and features associated with Content Based Fake
News Detection (CBFND) have been analysed and identified. To answer the research ques-
tions, a systematic literature review methodology, that allowed to select and organize the
works, was adopted. Specifically, the features and machine learning algorithms were de-
scribed and associated with the reviewed work. Furthermore, a description of the most
important datasets used is provided. For each work, the best-performing algorithms and fea-
tures were considered and an average performance analysis of the extrapolated data has been
conducted. The outcome of the analysis shows which features and models perform better
over multiple datasets. More in detail, we found the most performing models to be Gra-
dient Boosting, eXtreme Gradient Boosting, Multilayer perceptron, and Naive Bayes, and
the most performing features to be Automated Readability Index, Bag of Words, Number
of words, and TF-IDF. Furthermore, we identify as promising the models XLNet, ALBERT,
and LSTM with BERT as well as the Coleman Liau Index feature; however, more research
is needed to confirm their effect on CBFND. The analysis facilitates the work of researchers
in improving CBFND performance and, more importantly, indicates the models and features
that are more prone to have high performance on multiple datasets and, therefore, have a
higher probability to perform well also on real-world CBFND systems. Finally, challenges
and opportunities in this research field are described to indicate areas where further research
is needed.
References
[1] X. Zhang, A. A. Ghorbani, An overview of online fake news: Characterization, detection,
and discussion, Information Processing & Management 57 (2020) 102025.
[2] J. A. Nasir, O. S. Khan, I. Varlamis, Fake news detection: A hybrid cnn-rnn based deep
learning approach, International Journal of Information Management Data Insights 1
(2021) 100007.
[3] A. Abedalla, A. Al-Sadi, M. Abdullah, A closer look at fake news detection: A deep
learning perspective, in: Proceedings of the 2019 3rd International Conference on Ad-
vances in Artificial Intelligence, 2019, pp. 24–28.
[4] S. A. Alameri, M. Mohd, Comparison of fake news detection using machine learning
and deep learning techniques, in: 2021 3rd International Cyber Resilience Conference
(CRC), IEEE, 2021, pp. 1–6.
[7] M. Lahby, S. Aqil, W. Yafooz, Y. Abakarim, Online fake news detection using ma-
chine learning techniques: A systematic mapping study, Combating Fake News with
Computational Intelligence Techniques (2022) 3–37.
[8] C. Agrawal, A. Pandey, S. Goyal, A survey on role of machine learning and nlp in
fake news detection on social media, in: 2021 IEEE 4th International Conference on
Computing, Power and Communication Technologies (GUCON), IEEE, 2021, pp. 1–7.
[11] A. Chokshi, R. Mathew, Deep learning and natural language processing for fake news
detection: a survey (2020).
[12] F. Cardoso Durier da Silva, R. Vieira, A. C. Garcia, Can machines learn to detect fake
news? a survey focused on social media (2019).
[13] R. Oshikawa, J. Qian, W. Y. Wang, A survey on natural language processing for fake
news detection, arXiv preprint arXiv:1811.00770 (2018).
[14] K. Shu, A. Sliva, S. Wang, J. Tang, H. Liu, Fake news detection on social media: A
data mining perspective, ACM SIGKDD explorations newsletter 19 (2017) 22–36.
[16] W. Y. Wang, ” liar, liar pants on fire”: A new benchmark dataset for fake news detection,
arXiv preprint arXiv:1705.00648 (2017).
[17] M. Granik, V. Mesyura, Fake news detection using naive bayes classifier, in: 2017 IEEE
first Ukraine conference on electrical and computer engineering (UKRCON), IEEE, 2017,
pp. 900–903.
[18] H. Ahmed, I. Traore, S. Saad, Detection of online fake news using n-gram analysis
and machine learning techniques, in: International conference on intelligent, secure, and
dependable systems in distributed and cloud environments, Springer, 2017, pp. 127–138.
[20] B. Al Asaad, M. Erascu, A tool for fake news detection, in: 2018 20th International
Symposium on Symbolic and Numeric Algorithms for Scientific Computing (SYNASC),
IEEE, 2018, pp. 379–386.
[21] A. Thota, P. Tilak, S. Ahluwalia, N. Lohia, Fake news detection: a deep learning
approach, SMU Data Science Review 1 (2018) 10.
[27] E. M. Mahir, S. Akhter, M. R. Huq, et al., Detecting fake news using machine learning
and deep learning algorithms, in: 2019 7th International Conference on Smart Comput-
ing & Communications (ICSCC), IEEE, 2019, pp. 1–5.
[29] V. M. Krešňáková, M. Sarnovskỳ, P. Butka, Deep learning methods for fake news
detection, in: 2019 IEEE 19th International Symposium on Computational Intelligence
and Informatics and 7th IEEE International Conference on Recent Achievements in
Mechatronics, Automation, Computer Sciences and Robotics (CINTI-MACRo), IEEE,
2019, pp. 000143–000148.
[30] P. Ksieniewicz, M. Choraś, R. Kozik, M. Woźniak, Machine learning methods for fake
news classification, in: International Conference on Intelligent Data Engineering and
Automated Learning, Springer, 2019, pp. 332–339.
[31] K. Poddar, K. Umadevi, et al., Comparison of various machine learning models for
accurate detection of fake news, in: 2019 Innovations in Power and Advanced Computing
Technologies (i-PACT), volume 1, IEEE, 2019, pp. 1–5.
[32] P. Bharadwaj, Z. Shao, Fake news detection with semantic features and text mining,
International Journal on Natural Language Computing (IJNLC) Vol 8 (2019).
[33] K. Shu, L. Cui, S. Wang, D. Lee, H. Liu, defend: Explainable fake news detection, in:
Proceedings of the 25th ACM SIGKDD international conference on knowledge discovery
& data mining, 2019, pp. 395–405.
[34] H. Karimi, J. Tang, Learning hierarchical discourse-level structure for fake news detec-
tion, arXiv preprint arXiv:1903.07389 (2019).
[35] O. Ajao, D. Bhowmik, S. Zargari, Sentiment aware fake news detection on online social
networks, in: ICASSP 2019-2019 IEEE International Conference on Acoustics, Speech
and Signal Processing (ICASSP), IEEE, 2019, pp. 2507–2511.
[36] H. E. Wynne, Z. Z. Wint, Content based fake news detection using n-gram models,
in: Proceedings of the 21st International Conference on Information Integration and
Web-based Applications & Services, 2019, pp. 669–673.
[37] I. Ahmad, M. Yousaf, S. Yousaf, M. O. Ahmad, Fake news detection using machine
learning ensemble methods, Complexity 2020 (2020).
[38] N. Smitha, R. Bharath, Performance comparison of machine learning classifiers for fake
news detection, in: 2020 Second International Conference on Inventive Research in
Computing Applications (ICIRCA), IEEE, 2020, pp. 696–700.
[39] P. H. A. Faustini, T. F. Covões, Fake news detection in multiple platforms and languages,
Expert Systems with Applications 158 (2020) 113503.
[41] S. H. Kong, L. M. Tan, K. H. Gan, N. H. Samsudin, Fake news detection using deep
learning, in: 2020 IEEE 10th Symposium on Computer Applications & Industrial Elec-
tronics (ISCAIE), IEEE, 2020, pp. 102–107.
[43] A. Agarwal, A. Dixit, Fake news detection: an ensemble learning approach, in: 2020
4th International Conference on Intelligent Computing and Control Systems (ICICCS),
IEEE, 2020, pp. 1178–1183.
[44] X. Zhou, A. Jain, V. V. Phoha, R. Zafarani, Fake news early detection: A theory-driven
model, Digital Threats: Research and Practice 1 (2020) 1–25.
[45] Z. Khanam, B. Alwasel, H. Sirafi, M. Rashid, Fake news detection using machine
learning approaches, in: IOP Conference Series: Materials Science and Engineering,
volume 1099, IOP Publishing, 2021, p. 012040.
[46] A. Choudhary, A. Arora, Linguistic feature based learning model for fake news detection
and classification, Expert Systems with Applications 169 (2021) 114171.
[48] T. Felber, Constraint 2021: Machine learning models for covid-19 fake news detection
shared task, arXiv preprint arXiv:2101.03717 (2021).
[49] P. K. Verma, P. Agrawal, I. Amorim, R. Prodan, Welfake: word embedding over lin-
guistic features for fake news detection, IEEE Transactions on Computational Social
Systems 8 (2021) 881–893.
[50] S. Gundapu, R. Mamidi, Transformer based automatic covid-19 fake news detection
system, arXiv preprint arXiv:2101.00180 (2021).
[51] R. K. Kaliyar, A. Goswami, P. Narang, Fakebert: Fake news detection in social media
with a bert-based deep learning approach, Multimedia tools and applications 80 (2021)
11765–11788.
[52] A. Wani, I. Joshi, S. Khandve, V. Wagh, R. Joshi, Evaluating deep learning approaches
for covid19 fake news detection, in: International Workshop on Combating Online
Hostile Posts in Regional Languages during Emergency Situation, Springer, 2021, pp.
153–163.
[54] G. C. Santia, J. R. Williams, Buzzface: A news veracity dataset with facebook user
commentary and egos, in: Twelfth international AAAI conference on web and social
media, 2018.
[55] A. Zubiaga, M. Liakata, R. Procter, G. Wong Sak Hoi, P. Tolmie, Analysing how people
orient to and spread rumours in social media by looking at conversational threads, PloS
one 11 (2016) e0150989.
[57] T. de Melo, C. M. Figueiredo, A first public dataset from brazilian twitter and news on
covid-19 in portuguese, Data in brief 32 (2020) 106179.
[63] Z. Zhou, H. Guan, M. M. Bhat, J. Hsu, Fake news detection via nlp is vulnerable to
adversarial attacks, arXiv preprint arXiv:1901.09657 (2019).