doxa.comunicación | nº 33, pp. 321-346 | July-December of 2021

ISSN: 1696-019X / e-ISSN: 2386-3978

Research on disinformation: topics, methodology and impact. A systematic literature review (2016-2020)1

Agenda temática, metodologías e impacto de la investigación sobre desinformación. Revisión sistemática de la literatura (2016-2020)

David García-Marín. PhD from UNED (Universidad Nacional de Educación a Distancia, Spain). Lecturer in the Department of Journalism & Corporate Communication at Universidad Rey Juan Carlos, where he lectures on new technologies and information society, as well as radio production. He teaches digital communication and transmedia journalism at UNED, where he has directed several courses on digital media and disinformation. Researcher on the Cátedra Jean Monnet EUDFAKE: EU, disinformation and fake news financed by the European Commission’s Erasmus+ programme, and on the Rationality and Counter-knowledge Project Epistemology of the detection of falsehoods in news stories financed by the Spanish Ministry of Science, Innovation & Universities. His latest book is La posverdad. Una cartografía de los medios, las redes y la política (Gedisa, 2019).

University Rey Juan Carlos, Madrid, Spain

[email protected]

ORCID: 0000-0002-4575-1911

How to cite this article:

García-Marín, D. (2021). Research on disinformation: topics, methodology and impact. A systematic literature review (2016-2020). Doxa Comunicación, 33, pp. 321-346.

https://doi.org/10.31921/doxacom.n33a854

Received: 20/05/2021 - Accepted: 23/09/2021 - Early access: 26/10/2021 - Published: 01/12/2021

Abstract:

Since 2016, the analysis of disinformation and fake news has become one of the main trends in scientific research fields. In order to discover the topics, areas of knowledge, countries of origin and methodologies of these studies, as well as to analyse their impact, a systematic literature review was conducted in the Web of Science database. A total of 605 articles published between 2016 and 2020 were analysed. Nonparametric descriptive and inferential statistical methods (Kruskal-Wallis and Mann-Whitney U tests) were applied with the aim of discovering significant differences in the number of references received for each of the aforementioned variables. The most frequent topics are possible solutions to tackle disinformation, but papers on the propagation patterns of false content achieve greater impact. There is a notable presence of quantitative studies, mainly from English-speaking sources. The methodological approach is the only variable that can be considered relevant to the impact of this research.

Keywords:

Disinformation; fake news; systematic literature review; research methodology; scientific impact.

Recibido: 20/05/2021 - Aceptado: 23/09/2021 - En edición: 26/10/2021 - Publicado: 01/12/2021

Resumen

Desde el año 2016, el análisis de la desinformación y las fake news se ha convertido en una de las principales tendencias en la investigación científica producida desde diferentes campos del saber. A fin de conocer cuáles son los temas, las áreas del conocimiento, los países de procedencia y los tipos de trabajo y metodologías de estos estudios, así como analizar su impacto, se presenta una revisión sistemática de la literatura realizada en la base de datos Web of Science. Se analizó un total de 605 artículos publicados entre 2016 y 2020. Se aplicaron cálculos estadísticos descriptivos e inferenciales no paramétricos (pruebas de Kruskal-Wallis y U de Mann-Whitney) con el objetivo de descubrir diferencias significativas en el número de citas recibidas en cada una de las variables anteriormente mencionadas. Las soluciones para afrontar el fenómeno de la desinformación son las temáticas más frecuentes, pero los trabajos sobre los patrones de propagación del contenido falso presentan mayor impacto. Destaca la alta presencia de estudios de tipo cuantitativo, sobre todo procedentes del ámbito anglosajón. La aproximación metodológica de los trabajos es la única variable que resulta relevante en el impacto de este tipo de estudios.

Palabras clave:

Desinformación; fake news; revisión sistemática; metodología de la investigación; impacto científico.

1. Introduction and review of the literature

The analysis of disinformation and fake news has been a major trend in recent years in scientific research in several fields. Since Donald Trump’s victory in the 2016 election and the Brexit referendum in United Kingdom, the bulk of researchers’ attention has gone to revealing the causes, formal characteristics, production and propagation patterns as well as possible solutions to the phenomenon, seen as one of current democracies’ biggest problems (Elías, 2021). In liberal democratic systems, interest in disinformation “lies in its presumed capacity to raise levels of political polarization, and at the end of the day, damage the principals which underpin coexistence“ (Fernández-Roldán, 2021: 60), this explaining the concern this problem causes among scientists, particularly in the field of communication.

Though digital environments were initially welcomed as offering emancipating and democratic possibilities (Jenkins, 2006; Shirky, 2010) and to drive social movements (Castells, 2012) and the blossoming of so-called “citizen journalism” (Gillmor, 2004), over the last five years there has been a proliferation of research which places new digital technical developments applied to communication as one of the chief causes of the current phenomenon of disinformation. Most of these studies have focused on how to put a stop to the production and diffusion of false content in digital spaces (Margolin, Hannak & Weber, 2018; Zhang et al., 2019), the analysis of disinformation campaigns in virtual settings (Endsley, 2018) and how to detect false rumours on social networks (Zubiaga et al., 2018). These studies also cover the evaluation of the quality of the sources in these environments (Kim, Moravec & Dennis, 2019; Pennycook & Rand, 2019) and even, the appropriateness of censuring content on these platforms (Citron, 2018).

Research into disinformation has also looked kindly on technology, seeing digital tools as allies in the struggle against disinformation. In fact, an increasingly important area of analysis in recent years has been the effectiveness of algorithms and automated solutions for the detection of fake news (Soe, 2018; Ko et al., 2019). Some of these studies, such as that of Del Vicario et al. (2019), focus on the analysis of predictive models which help to better understand the production patterns of fake news. Research into the detection of false content on the networks has been particularly prolific, building on proposals for automatic identification and visualization tools for fake news (Shu, Mahudeswaran & Liu, 2019). These approaches have come from two points of view. Firstly, there are numerous studies which analyse content characteristics in order to train the algorithms in the detection of disinformation via machine learning techniques (Wang et al, 2018), textual analysis (Horne & Adali, 2017), stylometric tools (Potthast et al., 2018) and quantitative methods for an evaluation of the message’s credibility (Fairbanks et al., 2018). Other work proposes deep learning techniques as an approach for the identification of malicious content (Riedel et al., 2017; Popat et al., 2018). Secondly, studies focused on the detection of disinformation from the contextual characteristics of its place of production (Tacchini et al., 2017; Volkova et al., 2017) and circulation (Wu & Liu, 2018).

Also within the area of technology, studies by Fletcher, Schifferes and Thurman (2020) seek to analyse the effectiveness of algorithms in performing automated evaluations of source credibility. Other research in the same field has focused on algorithmic moderation of content (Gorwa, Binns & Katzenbach, 2020), the use of big data to detect disinformation (Torabi-Asr & Taboada, 2019), the automatic identification of bots in digital contexts (Beskow & Carley, 2019) and the applications of the blockchain against false information (Chen et al., 2020).

Apart from the technological implications of this phenomenon, the communication strategies used to fight against disinformation have also generated a considerable number of studies. Specifically, these studies have analysed the effective communication of science (Rose, 2018), counterpropaganda against disinformation campaigns in certain countries (Haigh, Haigh & Kozak, 2018) or how to combat populist discourse (Bosworth, 2019). A large number of the papers on this question focus on the role of the media in disinformation crises, in particular as regards fact-checking. Numerous studies have been caried out in this field about the efficacy of these entities in disproving fake news concerning science (Peter & Koch, 2016), health (Vraga & Bode, 2017), and in electoral (Nyhan et al., 2020) and highly polarized contexts (Zollo et al., 2017). The study of fact-checking has also been done from an analysis of its mechanisms of transparency when explaining the verification process to citizens (Humprecht, 2020), its effects on politicians’ credibility (Agadjanian et al., 2019) or the use of social networks or real-time messaging services (Bernal-Triviño & Clares-Gavilan, 2019). Other research has evaluated the role of collaborative investigative journalism (Carson y Farhall, 2018), the establishment of limits and journalistic standards that can contribute to clearly differentiating between true information and content that is malicious and of doubtful quality (Luengo & García-Marín, 2020), strategies that the media themselves can put into place to improve their credibility and citizens’ trust (Vos & Thomas, 2018; Pickard, 2020) and the adoption of more horizontal and participative communication models for their relationship with their audiences (Meier, Kraus & Michaeler, 2018).

As we can see, disinformation is an object of study that can be analysed from different perspectives and areas of knowledge. The multidimensional character of the phenomenon (Aparici & García-Marín, 2019) requires the adoption of approaches from different fields, as well as that of communication (and in combination with this latter), such as computational sciences, psychology, sociology or education. Citizens, different levels of government, researchers, technology companies and news organizations must, therefore, join forces to mitigate false content which circulates in digital spaces and the traditional media (Dahlgren, 2018).

The need to implement digital and news literacy programmes using critical pedagogy (Higdon, 2020) has considerable presence in research into this question. The contributions of Roozenbeek and Van der Linden (2019) as well as Roozenbeek et al. (2020) stand out here. They studied the effectiveness of inoculation strategies in putting citizens in the place of the creators of disinformation with the object of detecting it. From a more general perspective, other researchers have demonstrated the efficiency of news literacy projects in the fight against malicious content (Amazeen & Bucy, 2019), on social networks such as Twitter (Vraga, Bode & Tully, 2020) and in specific countries, such as United States and India (Guess et al., 2020). Khan & Idris (2019) have shown the correlation between the news education of the population and their capacity to identify fake news, while other studies such as those of Gómez-Suárez (2017) and Buckingham (2019) put the onus on the importance of critical thinking in any media education project against malicious content. Additionally, the outbreak since early 2020 of made-up and misleading news about Covid-19 has led to calls for action to be taken to strengthen citizens’ scientific education (Hughes, 2019; López-Borrull, 2020; Sharon & Baram-Tsabari, 2020).

In recent years, there have also been numerous studies into the regulatory frameworks which some countries have introduced to minimize the impact of fake news, especially in digital media (Baade, 2018; Perl, Howlett & Ramesh, 2018). A smaller number of papers have looked at the role of libraries as content curators (Walter et al., 2020), action designed to regain trust in and the credibility of political (Hunt, Wang & Zhuang, 2020), social (Bjola & Papadakis, 2020) and educational institutions, and the ethical implications of the struggle against disinformation (Bjola, 2018).

This prolific academic output over the past five years has favoured the publication of numerous reviews of the scientific literature focused on the study of disinformation. Some of this work, such as that of Tandoc (2019) and Jankowski (2018), has been performed non-systematically. The -few- systematic reviews are characterised by their wide range of subjects and approaches. Of particular interest are those studies on fake news (Ha, Andreu-Pérez & Ray, 2019), research into fact-checking (Chan et al., 2017; Nieminen & Rapeli, 2019), analysis of disinformation in the social media (Pierri & Ceri, 2019), false product reviews in the framework of electronic commerce (Wu et al., 2020) and, finally, definitions of the different types of disinformation in order to draw up a unified classification of information disorders (Kapantai et al., 2020).

In line with these studies, ours consists of a systematic literature review (SLR) concerning the phenomenon of disinformation, from a general and quantitative perspective. However, as explained in the following section, our study presents clear differences from previous work of this nature. On one hand, it is not focused on one single aspect of the question but covers it from a multi-dimensional and multi-disciplinary viewpoint, integrating various perspectives and areas of study. Moreover, the sample includes a far higher number of papers analysed (n=605) than in the aforementioned reviews. Finally, it integrates a statistical analysis of the impact of this research (measured by the number of references made and normalised by the respective year of publication) from the intersection of several variables: subject matter, fields of study, types of work and country of origin.

2. Research questions & methodology

Our work, eminently quantitative, seeks to answer the following questions:

To answer these questions, a systematic literature review has been performed following the principles proposed by Kitchenham (2004), characterized by transparency and systemization in each of its phases (Codina, 2020). This approach allows us to deeply study, to evaluate and to interpret the scientific work conducted on a specific object of study in a previously established timeframe (Ramírez-Montoya & García-Peñalvo, 2018).

Our methodological process was based on the contributions of Higgins et al. (2019), Brereton et al. (2007) and the Centre for Reviews and Dissemination of York University (2009), which divide studies with systematic reviews into three stages: (1) work planning from a pre-ordained definition of a series of research questions, (2) carrying-out of a documentary search and the creation of a definitive data base following the application of a set of criteria for inclusion and exclusion, and (3) drawing-up of a results report to answer the previously established questions.

To identify the articles to be analysed, we focused our study on the Web of Science (WoS) data base. Just as in previous studies of a similar nature, such as that of López-García et al. (2019), we use this data base exclusively as it is the most relevant scientifically and because it seeks to gather together research published in the journals of greatest impact and visibility internationally. For the first stage of compiling publications, the following criteria of inclusion were adopted:

  1. Studies published in the Social Sciences Citation Index (SSCI) data base of WoS.
  2. Period of publication: 2016-2020. 2016 was chosen as our start point as it saw the emergence of this phenomenon as we now know it (Aparici & García-Marín, 2019), with Donald Trump’s victory in the American presidential election and the Brexit referendum in United Kingdom. The term post-truth was elected word of the year by the Oxford English Dictionary. Following Trump’s victory in late 2016, use of the term fake news “started to become common in the media, and from there, in society in general” (Rodríguez-Fernández, 2021: 48), as shown by the graph of searches for the term on Google between January 2004 and May 2021 (figure 1). The growth in the number of searches became significant from November 2016.
  3. Studies that approach any aspect of the phenomenon of disinformation and from any field of knowledge (communication, psychology, sociology, education, political science, etc.).
  4. Inclusion in the title, abstract or keywords of at least one of the following terms: “disinformation”, “fake news”, “post-truth” or “misinformation”. To this end, advanced searches were made using the Boolean data type operators: “disinformation” OR “fake news” OR “post-truth” OR “misinformation”.
  5. Not only empirical work was included, but also studies of theoretical review and essays published by scientific journals.

The initial compilation of papers took place on December 7, 2020. 2650 publications were amassed (figure 2), which were then analysed by an exhaustive review of the title, abstract and keywords. From this first corpus of papers, a filtering process took place using two criteria of exclusion:

  1. Editorials presenting monographs and special editions.
  2. Papers not related to the object of our study were excluded, although they may have included some of the search terms. The search for the term “misinformation” was particularly problematic, as the concept encompasses papers on how some collectives (people with disability, for example) establish communication, and communication with non-human actors (information transmission between machines and/or digital technological devices), considerations that have nothing to do with the object of our study. Numerous papers containing the term “post-truth” were also discarded as the word is employed in many articles to characterise current times, but, in reality, were not dealing with disinformation.

Following the review phase, the definitive sample was reduced to 605 papers. Coding categories were now identified (see table 1) for each of the seven variables present in our research questions: (1) subject, (2) year of publication, (3) areas of study, (4) types of study / methodology, (5) country of origin of the study, (6) case studies and (7) normalised impact. In the case of countries of origin, the nationality of the institution of the lead researcher was taken as a reference. The data extracted was analysed using descriptive and referential statistics with the SPSS v.26 statistical package.

To measure the studies’ impact, the absolute number of references received was not used, in order to avoid distortions deriving from the year of publication. In its place, the number of references to each article was relativized with regard to the average number of references in each of the publication years (2016, 2017, 2018, 2019 and 2020). This average of annual references was taken from the sample analysed. Therefore, the normalised impact of each article is the result of dividing the number of references by the average number of references in its year of publication. It was decided to normalise the impact using this procedure instead of utilizing the figure for worldwide references in that specific area of knowledge due to the non-availability of such data corresponding to the years 2019 and 2020 (see the updated tables of worldwide references at this link: https://cutt.ly/3Ejykus).

Figure 1. Searches for the term “fake news” in Google between January 2004 and May 2021

Source: Google Trends

Figure 2. Process of systematic review adopted for this paper

Source: created by the author

Table 1. The relationship of each variable to its coding categories

Variable

Coding category

Subject

General

Solutions & strategies

Cognitive & ideological bias

Production of disinformation

Consequences / impact

Propagation of misleading content

Causes of disinformation

Definitions of key concepts

Perceptions of disinformation

Role of journalism

Emotions & disinformation

Deepfakes

Negationism

Cybersecurity

Year

2016

2017

2018

2019

2020

Academic field

Communication (no related area)

Psychology

Politics

Analysis of social networks

Health

Education

Science

Sociology

Philosophy

Economics

Type of study / methodology

Quantitative

Experiment

Qualitative

Mixed (quantitative-qualitative)

Review of theory

Essay

Country

The country of affiliation of the researcher who signed the article was taken into consideration

Case study

Yes / No. Indicate the case.

Normalised impact

Number of times the article was cited as of 07/12/2020, divided by the average number of references in the year it was published (obtained from the sample).

Source: created by the author

To perform the inferential statistical calculations and, that way, check for the existence of significant differences in the variables of the study, we had previously carried out a normality test of the sample in terms of normalised impact. This is necessary to choose the application of parametric and nonparametric tests in the statistical calculations. Kolgomorov-Smirnov tests were performed to determine the absence of normal distributions in the variable relative to the normalised impact of the studies (p < .01). It was decided to use Kruskal-Wallis and Mann-Whitney U nonparametric tests for this, to discover major differences on the impact in terms of the different categories of the variables, and, where necessary, to conduct multiple comparisons between these categories.

3. Results

Given the number of questions our study seeks to answer and with the aim of easing understanding of this section, the results are presented after first introducing the relevant research question.

Q1. What topics were researched in the field of disinformation over the period 2016-2020? Which had most impact (number of references made and normalised by year of publication)? Is the topic under research a determinant variable of the impact?

The most prevalent studies are those that analyse possible solutions and strategies for dealing with the problem of disinformation (n=144; 23.80%) (table 2). These studies focus on a total of 12 subcategories: (1) solutions related to social networks and platforms, (2) communication strategies, (3) the role of the media, (4) policies against disinformation, (5) technological, algorithmic and automated solutions, (6) education against fake news, (7) the role of libraries, (8) the regaining of trust in institutions, (9) the role of art against disinformation, (10) ethical implications in the fight against fake news, (11) fact-checking and evaluation of news sources and (12) meta-analysis of the proposed solutions. The appendix features all the solutions and strategies against disinformation covered by the studies analysed.

Table 2. Average and standard deviation from the normalised impact by subject

Subject

Frequency

Normalised Impact

Average

SD

Solutions & strategies

144 (23.80%)

0.84

0.92

Cognitive bias

119 (19.67%)

0.96

1.41

Production

84 (13.88%)

0.69

0.78

Consequences. impact

57 (9.42%)

1.17

1.91

Propagation

47 (7.77%)

2.29

5.38

Causes

46 (7.60%)

0.90

0.99

General

34 (5.62%)

0.95

1.25

Definitions

22 (3.63%)

1.33

2.24

Perceptions (of disinformation)

22 (3.63%)

0.73

0.73

Role of journalism

11 (1.81%)

0.38

0.19

Emotions

7 (1.16%)

0.45

0.31

Negationism

5 (0.82%)

0.65

0.50

Deepfakes

5 (0.82%)

0.91

0.48

Cybersecurity

2 (0.33%)

0.43

0.41

p value (χ2)

< .00*

p value (Kruskal-Wallis)

.19**

* Very significant differences are observed when p < .01.

** Significant differences are observed when p < .05.

Source: created by the author

After solutions, the most common studies are those related to cognitive and ideological bias (n=119; 19.67%). The analyses of the production mechanisms of disinformation come to 13.88% of the studies (n=84), while those relative to the repercussions or consequences of the question make up 9.42% (n=57). Chi-squared distribution tests determine highly significant differences in the frequency of this variable [χ2(13, N = 605) = 595.41, p < .00].

The papers with the greatest impact are those which address propagation patterns of false content (M=2.29; SD=5.38). In the next place are the studies which analyse the definitions of the key concepts related to this question (fake news, disinformation, post-truth, information disorders, etc.) from a theorical perspective, with an average impact of 1.33 and a standard deviation of 2.24. The studies on the consequences or repercussions of disinformation form the third most referenced topic (M=1.17; SD=1.91). The Kruskal-Wallis tests observe that the topic of a study does not constitute an important variable of its impact [H(13) = 17.12, p = .19].

Q2. What is the frequency of these studies by year of publication? Which topics were published most each year?

Studies on disinformation published in WoS have grown in number during the period 2016-2020. 37 studies published in 2016 have been compiled (6.11%), 84 in 2017 (13.88%), 151 in 2018 (24.95%), 198 in 2019 (32.72%) and 135 in 2020 (22.31%). Between 2016 and 2020, research published on this object of study multiplied five-fold.

Table 3 covers the topics in our sample by year. Two opposing tendencies stand out. On one hand, the percentage of studies on solutions and strategies against disinformation grew notably over the period. This topic represented 8.10% of the studies published in 2016, while by 2020 this aspect was addressed in 28.89% of that year’s research. In the opposite direction, the percentage of papers on cognitive bias, which made up 35.13% of research in 2016 had fallen considerably to only 17.03% by 2020. We infer from this that before the proliferation of studies on disinformation that started in 2017, there was already notable interest in psychological bias and news credibility.

Table 3. Number of papers published by topic over the period 2016-2020. The relative frequencies refer to
the percentage of papers published on the topic in that year

Topic

2016

2017

2018

2019

2020

Causes

2 (5.40%)

9 (10.71%)

14 (9.27%)

17 (8.58%)

4 (2.96%)

Consequences, impact

2 (5.40%)

6 (7.14%)

15 (9.93%)

16 (8.08%)

18 (13.33%)

Definitions

0 (0.00%)

5 (5.95%)

10 (6.62%)

5 (2.52%)

2 (1.48%)

Propagation

4 (10.81%)

7 (8.33%)

9 (5.96%)

16 (8.08%)

11 (8.14%)

Cognitive bias

13 (35.13%)

24 (28.57%)

23 (15.23%)

36 (18.18%)

23 (17.03%)

Solutions & strategies

3 (8.10%)

13 (15.47%)

36 (23.84%)

53 (26.76%)

39 (28.89%)

General

0 (0.00%)

5 (5.95%)

9 (5.96%)

14 (7.07%)

6 (4.44%)

Production

6 (16.21%)

7 (8.33%)

23 (15.23%)

24 (12.72%)

24 (17.78%)

Perceptions

3 (8.10%)

4 (4.76%)

6 (3.97%)

7 (3.53%)

2 (1.48%)

Negationism

0 (0.00%)

1 (1.19%)

1 (0.66%)

3 (1.51%)

0 (0.00%)

Role of journalism

1 (2.70%)

1 (1.19%)

3 (1.98%)

3 (1.51%)

3 (2.22%)

Emotions

3 (8.10%)

2 (2.38%)

1 (0.66%)

0 (0.00%)

1 (0.74%)

Deepfakes

0 (0.00%)

0 (0.00%)

0 (0.00%)

3 (1.51%)

2 (1.48%)

Cybersecurity

0 (0.00%)

0 (0.00%)

1 (0.66%)

1 (0.50%)

0 (0.00%)

Total

37

84

151

198

135

Source: created by the author

Q3. Which fields of knowledge conduct studies on disinformation? Which have most impact? Is the field of knowledge a determining variable on the impact of the work?

Very clear differences can be seen in the number of studies by area of knowledge [χ2(9, N = 605) = 310.62, p < .00]. The field of psychology (especially the studies on cognitive biases) is the area which has provided the highest number of papers (n=131; 21.65%) (table 4). Research from political science (n=119; 19.67%) and studies about social networks (n=112; 18.51%) have also been prominent. Following them, come papers on health (n=71; 11.73%), doubtless as a consequence of the proliferation of research analysing disinformation deriving from the Covid-19 pandemic, a matter greatly studied throughout 2020.

It is precisely these papers from the health field which have the greatest impact (M=1.17; SD=1.39). Studies about disinformation on social networks and digital platforms also had a major impact (M=1.13; SD=3.33) as well as those relative to politics (M=1.10; SD=1.90). Papers found exclusively in the area of communication (therefore, without another related area) reached an average impact of 1.05 (SD=1.66). The nonparametric tests conducted determine that the field of study is not a relevant factor in the impact of these papers [H(9) = 13.60, p = .13].

Table 4. Average and standard deviation of the normalised impact by fields of knowledge

Areas

Frequencies

Normalised Impact

Average

SD

Psychology

131 (21.65%)

0.92

1.36

Politics

119 (19.67%)

1.10

1.90

Networks

112 (18.51%)

1.13

3.33

Health

71 (11.73%)

1.17

1.39

Communication (no related area)

56 (9.25%)

1.05

1.66

Education

38 (6.28%)

0.70

0.59

Science

32 (5.29%)

1.04

1.12

Sociology

26 (4.29%)

0.52

0.41

Philosophy

12 (1.98%)

0.42

0.30

Economics

8 (1.32%)

0.80

0.44

p value (χ2)

< .00*

p value (Kruskal-Wallis)

.13**

*Highly significant differences are observed when p < .01.

** Significant differences are observed when p < .05.

Source: created by the author

Q4. What types of studies are they and what methodology is utilised? Which types have greatest impact? Is the type of work a determinant variable on the impact?

The essay is the commonest type of paper (n=168; 27.77%), followed by quantitative (n=159; 26.28%) and experimental studies (n=135; 22.31%) (table 5). Note that experiments are also quantitative, and therefore quantitative methodologies appear in almost half of the sample analysed (48.59%). The publication of qualitative studies is far less frequent (n=55; 9.09%) as is that of those conducted with mixed methodologies (articulation of qualitative and quantitative methods) (n=48; 7.93%). As occurred in the case of topics and areas of knowledge, there are highly significant differences in the frequencies of this variable [χ2(5, N = 605) = 175.09, p < .00].

Non-experimental quantitative studies receive a greater number of relative references and, therefore, achieve greater impact (M=1,42; SD=3.18). Following these come review studies of the literature (M=1,24; SD=1.90), which have considerable impact despite being the most infrequently published type of paper. The studies with least impact are the qualitative ones (M=0,75; SD=0.94) and essays (M=0,76; SD=0.86). It can be seen that the type of study or methodology employed constitutes a determinant factor on its impact [H(5) = 15.78, p < .00]. Table 6 shows the specific types of study where the differences in the impact are significant (p < .05) or highly significant (p < .01).

Table 5. Average and standard deviation from the normalised impact by type of study

Type of study / methodology

Frequency

Normalised impact

Average

SD

Essay

168 (27.77%)

0.76

0,86

Quantitative

159 (26.28%)

1.42

3,18

Experimental

135 (22.31%)

0.84

1,06

Qualitative

55 (9.09%)

0.75

0,94

Mixed (quantitative-qualitative)

48 (7.93%)

0.93

1,83

Review

40 (6.61%)

1.24

1,90

p value (χ2)

< .00*

p value (Kruskal-Wallis)

< .00*

*Highly significant differences are observed when p < .01.

Source: created by the author

Table 6. Multiple comparatives by the Kruskal-Wallis test on the normalised impact for each of the types of research /
methodologies employed

Types / methods compared

p value (Kruskal-Wallis)

Mixed- Quantitative

.031*

Qualitative- Quantitative

.008**

Experiment- Quantitative

.007**

Essay- Quantitative

.002**

*Significant differences are observed when p < .05.

**Highly significant differences are observed when p < .01.

Source: created by the author

Q5. In which countries are most studies carried out? Which accumulate most impact? Is a study’s country of origin a determining variable on its impact?

As might be expected, papers from the English-speaking world make up the bulk of this area, with the vast majority of research coming from United States (US). The four countries with the largest number of studies are English-speaking: US (n=272; 44.95%), United Kingdom (UK) (n=74; 12.23%), Canada (n=34; 5.61%) and Australia (27; 4.46%). Following them we find Germany (n=27; 4.46%) and Spain, with a total of 25 papers (4.13%). Table 7 shows the data from the countries which registered over 10 studies published.

Among the nations with more than 10 studies published, the country with the greatest impact is Italy, with an average of 1.51. Nonetheless, their high value for standard deviation on this variable (SD= 2.56), warns us of the existence of a highly unequal distribution in the number of references, thus Italy publishes some research of great impact but others of much less relevance. After Italy come US (M=1.23; SD=2.56), Canada (M=1.04; SD=1.23), UK (M=0.80; SD=0.99) and Australia (M=0.80; SD=0.64). They are followed by Germany (M=0.75; SD=0,66) and Holland (M=0.73; SD=0.90).

There are notable differences in the number of papers published by country [χ2(47, N = 605) = 5937.99, p < .00]. On the other hand, the country of origin of the studies is not a determining variable on its impact [H(47) = 54.53, p = .21].

Table 7. Average and standard deviation of normalised impact by country

Country

Frequency

Normalised Impact

Average

SD

United States

272 (44.95%)

1.23

2.56

United Kingdom

74 (12.23%)

0.80

0.99

Canada

34 (5.61%)

1.04

1.23

Australia

27 (4.46%)

0.80

0.64

Germany

27 (4.46%)

0.75

0.66

Spain

25 (4.13%)

0.50

0.58

Italy

18 (2.97%)

1.51

2.56

China

12 (1.98%)

0.44

0.22

Holland

11 (1.81%)

0.73

0.90

p value (χ2)

< .00*

p value (Kruskal-Wallis)

.21**

*Highly significant differences are observed when p < .01.

** Significant differences are observed when p < .05.

Source: created by the author

Q6. Which are the most frequent case studies? Is the carrying out of the case studies a determinant variable of their impact?

A total of 120 case studies were found in our sample, representing 19.83% of the total. Disinformation linked to Covid-19 is the most analysed case (n=26; 21.66%), followed by the 2016 American election (n=15; 12.505), Donald Trump’s mandate in the White House (10.83%) and Russian interference in American politics (n=8; 6.67%). A geographical bias is apparent in the most analysed cases, given that all of them (except the Covid-19 pandemic) are related to political processes linked to US, the leading country in the scientific production in the field of disinformation. The complete data on the frequencies of the case studies analysed is available at the following link: https://cutt.ly/Bb9Ms5d

The case studies have greater impact (M=1.35; SD= 2.01) than the papers that do not include studies of this type (M=0.91; SD=1.91). The nonparametric tests (in this case, Mann-Whitney U as we are only comparing two groups) detect highly significant differences in the normalized impact of both types of study (z = -2.895, p = .004).

Q7. Which types of work and research methodologies are used in the different countries?

Considerable differences can be seen in the types of research published in different countries (table 8). In this section we seek to highlight some of the more noteworthy questions. Firstly, the high percentage of quantitative studies carried out in US (57.35% of research from that nation). 29.41% of research in this country are experimental studies. In general, we can see a greater tendency in the English-speaking countries towards quantitative and experimental research, being 37.82% of studies in UK, 35.29% in Canada and 40.74% in Australia.

It is also worth pointing out the extremely high percentage of non-experimental quantitative research performed in Italy (72.22%). The trend in that country towards the publication of quantitative studies may explain the high average impact of Italian papers; as quantitative studies are those which receive the most references. Over half the studies published in Germany (51.85%) are essays, while Spain is the country with the highest percentage of mixed approach studies (36.00%). In comparative terms, Spain publishes far fewer quantitative studies (only 16.00%, 8.00% of papers are experimental and another 8.00% are non-experimental and quantitative). It is noteworthy that the types of research with greatest impact (quantitative and theoretical reviews) are the least common in Spain.

Table 8. Types of paper by country. Relative frequencies refer to the type of paper published in each country

Country

Quant.

Qual.

Exper.

Mix.

Rev.

Essay

United States

n=272

76 (27.94%)

15 (5.51%)

80 (29.41%)

11 (4.04%)

14 (5.88%)

74 (27.20%)

United Kingdom

n=74

14 (18.91%)

10 (13.51%)

14 (18.91%)

6 (8.11%)

2 (2.70%)

28 (37.83%)

Canada

n=34

9 (26.47%)

1 (2.94%)

3 (8.82%)

3 (8.82%)

2 (5.88%)

16 (47.05%)

Australia

n=27

7 (25.92%)

5 (18.51%)

4 (14.81%)

2 (7.40%)

1 (3.70%)

8 (29.63%)

Germany

n=27

2 (7.40%)

3 (11.11%)

6 (22.22%)

1 (3.70%)

1 (3.70%)

14 (51.85%)

Spain

n=25

2 (8.00%)

5 (20.00%)

2 (8.00%)

9 (36.00%)

1 (4.00%)

6 (24.00%)

Italy

n=18

13 (72.22%)

1 (5.55%)

1 (5.55%)

0 (0.00%)

2 (11.11%)

1 (5.55%)

China

n=12

7 (58.33%)

0 (0.00%)

1 (8.33%)

1 (8.33%)

1 (8.33%)

2 (16.66%)

Holland

n=11

1 (9.09%)

2 (18.18%)

4 (36.36%)

1 (9.09%)

2 (18.18%)

1 (9.09%)

Source: created by the author

Q8. What types of work / methodologies are used depending on the topic under study?

Finally, our results indicate that studies about solutions to the question of disinformation generally take on an essay form (33.33% of publications on the topic) and come via experimental methods (25.00%) (table 9). The causes of and the consequences of disinformation and the definitions of key concepts are generally analysed in essays (47.82%; 35.08% and 81.81% respectively). The studies on bias are conducted, fundamentally, with experimental designs (65.54%), while propagation patterns of false content are considered with non-experimental quantitative research (65.96%). Also notable is the percentage of this type of quantitative methodology utilised to analyse the production of disinformation (40.47%).

Table 9. Topics covered by research type. Relative frequencies refer to the topic of each type of research

Topic

Quant.

Qual.

Exper.

Mix.

Rev.

Essay

Causes

n=46

13 (28.26%)

4 (8.69%)

1 (2.17%)

2 (4.34%)

4 (8.69%)

22 (47.82%)

Consequences. impact

n=57

16 (28.07%)

8 (14.03%)

7 (12.28%)

3 (5.26%)

3 (5.26%)

20 (35.08%)

Definitions

n=22

1 (4.54%)

0 (0.00%)

0 (0.00%)

1 (4.54%)

2 (9.09%)

18 (81.81%)

Propagation

n=47

31 (65.95%)

3 (6.38%)

6 (12.76%)

3 (6.38%)

2 (4.25%)

2 (4.25%)

Cognitive bias

n=119

25 (21.01%)

0 (0.00%)

78 (65.54%)

3 (2.52%)

6 (5.04%)

7 (5.88%)

Solutions & strategies

n=144

25 (17.36%)

15 (10.415%)

36 (25.00%)

7 (4.86%)

13 (9.02%)

48 (33.33%)

General

n=34

1 (2.94%)

1 (2.94%)

0 (0.00%)

0 (0.00%)

6 (17.64%)

26 (76.47%)

Production

n=84

34 (40.47%)

16 (19.04%)

2 (2.38%)

21 (25.00%)

2 (2.38%)

9 (10.71%)

Perceptions

n=22

9 (40.90%)

2 (9.09%)

4 (18.18%)

5 (22.72%)

1 (4.54%)

1 (4.54%)

Negationism

n=5

1 (20.00%)

0 (0.00%)

0 (0.00%)

0 (0.00%)

0 (0.00%)

4 (80.00%)

Role of journalism

n=11

1 (9.09%)

6 (54.54%)

0 (0.00%)

1 (9.09%)

0 (0.00%)

3 (27.27%)

Emotions

n=7

1 (14.28%)

0 (0.00%)

1 (14.28%)

1 (14.28%)

1 (14.28%)

3 (42.85%)

Deepfakes

n=5

1 (20.00%)

0 (0.00%)

0 (0.00%)

1 (20.00%)

0 (0.00%)

3 (60.00%)

Cybersecurity

n=2

0 (0.00%)

0 (0.00%)

0 (0.00%)

0 (0.00%)

0 (0.00%)

2 (100%)

Source: created by the author

4. Discussion and conclusions

This study establishes a complete cartography of the subject matter, types of research, areas of knowledge, countries of origin and the impact of the scientific literature on disinformation between 2016 and 2020, so far, the period of greatest attention paid by researchers to this phenomenon. The commonest topics are (1) those that look for solutions to the question and (2) experimental studies into psychological and ideological biases in the reception processes of false content. As for solutions, the studies show a wide variety of focuses over the period observed (a total of 12 different aspects have been found, from which this challenge can be met). The number of studies analysing ways of facing fake news have increased remarkably since 2016. These results show the concern generated by the problem, as well as the emergence of the search for mechanisms to contain it. On the other hand, the studies into cognitive bias, highly prevalent at first, have decreased in number since 2016. This decrease is not an obstacle to placing the field of psychology as one of the areas that generates most papers, confirming previous studies such as that of Ha, Andreu-Pérez and Ray (2019). This research from the field of psychology offers extremely valuable knowledge for analysing the differences in the perception of false content and true information. The bulk of these papers analyse how psychological and ideological biases are activated when citizens tend to believe all that which fits with their worldview (Elías, 2021). These biases also act on the subjects when believing false content, even after having been refuted by fact-checkers (Chan et al., 2017). Processes like backfire effect of fact-checking and motivated reasoning (creating alternative explanations to make reality compatible with ideological bias) have been used to explain resistance to verified information (Cook, Lewandowsky & Ecker, 2017; Swire et al., 2017).

However, none of these aspects -solutions and cognitive bias- are amongst the most referenced. The topic of greatest impact is that focused on the propagation patterns of false content. In this area, the study by Vosoughi, Roy and Aral (2018) is the one that registers the highest number of relative references. The definitions of the key concepts linked to the phenomenon of disinformation is the second category with greatest impact. These definitions are usually dealt with in essays, the most published type of paper. From this last fact it can be inferred that disinformation –in its current form – is a new object of study which still requires the establishment of clear conceptual limits agreed by a scientific community which over the last five years has held incessant discussions about the theoretical underpinning of this phenomenon.

After the field of psychology, political science and the analysis of the social networks are at the forefront in terms of number of papers published. The utilization of misleading strategies in electoral processes and, in general, for political ends, and the importance of digital platforms in these strategies explain these results. The considerable number of papers on health is justified by the avalanche of studies on disinformation which the Covid-19 health crisis has generated since March 2020. The diversity of fields from which the phenomenon has been approached shows its multi-dimensional nature, that has necessitated the collaboration of experts from different disciplines.

Another important aspect is the large number of quantitative papers, especially from the English-speaking world. It would however be preferable to see a greater amount of research using mixed methodology (very little incidence in the studies analysed), whose qualitative perspective could provide greater depth and detail to the statistical results obtained through quantitative methods (Jankowski, 2018). Neither are reviews of the literature particularly common, despite having (after non-experimental quantitative studies) the greatest impact.

From a geographical point of view, our study confirms the leadership of the United States in the study of this question. Previous research had already pointed to the hegemony of US in the scientific and academic approach to false content (Ha, Andreu-Pérez & Ray (2019). In reality, the leadership of the North American country can be extrapolated to the whole English-speaking context (UK, Canada and Australia). Likewise, US provides most of the case studies, focused on the political processes of the American agenda, such as the 2016 election, Donald Trump’s mandate and Russian interference in their politics. This aspect coincides with the results obtained by other studies, such as that of Nieminen and Rapeli (2019).

Calculations made using inferential statistics determine that the type of research (methodological approach) is the only important variable effecting impact. Neither the topic being studied, nor the area of knowledge from which it is approached, nor the country in which it is studied are determining factors.

To finish, it is useful to identify some of the aspects related to disinformation that have not yet received enough attention, and which, therefore, may become lines of research to be explored in future work on this phenomenon. Firstly, there is a noteworthy lack of studies on the legal aspects related to the tagging and elimination of fake news and misleading sources, be they on digital platforms or in other spaces. It is also puzzling that, as opposed to the predominance of research from a political standpoint, the other great motivation behind the generation of false content (economic interest) does not have greater presence in this type of research. It would be interesting to see a strengthening of the approach to such important aspects as black PR –fictitious comments and reviews of products and services available online (Rodríguez-Fernández, 2021)– and, in general, the utilization of disinformation with reputational ends and commercial interests. Neither has there been great prevalence of the study of disinformation in audio formats. The growing importance of audio content in the media environment would make it especially interesting to see an analysis of false content circulating through voice notes on instant messaging services like WhatsApp or Telegram (García-Marín, 2021) or on media such as podcasts, which are increasingly popular. Additionally, it is foreseeable that research into other areas of new technological development, such as deepfakes, or the use of the blockchain as a weapon against disinformation (Tapscott & Tapscott 2016; Dickson, 2017) may acquire heightened importance in coming years.

5. Acknowledgements

This article has been translated into English by Brian O’Halloran.

6. Bibliographic References

Agadjanian, A., Bakhru, N., Chi, V., Greenberg, D., Hollander, B., Hurt, A., & Woodruff, A. (2019). Counting the Pinocchios: The effect of summary fact-checking data on perceived accuracy and favorability of politicians. Research & Politics, 6(3). https://doi.org/10.1177/2053168019870351

Amazeen, M. A., & Bucy, E. P. (2019). Conferring resistance to digital disinformation: The inoculating influence of procedural news knowledge. Journal of Broadcasting & Electronic Media, 63(3), 415-432. https://doi.org/10.1080/08838151.2019.1653101

Aparici, R., & Marín, D. G. (2019). La posverdad: una cartografía de los medios, las redes y la política. Editorial Gedisa.

Baade, B. (2018). Fake News and International Law. European Journal of International Law, 29(4), 1357-1376. https://doi.org/10.1093/ejil/chy071

Bernal-Triviño, A., & Clares-Gavilán, J. (2019). Using mobile devices and social networks as checking channels of fake news. The case of Maldita. es. Profesional de la información, 28(3), e280312. https://doi.org/10.3145/epi.2019.may.12

Beskow, D. M., & Carley, K. M. (2019). Its all in a name: detecting and labeling bots by their name. Computational and Mathematical Organization Theory, 25(1), 24-35. https://doi.org/10.1007/s10588-018-09290-1

Bjola, C. (2018). The ethics of countering digital propaganda. Ethics & international affairs, 32(3), 305-315. https://doi.org/10.1017/S0892679418000436

Bjola, C., & Papadakis, K. (2020). Digital propaganda, counterpublics and the disruption of the public sphere: the Finnish approach to building digital resilience. Cambridge Review of International Affairs, 33(5), 638-666. https://doi.org/10.1080/09557571.2019.1704221

Bosworth, K. (2019). The people know best: Situating the counterexpertise of populist pipeline opposition movements. Annals of the American Association of Geographers, 109(2), 581-592. https://doi.org/10.1080/24694452.2018.1494538

Brereton, P., Kitchenham, B.A., Budgen, D., Turner, M., & Khalil, M. (2007). Lessons from applying the systematic literature review process within the software engineering domain. Journal of Systems and Software, 80(4), 571-583. https://doi.org/10.1016/j.jss.2006.07.009

Buckingham, D. (2019). Teaching media in a ‘post-truth’age: fake news, media bias and the challenge for media/digital literacy education/La enseñanza mediática en la era de la posverdad: fake news, sesgo mediático y el reto para la educación en materia de alfabetización mediática y digital. Cultura y Educación, 31(2), 213-231. https://doi.org/10.1080/11356405.2019.1603814

Carson, A., & Farhall, K. (2018). Understanding collaborative investigative journalism in a “post-truth” age. Journalism Studies, 19(13), 1899-1911. https://doi.org/10.1080/1461670X.2018.1494515

Castells, M. (2012). Redes de indignación y esperanza. Los movimientos sociales en la era de Internet. Alianza Editorial.

Centre for Reviews and Dissemination. (2009). Systematic Reviews. CRD’s guidance for undertaking reviews in health care. University of York.

Chan, M. P. S., Jones, C. R., Hall Jamieson, K., & Albarracín, D. (2017). Debunking: A meta-analysis of the psychological efficacy of messages countering misinformation. Psychological science, 28(11), 1531-1546. https://doi.org/10.1177/0956797617714579

Chen, Q., Srivastava, G., Parizi, R. M., Aloqaily, M., & Al Ridhawi, I. (2020). An incentive-aware blockchain-based solution for internet of fake media things. Information Processing & Management, 57(6). https://doi.org/10.1016/j.ipm.2020.102370

Citron, D.K. (2018). Extremist Speech, Compelled Conformity, and Censorship Creep. Notre Dame Law Review, 93, 1035-1072. https://cutt.ly/wbVP5lP

Codina, L. (2020). Cómo hacer revisiones bibliográficas tradicionales o sistemáticas utilizando bases de datos académicas: estudio de una serie de casos. Revista ORL, 11(2), 139-153. https://dx.doi.org/10.14201/orl.22977

Cook, J., Lewandowsky, S., & Ecker, U. K. (2017). Neutralizing misinformation through inoculation: Exposing misleading argumentation techniques reduces their influence. PLoS One, 12, e0175799. https://doi.org/10.1371/journal.pone.0175799

Del Vicario, M., Quattrociocchi, W., Scala, A. & Zollo, F. (2019). Polarization and Fake News: Early Warning of Potential Misinformation Targets. ACM Transactions on the Web, 13(2). https://doi.org/10.1145/3316809

Dahlgren, P. (2018). Media, knowledge and trust: The deepening epistemic crisis of democracy. Javnost-The Public, 25(1-2), 20-27. https://doi.org/10.1080/13183222.2018.1418819

Dickson B. (2017). How blockchain helps fight fake news and filter bubbles. The Next Web. Recuperado de: https://bit.ly/2OLsNvq

Elías, C. (2021). El periodismo como herramienta contra las fake news. En Elías, C. & Teira, D. (Eds.), Manual de periodismo y verificación de noticias en la era de las fake (pp. 19-57). UNED.

Endsley, M.R. (2018). Combating Information Attacks in the Age of the Internet: New Challenges for Cognitive Engineering. Human Factors, 60(8), 1081-1094. https://doi.org/10.1177/0018720818807357

Fairbanks, J., Fitch, N., Knauf, N. & Briscoe, E. (2018). Credibility assessment in the news: Do we need to read? Proceedings of the MIS2 Workshop held in conjuction with 11th International Conference on Web Search and Data Mining. https://cutt.ly/hbVHlsx

Fernández-Roldán, A. (2021). El papel de las plataformas digitales en la difusión de contraconocimiento. En Elías, C. & Teira, D. (Eds.), Manual de periodismo y verificación de noticias en la era de las fake (pp. 59-84). UNED.

Fletcher, R., Schifferes, S., & Thurman, N. (2020). Building the ‘Truthmeter’: Training algorithms to help journalists assess the credibility of social media sources. Convergence, 26(1), 19-34. https://doi.org/10.1177/1354856517714955

García-Marín, D. (2021). El whatsapp de Odiseo. Potencial desinformación y estrategias retóricas del audio fake. En Elías, C. & Teira, D. (Eds.), Manual de periodismo y verificación de noticias en la era de las fake (pp. 99-132). UNED.

Gillmor, D. (2004). We the media: Grassroots Journalism by the people, for the people. O’Reilly Media.

Gómez-Suárez, A. (2017). Peace process pedagogy: lessons from the no-vote victory in the Colombian peace referendum. Comparative Education, 53(3), 462-482. https://doi.org/10.1080/03050068.2017.1334425

Gorwa, R., Binns, R., & Katzenbach, C. (2020). Algorithmic content moderation: Technical and political challenges in the automation of platform governance. Big Data & Society, 7(1). https://doi.org/10.1177/2053951719897945

Guess, A. M., Lerner, M., Lyons, B., Montgomery, J. M., Nyhan, B., Reifler, J., & Sircar, N. (2020). A digital media literacy intervention increases discernment between mainstream and false news in the United States and India. Proceedings of the National Academy of Sciences, 117(27), 15536-15545. https://doi.org/10.1073/pnas.1920498117

Ha, L., Andreu-Pérez, L., & Ray, R. (2019). Mapping recent development in scholarship on fake news and misinformation, 2008 to 2017: Disciplinary contribution, topics, and impact. American behavioral scientist, 65(2), 290-315. https://doi.org/10.1177/0002764219869402

Haigh, M., Haigh, T., & Kozak, N. I. (2018). Stopping fake news: The work practices of peer-to-peer counter propaganda. Journalism studies, 19(14). https://doi.org/10.1080/1461670X.2017.1316681

Higdon, N. (2020). The Anatomy of Fake News: A Critical News Literacy Education. University of California Press.

Higgins, J. P., Thomas, J., Chandler, J., Cumpston, M., Li, T., Page, M. J., & Welch, V. A. (2019). Cochrane handbook for systematic reviews of interventions. John Wiley & Sons.

Horne, B.D. & Adali. S. (2017). This just in: fake news packs a lot in title, uses simpler, repetitive content in text body, more similar to satire than real news. Proceedings of the International AAAI Conference on Web and Social Media, 11(1). https://cutt.ly/PbVF9LU

Hughes, G. (2019). Developing student research capability for a ‘post-truth’world: Three challenges for integrating research across taught programmes. Teaching in Higher Education, 24(3), 394-411. https://doi.org/10.1080/13562517.2018.1541173

Humprecht, E. (2020). How do they debunk “fake news”? A cross-national comparison of transparency in fact checks. Digital Journalism, 8(3), 310-327. https://doi.org/10.1080/21670811.2019.1691031

Hunt, K., Wang, B., & Zhuang, J. (2020). Misinformation debunking and cross-platform information sharing through Twitter during Hurricanes Harvey and Irma: a case study on shelters and ID checks. Natural Hazards, 103, 861-883. https://doi.org/10.1007/s11069-020-04016-6

Jankowski, N. W. (2018). Researching fake news: A selective examination of empirical studies. Javnost-The Public, 25(1-2), 248-255. https://doi.org/10.1080/13183222.2018.1418964

Jenkins, H. (2006). Convergence Culture. La cultura de la convergencia en los medios de comunicación. Paidós.

Kapantai, E., Christopoulou, A., Berberidis, C., & Peristeras, V. (2020). A systematic literature review on disinformation: Toward a unified taxonomical framework. New media & society, 23(5), 1301-1326. https://doi.org/10.1177/1461444820959296

Khan, M. L., & Idris, I. K. (2019). Recognise misinformation and verify before sharing: a reasoned action and information literacy perspective. Behaviour & Information Technology, 38(12), 1194-1212. https://doi.org/10.1080/0144929X.2019.1578828

Kim, A., Moravec, L. & Dennis, A.R. (2019). Combating Fake News on Social Media with Source Ratings: The Effects of User and Expert Reputation Ratings. Journal of Management Information Systems, 36(3), 931-968.https://doi.org/10.1080/07421222.2019.1628921

Kitchenham, B. (2004). Procedures for performing systematic reviews. Keele, UK, Keele University, 33, 1-26. https://cutt.ly/TbV4nRW

Ko, H., Hong, J.Y., Kim, S., Mesicek, L. & Na, I.S. (2019). Human-machine interaction: A case study on fake news detection using a backtracking based on a cognitive system. Cognitive Systems Research, 55, 77-81. https://doi.org/10.1016/j.cogsys.2018.12.018

López-Borrull, A. (2020). Fake news e infodemia científica durante la Covid-19,¿ dos caras de la misma crisis informacional? Anuario ThinkEPI, 14(1), e14e07. https://doi.org/10.3145/thinkepi.2020.e14e07

López-García, X., Silva-Rodríguez, A., Vizoso-García, Á. A., Westlund, O., & Canavilhas, J. (2019). Periodismo móvil: Revisión sistemática de la producción científica. Comunicar, 27(59), 9-18. https://doi.org/10.3916/C59-2019-01

Luengo, M., & García-Marín, D. (2020). The performance of truth: politicians, fact-checking journalism, and the struggle to tackle COVID-19 misinformation. American Journal of Cultural Sociology, 8(3), 405-427. https://doi.org/10.1057/s41290-020-00115-w

Margolin, D.B., Hannak, A. & Weber, I. (2018). Political Fact-Checking on Twitter: When Do Corrections Have an Effect? Political Communication, 35(2), 196-219. https://doi.org/10.1080/10584609.2017.1334018

Meier, K., Kraus, D., & Michaeler, E. (2018). Audience engagement in a post-truth age: What it means and how to learn the activities connected with it. Digital Journalism, 6(8), 1052-1063. https://doi.org/10.1080/21670811.2018.1498295

Nieminen, S., & Rapeli, L. (2019). Fighting misperceptions and doubting journalists’ objectivity: A review of fact-checking literature. Political Studies Review, 17(3), 296-309. https://doi.org/10.1177/1478929918786852

Nyhan, B., Porter, E., Reifler, J., & Wood, T. J. (2020). Taking fact-checks literally but not seriously? The effects of journalistic fact-checking on factual beliefs and candidate favorability. Political Behavior, 42(3), 939-960. https://doi.org/10.1007/s11109-019-09528-x

Pennycook, G. & Rand, D.G. (2019). Fighting misinformation on social media using crowdsourced judgments of news source quality. Proceedings of the National Academy of Sciences, 116(7), 2521-2526. https://doi.org/10.1073/pnas.1806781116

Perl, A., Howlett, M., & Ramesh, M. (2018). Policy-making and truthiness: Can existing policy models cope with politicized evidence and willful ignorance in a “post-fact” world?. Policy Sciences, 51(4), 581-600. https://doi.org/10.1007/s11077-018-9334-4

Peter, C., & Koch, T. (2016). When debunking scientific myths fails (and when it does not) The backfire effect in the context of journalistic coverage and immediate judgments as prevention strategy. Science Communication, 38(1), 3-25. https://doi.org/10.1177/1075547015613523

Pickard, V. (2020). Restructuring democratic infrastructures: A policy approach to the journalism crisis. Digital Journalism, 8(6), 704-719. https://doi.org/10.1080/21670811.2020.1733433

Pierri, F. & Ceri, S. (2019). False News On Social Media: A Data-Driven Survey. ACM SIGMOD Record, 48(2). https://doi.org/10.1145/3377330.3377334

Popat, K., Mukherjee, S., Yates, A., & Weikum, G. (2018). Declare: Debunking fake news and false claims using evidence-aware deep learning. Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, 22–32. https://cutt.ly/DbVJnMQ

Potthast, M., Kiesel, J., Reinartz, J., Bevendor, J. & Stein, B. (2018). A stylometric inquiry into hyperpartisan and fake news. Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics, 1, 231–240. https://cutt.ly/ibVGDy6

Ramírez-Montoya, M.S., & García-Peñalvo, F.J. (2018). Co-creación e innovación abierta: Revisión sistemática de literatura. Comunicar, 26(54), 9-18. https://doi.org/10.3916/C54-2018-01

Riedel, B., Augenstein, I., Spithourakis, G. P., & Riedel, S. (2017). A simple but tough-to-beat baseline for the Fake News Challenge stance detection task. arXiv. https://cutt.ly/PbVJyUi

Rodríguez-Fernández, L. (2021). Propaganda digital: Comunicación en tiempos de desinformación. Editorial UOC.

Rose, D. C. (2018). Avoiding a post-truth world: Embracing post-normal conservation. Conservation and Society, 16(4), 518-524. https://cutt.ly/2bVXMvZ

Roozenbeek, J., Maertens, R., McClanahan, W., & van der Linden, S. (2020). Disentangling item and testing effects in inoculation research on online misinformation: Solomon revisited. Educational and Psychological Measurement, 81(2), 340-362. https://doi.org/10.1177/0013164420940378

Roozenbeek, J., & Van Der Linden, S. (2019). The fake news game: actively inoculating against the risk of misinformation. Journal of Risk Research, 22(5), 570-580. https://doi.org/10.1080/13669877.2018.1443491

Sharon, A. J., & Baram‐Tsabari, A. (2020). Can science literacy help individuals identify misinformation in everyday life?. Science Education, 104(5), 873-894. https://doi.org/10.1002/sce.21581

Shirky, C. (2010). Cognitive Surplus. How Technology Makes Consumers into Collaborators. Penguin.

Shu, K., Mahudeswaran, D. & Liu, H. (2019). FakeNewsTracker: a tool for fake news collection, detection, and visualization. Computational and Mathematical Organization Theory, 25, 60–71. https://doi.org/10.1007/s10588-018-09280-3

Soe, S.O. (2018). Algorithmic detection of misinformation and disinformation: Gricean perspectives. Journal of Documentation, 74(2), 309-332. https://doi.org/10.1108/JD-05-2017-0075

Swire, B., Berinsky, A. J., Lewandowsky, S., & Ecker, U. K. (2017). Processing political mis- information: Comprehending the Trump phenomenon. Royal Society Open Science, 4(3). https://doi.org/10.1098/rsos.160802

Tacchini, E., Ballarin, G., Della Vedova, M. L., Moret, S., & de Alfaro, L. (2017). Some like it hoax: Automated fake news detection in social networks. arXiv. https://cutt.ly/wbVJVHt

Tandoc, E. C. (2019). The facts of fake news: A research review. Sociology Compass, 13(9), e12724. https://doi.org/10.1111/soc4.12724

Tapscott D. & Tapscott A. (2016). Blockchain Revolution: How the Technology Behind Bitcoin is Changing Money, Business and the World. Random House, New York

Torabi-Asr, F., & Taboada, M. (2019). Big Data and quality data for fake news and misinformation detection. Big Data & Society, 6(1). https://doi.org/10.1177/2053951719843310

Volkova, S., Shaffer, K., Jang, J. Y., & Hodas, N. (2017). Separating facts from fiction: Linguistic models to classify suspicious and trusted news posts on Twitter. Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics, 2, 647-653. https://cutt.ly/DbVKpTG

Vos, T. P., & Thomas, R. J. (2018). The discursive construction of journalistic authority in a post-truth age. Journalism Studies, 19(13), 2001-2010. https://doi.org/10.1080/1461670X.2018.1492879

Vosoughi, S., Roy, D., & Aral, S. (2018). The spread of true and false news online. Science, 359(6380), 1146-1151. https://dx.doi.org/10.1126/science.aap9559

Vraga, E. K., & Bode, L. (2017). Using expert sources to correct health misinformation in social media. Science Communication, 39(5), 621-645. https://doi.org/10.1177/1075547017731776

Vraga, E. K., Bode, L., & Tully, M. (2020). Creating news literacy messages to enhance expert corrections of misinformation on Twitter. Communication Research. https://doi.org/10.1177/0093650219898094

Wang, Y., Ma, F., Jin, Z., Yuan. Y., Xun, G., Jha, K., Su, L. & Gao, J. (2018). EANN: Event adversarial neural networks for multi-modal fake news detection. Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, 849-857. https://cutt.ly/vbVFdE8

Walter, N., Brooks, J. J., Saucier, C. J., & Suresh, S. (2020). Evaluating the impact of attempts to correct health misinformation on social media: A meta-analysis. Health Communication, 1-9. https://doi.org/10.1080/10410236.2020.1794553

Wu, L., & Liu, H. (2018). Tracing fake-news footprints: Characterizing social media messages by how they propagate. Proceedings of the eleventh ACM international conference on Web Search and Data Mining, 637-645. https://doi.org/10.1145/3159652.3159677

Wu, Y., Ngai, E. W., Wu, P., & Wu, C. (2020). Fake online reviews: Literature review, synthesis, and directions for future research. Decision Support Systems, 132, 113280. https://doi.org/10.1016/j.dss.2020.113280

Zhang, C.W.; Gupta, A., Kauten, C., Deokar, A.V. & Qin, X. (2019). Detecting fake news for reducing misinformation risks using analytics approaches. European Journal of Operational Research, 279(3), 1036-1052. https://doi.org/10.1016/j.ejor.2019.06.022

Zollo, F., Bessi, A., Del Vicario, M., Scala, A., Caldarelli, G., Shekhtman, L. & Quattrociocchi, W. (2017). Debunking in a world of tribes. PloS one, 12(7), e0181821. https://doi.org/10.1371/journal.pone.0181821

Zubiaga, A., Aker, A., Bontcheva, K., Liakata, M. & Procter, M. (2018). Detection and Resolution of Rumours in Social Media: A Survey. ACM Computing Surveys, 51(2). https://doi.org/10.1145/3161603


Notas

1This paper is supported by the Cátedra Jean Monnet EUDFAKE: EU, disinformation and fake news (Call 2019 – 610538-EPP1-2019-1-ES-EPPJMO-CHAIR) financed by the European Commission’s Erasmus+ programme and the “Rationality and counter-knowledge. Epistemology of the detection of falsehoods in news stories” (RTI2018-097709-B-I00) project financed by the Spanish Ministry of Science, Innovation and Universities.

7. Appendix

Appendix 1. Cartography of the solutions analysed by the research reviewed

Solutions and strategies against disinformation (n=144; 23.80%)

1.1. Related to social networks and platforms

1.1.1. Detection of disinformation on social networks

1.1.2. Correction of information on social networks

1.1.3. Evaluation of the quality of the sources on social networks

1.1.3.1. Collaborative evaluation of the sources on social networks

1.1.4. Appropriate uses for reporting

1.1.5. Censure

1.1.6. Labelling on social networks

1.1.7. Evaluation of the impact of attempts to correct information on social networks

1.1.8. Crowdsourcing to disprove false content

1.1.9. Limits on propagation

1.2. Communication strategies

1.2.1. Effective communication of science

1.2.2. Counter-propaganda

1.2.3. Discursive strategies against scientific negationism

1.2.4. Communication in reputational crisis due to fake news

1.2.5. Rhetorical strategies to debate

1.2.6. Counteract populist discourse

1.3. Role of the media

1.3.1. Fact-checking

1.3.1.1. Studies on its effectiveness

1.3.1.2. Perceptions of its effectiveness & credibility among professionals & citizens

1.3.1.3. Use of messaging networks & services for fact-checking

1.3.1.4. Transparency in fact-checking

1.3.1.5. Effects on the credibility of politicians

1.3.1.6. Comparative study of verifications with or without humour

1.3.2. Collaborative investigative journalism

1.3.3. Boundary work (establishment of journalistic limits and standards)

1.3.4. Increase confidence in journalism

1.3.5. Adoption of more horizontal communicative models with the audiences (from traditional media)

1.4. Policies against disinformation

1.4.1. Regulation of information

1.5. Algorithmic, automated & technological solutions

1.5.1. Predictive models of the production of fake news

1.5.2. Tools for detecting and visualising fake news

1.5.3. Algorithmic evaluation of source credibility

1.5.4. Algorithmic moderation of content

1.5.5. Use of Big Data to detect disinformation

1.5.6. Use of the blockchain

1.5.7. Neural networks (Artificial Intelligence) to detect fake news

1.5.8. Automatic detection of bots

1.6. Education

1.6.1. Educational games about fake news (inoculation)

1.6.2. Education & vaccines

1.6.3. Media education & the perception of veracity

1.6.4. Education & critical thought

1.6.6. Effectiveness of media & news education

1.6.7. Scientific education

1.7. The role of libraries

1.8. Increase trust in institutions

1.8.1. Scientific rigour & transparency

1.8.2. Trust in academia & universities

1.9. Role of art against disinformation

1.10. Ethical implications of the fight against disinformation

1.11. Verification & evaluation of sources

1.12. Meta-analysis of the solutions proposed


doxa.comunicación | nº 33, pp. 321-346 | July-December of 2021

ISSN: 1696-019X / e-ISSN: 2386-3978