Social media can have an impressive impact on civic engagement and political discourse. Yet increasingly we find political actors using digital media and automated scripts for social control. Computational propaganda—through bots, botnets, and algorithms—has become one of the most concerning impacts of technology innovation. Unfortunately, bot identification and impact analysis are among the most difficult research challenges facing the social and computer sciences. COMPROP objectives are to advance a) rigorous social and computer science on bot use, b) critical theory on digital manipulation and political outcomes, c) our understanding of how social media propaganda impacts social movement organization and vitality. This project will innovate through i) “real-time” social and information science actively disseminated to journalists, researchers, policy experts and the interested public, ii) the first detailed data set of political bot activity, iii) a deepened regional expert network able to detect bots and their impact in Europe. COMPROP will achieve this through multi-method and reflexive work packages: 1) international qualitative fieldwork with teams of bot makers and computer scientists working to detect bots; 2a) construction of an original event data set of incidents of political bot use and 2b) treatment of the data set with fuzzy set and traditional statistics; 3) computational theory for detecting political bots and 4) a sustained dissemination strategy. This project will employ state-of-the-art “network ethnography” techniques, use the latest fuzzy set / qualitative comparative statistics, and advance computational theory on bot detection via cutting-edge algorithmic work enhanced by new crowd-sourcing techniques. Political bots are already being deployed over social networks in Europe. COMPROP will put the best methods in social and computer science to work on the size of the problem and the possible solutions.
Misinformation, Science and Media
In this new, three-year programme, researchers from the Oxford Internet Institute and the Reuters Institute for the Study of Journalism will examine the interplay between systematic misinformation campaigns, news coverage, and increasingly important social media platforms for public understanding of science and technological innovation. In some key domains of public life there appears to be coordinated efforts to ruin the reputation of science and innovation. Scientists now protest in the streets just to get policymakers to embrace evidence-based policy making. Long-held consensus on the causes and consequences of climate change, tobacco-induced cancers, and value of public health strategies increasingly seem open for debate. We have political leaders who claim to be unable to discern what expert consensus is—even when experts organize to make explicit statements about levels of confidence and certainty around particular areas of research. Social media platforms have become a powerful venue for those aiming to deflating public support for action based on reliable research, and previously trusted technological innovations come to have negligible impact. In this three-year project, we will examine the interplay between systematic misinformation campaigns, news, and increasingly important social media platforms for public understandings of science and innovation. We aim to increase our understanding of the role of “junk science” and fake news in influencing—or even undermining—public understanding of scientific issues and develop evidence-based recommendations for scientists, journalists, and policymakers interested in effective science communication in the 21st century.
The Production and Detection of Bots
Political bots are manipulating public opinion over major social networking applications. This project enables a new team of social and information scientists to investigate the impact of automated scripts, commonly called bots, on social media. The PIs will study both the bot scripts and the people making such bots, and then work with computer scientists to improve the way we catch and stop such bots. Experience suggests that political bots are most likely to appear during an international crisis, and are usually designed to promote the interests of a government in trouble. Political actors have used bots to manipulate conversations, demobilize opposition, and generate false support on popular sites like Twitter and Facebook from the U.S. as well as Sina Weibo from China
- Professor Philip Howard, University of Oxford
- Funder: National Science Foundation
- Award: $218,825.00
- Proposal Number: 8060
- Dates: 2014-2016
- Proposal Text