How AI can help to prevent the spread of disinformation

Ben Lorica, chief data scientist at O’Reilly Media, looks at how artificial intelligence can stop the spread of disinformation — more commonly known as fake news How AI can help to prevent the spread of disinformation image

Our industry has a duty to discuss the dark side of technology. Yet many organisations — including some that wield enormous power and influence — are reluctant to acknowledge that their platforms are used to spread disinformation, foster hatred, facilitate bullying, and much else that makes our world a worse place in which to live.

Disinformation — what is sometimes called “fake news” — is a prime example of the unintended consequences of new technology. Its purpose is purely to create discord; it poisons public discourse and feeds festering hatreds with a litany of lies. What makes disinformation so effective is that it exploits characteristics of human nature such as confirmation bias, then seizes on the smallest seed of doubt and amplifies it with untruths and obfuscation.

UK government takes on fake news in effort to curb misinformation, and brands Facebook ‘digital gangsters’

Earlier this week, the UK government called out the tech giants, following an MPs report on fake news, as part of their efforts to curb disinformation. Read here

Disinformation has spawned a new sub-industry within journalism, with fact checkers working around the clock to analyse politicians’ speeches, articles from other publications and news reports, and government statistics among much else. But the sheer volume of disinformation, together with its ability to multiply and mutate like a virus on a variety of social platforms, means that thorough fact-checking is only possible on a tiny proportion of disputed articles.

While technology has provided the seedbed and distribution for disinformation, it also offers a solution to the issue. Artificial intelligence in particular offers powerful tools in the fight against disinformation, working on multiple levels to identify dubious content. These techniques are broadly split between content-based and response-based identification. The former works much like a human fact checker, by matching the content of an article with trusted sources of information to highlight errors or outright lies.

But disinformation is an insidious beast and doesn’t always include facts that can be checked. This could involve a distorted or mis-captioned image, highly tendentious or biased reporting, or misleading stories that are not based on facts but rather use specious arguments to promote a particular cause. Another issue is false positives generated by satire or parody articles (which can be hard enough for many humans to detect without a winking emoji).

This is where response-based identification brings real value. Rather than relying on the text of an article as the primary source of information, this technique examines patterns of propagation as the news spreads through social media. By looking at ‘likes’, comments, temporal patterns in the spread of stories and the reputation of those who post and engage with the content, analysts can build a very clear idea of how trustworthy it is.

These concepts harness many of the techniques developed in the field of digital media forensics, a discipline dedicated to identifying issues such as plagiarism and “forged media”, where genuine content is digitally manipulated, or fake articles fabricated to look like they come from reputable news sources.

Thanks to this pioneering work, there is now a range of incredibly sophisticated tools, many harnessing the power of machine learning. These include signal processing analysis, which can identify bad actors through their use of compression software; physics-level analysis, which examines inconsistencies in lighting, landscapes shadows, and the like; as well as techniques that look at semantic and even physiological signals.

Cyber populism: is social media damaging democracy?

In a talk at Chatham House, a panel of experts discussed the growing problem of the manipulation of information on social media. Read here

There was a compelling, although distressing, illustration of these techniques last summer. A team of fact checkers at the BBC used multiple open-source forensic investigation technologies (including Google Earth) to prove that a disturbing video of an atrocity had, in fact, been committed by government soldiers – a claim that the state had initially decried as “fake news”.

Tools such as these are crucial front-line weapons in the war against disinformation. But identifying disinformation is only the start: publishers and other organisations need to back these up with robust intervention strategies to take down or limit the spread of this content as soon as it appears, and to ensure that those who are exposed to it are alerted and served with content that counters the false information in the original – a process known as “decontamination”. One promising decontamination strategy that researchers are examining is the competing cascade, which places trusted, truthful information directly into a user’s newsfeed to compete with the lies in the original article.

The internet was supposed to usher in a new era for humanity, bringing unprecedented knowledge to the whole world — and in many ways it has. Disinformation is the antithesis of this dream, poisoning the well from which we all drink. But let’s not forget that we’re still in our digital infancy, still working out how to combat the raft of new societal problems that the Internet has created. With disinformation and “fake news”, we have the will and the technology to fight against it.

Written by Ben Lorica, chief data scientist at O’Reilly Media

Written by Ben Lorica, chief data scientist at O’Reilly Media

Latest news

divider
Digital Transformation
The four steps you need to take to kick-start the leadership revolution

The four steps you need to take to kick-start the leadership revolution

26 March 2019 / Organisations are significantly increasing their investment in new digital technologies. Yet, it would be easy [...]

divider
Data Analytics & Data Science
The unstructured data pandemic

The unstructured data pandemic

26 March 2019 / Gartner estimates that today over 80% of enterprise data is unstructured. That means that the [...]

divider
Business & Strategy
Majority of remote workers are being excluded from meetings

Majority of remote workers are being excluded from meetings

25 March 2019 / New data from Igloo Software released today shows 56% of remote employees missed out on [...]

divider
Business Skills
Bracing for the inevitable skills crisis

Bracing for the inevitable skills crisis

25 March 2019 / EU migration hit a six-year low following the Brexit vote. For industries projected to grow, [...]

divider
Cybersecurity
Lack of skills the main challenge in recruiting information security talent

Lack of skills the main challenge in recruiting information security talent

25 March 2019 / Despite, Raj Samani — chief data scientist at McAfee — last week claiming that there [...]

divider
AI & Machine Learning
Kasparov and AI: the gulf between perception and reality

Kasparov and AI: the gulf between perception and reality

25 March 2019 / Kasparov and AI “Chess used to be connected to the mysteries of human intelligence,” said [...]

divider
AI & Machine Learning
Machine learning is not real learning argues a new paper

Machine learning is not real learning argues a new paper

25 March 2019 / Artificial intelligence is a misleading phrase, or so many argue. It is misleading because there [...]

divider
Automotive
Overcoming simulation hurdles to expedite the fully autonomous vehicle

Overcoming simulation hurdles to expedite the fully autonomous vehicle

25 March 2019 / Not a day goes by without artificial intelligence (AI) and machine learning (ML) making headlines. [...]

divider
Data Analytics & Data Science
Data gurus comment on Revoke Article 50 petition

Data gurus comment on Revoke Article 50 petition

22 March 2019 / “You have to take the data with a pinch of salt, especially as the second [...]

Do NOT follow this link or you will be banned from the site!

Pin It on Pinterest