Facebook, Twitter and the Digital Disinformation Mess

By Shelly Banjo

Disinformation, now known as fake news, has tainted public discourse for centuries, even millennia. It’s been amplified in our digital age as a weapon of fearmongers, mob-baiters and election-meddlers to widen social fissures, subvert democracy and boost authoritarian regimes. Companies such as Facebook, Twitter and Google are under pressure to take action.

1. What is disinformation?

It’s often defined as false content spread with the specific intent to deceive, mislead or manipulate. (That’s different from misinformation, which is erroneous but spread unintentionally.) Disinformation can take the form of legitimate-looking news stories, tweets, Facebook or Instagram posts, advertisements and edited recordings distributed on social media or by messaging app. A new worry is what are called deepfakes: video or audio clips in which computers can literally put words in someone’s mouth.AD

2. What’s different in the internet age?

Barriers to mass communication are lower. With platforms such as Facebook and Twitter, modern-day purveyors of disinformation need only a computer or smartphone and an internet connection to reach a potentially huge audience — openly, anonymously or disguised as someone or something else, such as a genuine grassroots movement. In addition, armies of people, known as trolls, and so-called internet bots — software that performs automated tasks quickly — can be deployed to drive large-scale disinformation campaigns.

3. What’s the harm?

If the global reach of social media were being used merely to spread messages of peace and harmony — or just to make money — maybe there wouldn’t be any. But the purposes are often darker. In what’s known as state-sponsored trolling, for instance, governments create digital hate mobs to smear critical activists or journalists, suppress dissent, undermine political opponents, spread lies and control public opinion.AD

4. Who produces disinformation?

Researchers at the University of Oxford this year found evidence of “social media manipulation campaigns” by governments or political parties in 70 countries, up from 28 countries in 2017, with Facebook being the top venue where the disinformation is disseminated. Discussion of government-directed campaigns usually starts with Russia. But the Oxford report singles out China as having become “a major player in the global disinformation order.” Along with those two countries, five others — India, Iran, Pakistan, Saudi Arabia and Venezuela — have used Facebook and Twitter “to influence global audiences,” according to the Oxford report.

5. What has China done?

Twitter and Facebook, in August 2019, revealed a Chinese state-backed information operation launched globally to de-legitimize the pro-democracy movement in Hong Kong. Twitter said it had taken down 936 accounts that were “deliberately and specifically attempting to sow political discord in Hong Kong.” Facebook said it had found a similar Chinese government-backed operation and deleted fake accounts. It said it doesn’t want its services “to be used to manipulate people.”AD

6. What has Russia done?

A Rand Corp. study of the conflict in eastern Ukraine, which has claimed some 13,000 lives since 2014, found the Russian government under President Vladimir Putin ran a sophisticated social media campaign that included fake news, Twitter bots, unattributed comments on web pages and made-up hashtag campaigns to “mobilize support, spread disinformation and hatred and try to destabilize the situation.” Another Russian effort targeted the 2016 U.S. presidential election, reaching millions of American voters with phony posts and ads that sought to exploit divisions on hot-button issues.

7. Where else is this a problem?

Some examples:

• Before India’s 2019 elections, shadowy marketing groups connected to politicians used the WhatsApp messaging service to spread doctored stories and videos to denigrate opponents. The country also has been plagued with deadly violence spurred by rumors that spread via WhatsApp groups.AD

• A study of 100,000 political images shared on WhatsApp in Brazil in the run-up to its 2018 election found that more than half contained misleading or flatly false information; It’s unclear who was behind them.

• In countries such as Sri Lanka and Malaysia, fake news on Facebook has become a battleground between Buddhists and Muslims. In one instance in Sri Lanka, posts falsely alleging that Muslim shopkeepers were putting sterilization pills in food served to Buddhist customers led to a violent outburst in which a man was burned to death.

• In Myanmar, a study commissioned by Facebook blamed military officials for using fake news to whip up popular sentiment against the Rohingya minority, helping to set the stage for what UN officials have described as genocide.

8. How does digital disinformation work?AD

A blatant falsehood might spring up on something that resembles a legitimate news website — with names such as newsexaminer.net or WorldPoliticus.com — and go viral when it’s tweeted by someone with lots of followers or turned into a “trending” YouTube video. The most sophisticated disinformation operations use troll farms, artificial intelligence and internet bots — what the Oxford researchers call “cyber troops” — to flood the zone with social-media posts or messages to make a fake or doctored story appear authentic and consequential. Fake news can be a complete fabrication (the pope didn’t really endorse Donald Trump), but often there’s a kernel of truth that’s taken out of context or edited to change its meaning.

9. How are social-media companies responding?

Under pressure from lawmakers and regulators, Facebook and Google (a unit of Alphabet Inc.) have started requiring political ads in the U.S. and Europe to disclose who is behind them. In October Twitter said it would ban all political advertising globally, days after Facebook Chief Executive Officer Mark Zuckerberg defended his company’s policy of not fact-checking ads from politicians. Google’s YouTube division adjusted its “up next” algorithms to limit recommendations for suspected fake or inflammatory videos, a move it had resisted for years. WhatsApp now limits, to five, how many people or groups a message can be forwarded to. Its parent company, Facebook, said it spent 18 months preparing for India’s 2019 election: It blocked and removed fake accounts, looked for attempts at meddling and partnered with outside fact-checkers (albeit relatively few) to combat fake news. Facebook has developed artificial intelligence tools to help identify content that’s abusive or otherwise violates the site’s policies. In the wake of the March 15 shooting massacre in Christchurch, New Zealand, Facebook, Google and Twitter signed a voluntary agreement with world leaders pledging to fight hate speech online.AD

10. What are governments doing?

A Singapore law that took effect Oct. 2 allows for criminal penalties of up to 10 years in prison and a fine of up to S$1 million ($720,000) for anyone convicted of spreading online inaccuracies. The responsibility for identifying falsehoods detrimental to the public interest was given to government ministers. Malaysia enacted a similar law that the government, elected last year, is trying to repeal. Indonesia set up a 24-hour “war room” ahead of its 2019 elections to fight hoaxes and fake news. France has a new law that allows judges to determine what is fake news and order its removal during election campaigns. In the U.S., efforts to crack down on disinformation can run up against the guarantee of free speech, although some platforms have begun to restrict postings by anti-vaccine activists, for example. Then there’s the Philippines, where the government of President Rodrigo Duterte encourages “patriotic trolling” to undermine his critics.

Leave a Reply

Your email address will not be published. Required fields are marked *

This site uses Akismet to reduce spam. Learn how your comment data is processed.