- Instagram influencer says her account was banned over ‘sexual’ pregnancy photo 2 Years Ago
- YouTube time traveler emotionally describes floating cities in the year 2300 Today 10:15 AM
- Trump’s former campaign manager admits to lying to the media—gets CNN appearance Today 10:15 AM
- Kyrsten Sinema may face a censure vote—and net neutrality is a big reason why Today 8:36 AM
- Recreate a Hogwarts holiday with the LEGO ‘Harry Potter’ Advent calendar Today 8:27 AM
- How to stream Titans vs. Jaguars on Thursday Night Football Today 8:26 AM
- 24 Halloween costumes so weird all you can do is laugh Today 8:13 AM
- Night Monkey finally gets the trailer he deserves Today 8:04 AM
- All the TV series and films coming to AppleTV+ Today 8:00 AM
- How to watch ‘American Horror Story: 1984’ Today 7:00 AM
- What’s new in Call of Duty: Modern Warfare? Today 7:00 AM
- ‘Carole and Tuesday’ is a feast for the eyes, ears, and heart Today 6:30 AM
- Tara Booth’s Instagram art embraces the comedy in mental health struggles Today 6:00 AM
- Everything we know so far about Peacock, NBC’s new streaming service Tuesday 7:42 PM
- Selena Gomez producing docuseries about immigration for Netflix Tuesday 7:11 PM
Researchers at Indiana University studied 400,000 claims made by 122 alleged fake news sites like Breitbart, Infowars, PoliticusUSA, and satirical publications like the Onion. At the same time, they monitored more than 1 million Twitter posts referencing 15,000 stories written by fact-checking sites, including Snopes, Politifact, and FactCheck.
From that data set, researchers looked at which accounts were spreading the news and sampled 200 of their most recent tweets. Using machine learning, they were able to determine whether the account belonged to a human or an automated social bot.
“From this data we extract features capturing various dimensions of information diffusion as well as user metadata, friend statistics, temporal patterns, part-of-speech and sentiment analysis,” the study says. “These features are fed to a machine learning algorithm trained on thousands of examples of human and bot accounts.”
The algorithm, called Belometer, found that social bots play a key role in the spread of fake news.
“Relatively few accounts are responsible for a large share of the traffic that carries misinformation. These accounts are likely bots, and we uncovered several manipulation strategies they use.”
Researchers believe these strategies are why Twitter bots are so effective. First, they amplify fake news in its early stages, long before it goes viral. Then they target individual users through replies and mentions, instead of writing broad posts or retweeting. This increases the chances a post could go viral because it injects fake news directly into a closely connected human network. Finally, bots disguise themselves as human by changing their geographic location. These manipulations are largely why people spread false news from bots just as much as other humans, according to the study.
While the study only looked at Twitter, researchers believe other social platforms are just as vulnerable. Conspiracy theories are also spread on Facebook by accounts managed automatically and can spread just as fast as real news. Unfortunately, the increasing popularity of ephemeral platforms like Snapchat and Sarahah makes studying the spread of fake news nearly impossible.
Still, researchers are confident their findings will help social platforms make changes to end the spread of false information, even if it isn’t clear what route to take. The study presents two possible solutions: using machine learning algorithms to detect and shut down bots, or deploying CAPTCHAs, a proven method for distinguishing between a human and machine.
Phillip Tracy is a former technology staff writer at the Daily Dot. He's an expert on smartphones, social media trends, and gadgets. He previously reported on IoT and telecom for RCR Wireless News and contributed to NewBay Media magazine. He now writes for Laptop magazine.