The growing stream of reporting on and data about fake news, misinformation, partisan content, and news literacy is hard to keep up with. This weekly roundup offers the highlights of what you might have missed.
“States or counties experiencing more fake news visitations also tended to vote for Donald Trump.” A Microsoft Research team analyzed traffic to fake news websites leading up to the 2016 U.S. presidential election and found that — at least among users of desktop browsers Microsoft Internet Explorer 11 and Edge (browsers that combined make up just about 16 percent of desktop browser usage) — “states or counties experiencing more fake news visitations also tended to vote for Donald Trump.”
Correlation found between access of info from fake news sites & votes for Trump vs. Clinton. 99% clicks via FB https://t.co/bAuBVpkv9Q @EFF pic.twitter.com/mjft39NOc1
— Eric Horvitz (@erichorvitz) October 24, 2017
“It remains to be shown if similar trends occur for other browsers and in mobile scenarios — 51.7 percent of Facebook’s worldwide active monthly users access the site exclusively from mobile devices,” they write.
Facebook, Google, and Twitter face Congress. My colleague Shan Wang wrote up Wednesday’s hearings. (“What is a bot versus a troll?”)Future trends in “information disorder.” A report released this week by Council of Europe, with support from the Shorenstein Center and First Draft, looks at “information disorder”:
While the historical impact of rumors and fabricated content have been well documented, we argue that contemporary social technology means that we are witnessing something new: information pollution at a global scale; a complex web of motivations for creating, disseminating and consuming these “polluted” messages; a myriad of content types and techniques for amplifying content; innumerable platforms hosting and reproducing this content; and breakneck speeds of communication between trusted peers.
The report by Claire Wardle and Hossein Derakhshan is a long, good overview of the space. I was particularly interested in the fourth section on future trends: “Much of the focus has been on the Facebook News Feed. But even a cursory glance outside of the U.S. demonstrates that the next frontier for mis- and dis-information is closed-messaging apps.” And, they write, “our biggest challenge will be the speed at which technology is refining the creation of fabricated video and audio.” For instance:
Audio can be manipulated even more easily than video. Adobe has created Project VoCo, which has been nicknamed ‘Photoshop for audio.’ The product allows users to feed a 10-to-20 minute clip of of someone’s voice into the application and then dictate words in that person’s exact voice. Another company called Lyrebird is working on voice generation. On its site, it claims to “need as little as one minute of audio recording of a speaker to compute a unique key defining her/his voice. This key will then generate anything from its corresponding voice.” It also plans to create an API whereby other platforms could easily use those voices.
“A dazzling array of content generators and the potential for polarization.” Chi Zhang writes about how her team at the University of Southern California is collaborating with The Alhambra Source and Asian American Advancing Justice, with support from the Tow Center for Digital Journalism, to “assess the nature of bias and misinformation in WeChat and ethnic Chinese media and to explore strategies for intervention.” During the 2016 U.S. presidential election, “anti-Hillary memes and conspiracy theories about sharia law found their way onto the mobile messaging platform, which serves the growing number of Chinese immigrants in the United States.”
What could possibly go wrong https://t.co/IBqtOS1aB4 pic.twitter.com/Ihv1HItEH5
— Alexios (@Mantzarlis) October 31, 2017