Fixing What the Internet Broke

Eufemia Didonato

This article is part of the On Tech newsletter. You can sign up here to receive it weekdays.

January’s riot at the U.S. Capitol showed the damage that can result when millions of people believe an election was stolen despite no evidence of widespread fraud.

The Election Integrity Partnership, a coalition of online information researchers, published this week a comprehensive analysis of the false narrative of the presidential contest and recommended ways to avoid a repeat.

Internet companies weren’t solely to blame for the fiction of a stolen election, but the report concluded that they were hubs where false narratives were incubated, reinforced and cemented. I’m going to summarize here three of the report’s intriguing suggestions for how companies such as Facebook, YouTube and Twitter can change to help create a healthier climate of information about elections and everything else.

One broad point: It can feel as if the norms and behaviors of people online are immutable and inevitable, but they’re not. Digital life is still relatively new, and what’s good or toxic is the result of deliberate choices by companies and all of us. We can fix what’s broken. And as another threat against the Capitol this week shows, it’s imperative we get this right.

The hard part, though, is not necessarily making policies. It’s enforcing them when doing so could trigger a backlash.

2) Internet companies should tell us what they’re doing and why: Big websites like Facebook and Twitter have detailed guidelines about what’s not allowed — for example, threatening others with violence or selling drugs.

But internet companies often apply their policies inconsistently and don’t always provide clear reasons when people’s posts are flagged or deleted. The EIP report suggested that online companies do more to inform people about their guidelines and share evidence to support why a post broke the rules.

3) More visibility and accountability for internet companies’ decisions: News organizations have reported on Facebook’s own research identifying ways that its computer recommendations steered some to fringe ideas and made people more polarized. But Facebook and other internet companies mostly keep such analyses a secret.

The EIP researchers suggested that internet companies make public their research into misinformation and their assessments of attempts to counter it. That could improve people’s understanding of how these information systems work.

The report also suggested a change that journalists and researchers have long wanted: ways for outsiders to see posts that have been deleted by the internet companies or labeled false. This would allow accountability for the decisions that internet companies make.

There are no easy fixes to building Americans’ trust in a shared set of facts, particularly when internet sites enable lies to travel farther and faster than the truth. But the EIP recommendations show we do have options and a path forward.

Source Article

Next Post

My Sexuality Is As Fluid As My Netflix Queue

Design by Ingrid Frahm Early on in the pandemic, my girlfriend and I watch the movie Carol, and I tell her that I don’t find Cate Blanchett hot. My girlfriend scoffs when I admit this to her, and tells me that all lesbians find Cate Blanchett hot. I remind her […]