How a Twitter plan to counter extremism fell apart
Here’s your second free edition of Platformer this week: another investigation, reported with Zoe Schiffer and co-published with The Verge, into a Twitter project called Redirect that was showing promise in steering users away from extremism before Elon Musk’s purchase of the company threw it into turmoil. Do you value the work it takes to deliver two investigative reports in a week? If so, consider upgrading your membership. It will make more of this work possible — and help demonstrate that this kind of independent investigative reporting can thrive. Plus, I’ll send you a link to our chatty Discord server, where we have been talking about Twitter all week. How a Twitter plan to counter extremism fell apartA research team was midway through a project to help troubled users. Then Elon Musk bought the companyProgramming note: Platformer is off Monday for Labor Day. Zoe Schiffer and I reported this story for The Verge. It had been a long pandemic for Twitter’s research team. Tasked with solving some of the platform’s toughest problems around harassment, extremism, and disinformation, staffers absconded to Napa Valley in November 2021 for a company retreat. Despite a tumultuous change in leadership — Jack Dorsey had recently stepped down, appointing former chief technology officer Parag Agrawal to take his place — the group felt unified, even hopeful. After months of fighting bad actors online, employees took a moment to unwind. “We finally felt like we had a cohesive team,” one researcher says. But at the goodbye brunch on the last day, people’s phones started pinging with alarming news: their boss, Dantley Davis, Twitter’s vice president of design, had been fired. Nobody knew it was coming. “It was like a movie,” says one attendee, who asked to remain anonymous because they are not authorized to speak publicly about the company. “People started crying. I was just sitting there eating a croissant being like, ‘What’s up with the mood?’” The news foreshadowed a downward spiral for the research organization. Although the group was used to reorganizations, a shakeup in the middle of an outing meant to bond the team together felt deeply symbolic. The turmoil came to a head in April, when Elon Musk signed a deal to buy Twitter. Interviews with current and former employees, along with 70 pages of internal documents, suggest the chaos surrounding Musk’s acquisition pushed some teams to the breaking point, prompting numerous health researchers to quit, with some saying their colleagues were told to to deprioritize projects to fight extremism” in favor of focusing on bots and spam. The Musk deal might not even go through, but the effects on Twitter’s health efforts are already clear. The health team, once tasked with fostering civil conversations on the famously uncivil platform, went from 15 full-time staffers down to two. In 2019, Jack Dorsey asked a fundamental question about the platform he had helped create: “Can we actually measure the health of the conversation?” Onstage at a TED conference in Vancouver, the beanie-clad CEO talked earnestly about investing in automated systems to proactively detect bad behavior and “take the burden off the victim completely.” That summer, the company began staffing up a team of health researchers to carry out Dorsey’s mission. His talk convinced people who’d been working in academia, or for larger tech companies like Meta, to join Twitter, inspired by the prospect of working toward positive social change. When the process worked as intended, health researchers helped Twitter think through potential abuses of new products. In 2020, Twitter was working on a tool called “unmention” that allows users to limit who can reply to their tweets. Researchers conducted a “red team” exercise, bringing together employees across the company to explore how the tool could be misused. Unmention could allow “powerful people [to] suppress dissent, discussion, and correction” and enable “harassers seeking contact with their targets [to] coerce targets to respond in person,” the red team wrote in an internal report. But the process wasn't always so smooth. In 2021, former Twitter product chief Kayvon Beykpour announced the company’s number one priority was launching Spaces. (“It was a full on assault to kill Clubhouse,” one employee says.) The team assigned to the project worked overtime trying to get the feature out the door and didn’t schedule a red team exercise until August 10th — three months after launch. In July, the exercise was canceled. Spaces went live without a comprehensive assessment of the key risks, and white nationalists and terrorists flooded the platform, as The Washington Post reported. When Twitter eventually held a red team exercise for Spaces in January 2022, the report concluded: “We did not prioritize identifying and mitigating against health and safety risks before launching Spaces. This Red Team occurred too late. Despite critical investments in the first year and a half of building Spaces, we have been largely reactive to the real-world harms inflicted by malicious actors in Spaces. We have over relied on the general public to identify problems. We have launched products and features without adequate exploration of potential health implications.” Earlier this year, Twitter walked back plans to monetize adult content after a red team found that the platform had failed to adequately address child sexual exploitation material. It was a problem researchers had been warning about for years. Employees said that Twitter executives have been aware of the problem but noted the company has not allocated the resources necessary to fix it. By late 2021, Twitter’s health researchers had spent years playing whack-a-mole with bad actors on the platform and decided to deploy a more sophisticated approach to dealing with harmful content. Externally, the company was regularly criticized for allowing dangerous groups to run amok. But internally, it sometimes felt as though certain groups, like conspiracy theorists, were kicked off the platform too soon — before researchers could study their dynamics. “The old approach was almost comically ineffective, and very reactive — a manual process of playing catch,” says a former employee, who asked to remain anonymous because they are not authorized to speak publicly about the company. “Simply defining and catching ‘bad guys’ is a losing game.” Instead, researchers hoped to identify people who were about to engage with harmful tweets, and nudge them toward healthier content using pop-up messages and interstitials. “The pilot will allow Twitter to identify and leverage behavioral — rather than content — signals and reach users at risk from harm with redirection to supportive content and services,” read an internal project brief, viewed by The Verge. Twitter researchers partnered with Moonshot, a company that specializes in studying violent extremists, and kicked off a project called Redirect, modeled after work that Google and Facebook had done to curb the spread of harmful communities. At Google, this work had resulted in a sophisticated campaign to target people searching for extremist content with ads and YouTube videos aimed at debunking extremist messaging. Twitter planned to do the same. The goal was to move the company from simply reacting to bad accounts and posts to proactively guiding users toward better behavior. Read the rest of this story at The Verge. Governing
Industry
Those good tweetsi love when my cat pushes open the bathroom door the same way a cowboy would bust open the doors to a saloon Talk to meSend me tips, comments, questions, and edited tweets: casey@platformer.news. By design, the vast majority of Platformer readers never pay anything for the journalism it provides. But you made it all the way to the end of this week’s edition — maybe not for the first time. Want to support more journalism like what you read today? If so, click here: |
Older messages
Twitter's other porn problem
Thursday, September 1, 2022
Imagine the company did want to monetize adult content. Would Apple let it?
🚨 How Twitter’s child porn problem ruined its plans for an OnlyFans competitor
Tuesday, August 30, 2022
Internal documents and Twitter employees reveal the need for massive investment to remove illegal content — but executives haven't listened
Why Reddit is still betting on the blockchain
Friday, August 26, 2022
CEO Steve Huffman on making the web more user-friendly — and his plan to "not use any crypto words"
Mudge blows the whistle on Twitter
Wednesday, August 24, 2022
It's a gift to Elon Musk. But how much will it matter?
How a viral NFT project survived the crypto crash
Friday, August 19, 2022
Loot's Dom Hofmann on outlasting the hype cycle, public-domain video games, and why Creative Commons is the future
You Might Also Like
⏰ 48 hours left - the #1 reason an ecommerce venture fails
Friday, January 10, 2025
Don't risk your time and money—learn how to find and test winning products. Hi Friend , Less than 48 hours left—so please pay attention. Here's a hard truth: 90% of ecommerce stores fail. Not
Meta just killed its diversity, equity and inclusion program
Friday, January 10, 2025
What employees are saying about the company's embrace of MAGA ideology —and what Meta is telling them not to say Platformer Platformer Meta just killed its diversity, equity and inclusion program
quitters day
Friday, January 10, 2025
Read time: 51 sec. You gave up already, didn't you? I'm not trying to be ad*ck 😆 It's just a fact: today is National Quitter's Day. The day 80% of people give up on their New Year's
🌟 Social Media Trends, AI Tools, and Expert Marketing Tutorials!🚀
Friday, January 10, 2025
Discover the latest on social media trends for 2025, Google's evolving ad campaigns, and YouTube's 3-minute Shorts. Plus, explore AI-driven tools like TopView 2.0 and Fenado AI, alongside must-
We found the best time to post on Instagram
Friday, January 10, 2025
Plus, Creator Camp is back! ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏
10words: Top picks from this week
Friday, January 10, 2025
Today's projects: CareerCode.it • Lesson Bud • NorthPoll • Webtwizz • FineVoice • Converti • Seller Terminal • HabitStack • Ariglad • OutSkill • edesy.in • Grow My Small Business AI 10words
Issue #134: Building $1K-$10K MRR Micro SaaS Products: RAG-as-a-Service, AI Voice Agent for Appointments, Employee…
Friday, January 10, 2025
Build Profitable SaaS products!! ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏
Expect-AI-tions — The Bootstrapped Founder 369
Friday, January 10, 2025
Expectations around AI tooling are changing, and software founders will be the first to either provide what people want — or perish. ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏
The quantum talent squeeze
Friday, January 10, 2025
+ One of Europe's largest ever Series A rounds; Northvolt sister company Stegra 'fully financed' View in browser Sponsored by Zoom Good morning there, “We're expecting a spectacular
Inside Meta’s dehumanizing new speech policies for trans people
Friday, January 10, 2025
"A trans person isn't a he or she, it's an it," reads a new guideline telling moderators what is now allowed on Facebook and Instagram Platformer Platformer Inside Meta's