Facebook, Twitter, YouTube are pressed on ‘poisonous’ algorithms
Executives from Facebook Inc., Twitter Inc. and Alphabet Inc.’s YouTube were pressed by lawmakers Tuesday on how user content is shared and highlighted on their platforms through algorithms that one senator said can be misused, “driving us into poisonous echo chambers.”
Sen. Ben Sasse, the top Republican on the Senate Judiciary Committee’s panel on Privacy, Technology and the Law, made the comment as members examined algorithms — the lines of software code that determine how user-generated information is displayed and who gets to see it.
“Algorithms, like almost any technologies that are new, have costs and benefits” and can be abused, said Sasse, who is from Nebraska.
The hearing took place as Congress is considering how to overhaul Section 230, a provision of the 1996 communications law that protects internet companies from liability for user content. One House proposal would make social media platforms responsible for the way content is shared and amplified through algorithms.
“I plan to use this hearing as an opportunity to learn about how these companies’ algorithms work, what steps may have been taken to reduce algorithmic amplification that is harmful and what can be done better,” said Delaware Sen. Chris Coons, a Democrat and the subcommittee’s chairman, as he opened the hearing.
Illinois Sen. Richard J. Durbin, the Democratic chairman of the full Judiciary Committee, urged social media companies to do more to remove harmful content, citing the Jan. 6 attack on the U.S. Capitol. He said domestic extremists organized and shared disinformation on some of the platforms represented at Tuesday’s hearing.
Monika Bickert, Facebook’s vice president for content policy, testified that its tools make the platform’s algorithm more transparent, so users can see why certain posts appear on their news feeds.
“It is not in our interest financially or reputationally” to push people toward extremist content, Bickert said.
Lauren Culbertson, Twitter’s head of U.S. public policy, highlighted the positive uses for algorithms and machine learning, especially the ability to recognize harmful content to review and remove. She said in her opening statement that Twitter is committed to studying the unintended consequences of algorithms and to giving users more choice over how algorithms shape their experience.
“As members of Congress and other policymakers debate the future of internet regulation, they should closely consider the ways technology, algorithms, and machine learning make Twitter a safer place for the public conversation and enhance the global experience with the internet at large,” Culbertson said.
The social network’s crackdown on posts that may incite violence raises the question of why it doesn’t do more to discourage abuse all the time.
Alexandra Veitch, YouTube’s director of government affairs and public policy for the Americas and emerging markets, said the service uses an automated process to detect videos that violate the company’s policies, and algorithms can be used to promote trusted sources and minimize content that’s questionable. She described YouTube as “not just a hobby, but a business” for people who create and share videos on the platform.
But Tristan Harris, co-founder and president of the Center for Humane Technology, dismissed the testimony by the company executives, saying “it’s almost like having the heads of Exxon, BP and Shell asking about what are you doing to responsibly stop climate change.”
Harris, a former design ethicist at Google, said “their business model is to create a society that is addicted, outraged, polarized, performative and disinformed. That’s just the fundamentals of how it works.”
The role that algorithms play in sharing information — and disinformation — has taken on renewed importance as people turn to social media to learn and comment on issues such as COVID-19 vaccines, protests over police killings and election security. As Durbin indicated, the platforms have been under increased scrutiny since supporters of former President Trump amplified disinformation ahead of the Jan. 6 attack.
Trump was suspended by Facebook, Twitter and YouTube for comments that the companies said could lead to violence. Facebook’s Oversight Board is reviewing the decision, while YouTube has left open the possibility of reversing the suspension. Twitter said its ban of Trump’s account is permanent.
Frozen in time since the Jan. 6 insurrection, perhaps forever, Donald Trump’s Facebook page lives on as an internet destination for #MAGA fans and #Resistance types alike.
But Sen. Charles E. Grassley of Iowa, the top Republican on the full Senate Judiciary Committee, voiced the frequent GOP complaint that social media platforms censor conservatives. He described the companies as monopolies that don’t face the competition that would make them more responsible with user information. “We must look at the power and control that a handful of companies have over speech,” Grassley said.
Facebook has been advocating for updated internet regulation, including new privacy rules. It also has called for election protection measures and an overhaul of Section 230 to require more transparency, reporting requirements and best-practice guidelines for larger companies. As part of that campaign, Facebook is buying ads in the nation’s capital pointing out how much the internet has changed in the 25 years since current regulations became law.