Free Speech
Permanent Suspension of Donald Trump From Facebook Now in the Hands of its Oversight Board Subsidiary

February 7, 2021—After the insurrection at the U.S. Capitol on January 6, Facebook was one of the first big tech platforms to ban former President Donald Trump from posting, announcing on January 7 that his suspension would last at least until the inauguration of President Joe Biden.
Following the inauguration, Facebook CEO Mark Zuckerberg said that he would refer the decision about whether to permanently ban Trump to the tech company’s so-called “Supreme Court,” its relatively new oversight board subsidiary.
This “referral” to Oversight Board LLC is among the first test cases for the entity – and the concept – as the major tech giants face growing complaints about politicization and even censorship.
But considerations over whether or not to permanently ban Trump is only one of the many controversial cases that will be considered by the subsidiary, according to a panel of experts discussing Facebook’s new Oversight Board at the Cato Institute on Thursday. The group become functional last year.
Cato Vice President John Samples tried to assure the panel that Oversight Board operates independently from Facebook, and said that it was comprised of board members representing many cultures and ethnicities. He emphasized that Oversight Boards’s primary goal was to propose non-binding recommendations regarding Facebook’s content moderation around the world.
Despite the perception and claims that Oversight Board will be independent, a minimum of 75 percent of the entity’s voting shares are controlled by Facebook. Facebook reserves the right to appoint or remove members of Oversight Board.
Even though the decisions proposed by Oversight Board are not binding on Facebook, Samples believes that the company would not disregard them lightly. “If they don’t [adhere to the suggested solutions], then there will be a gigantic mess because they publicly committed themselves to” do so, he said.
Others on the panel agreed. Ignoring Oversight Boards’ recommendations would only serve to undercut its independence and legitimacy, said New America’s Spandana Singh. “That would really defeat the whole purpose of the board.”
The topics to be tackled by Oversight Board range from hate speech and incitement of violence – including the January 6 insurrection — to combatting the spread of fake news.
These situation may become particularly problematic regarding world leaders that lack legitimacy, panelists said.
Autocratic systems often eschew both democratic legitimacy and liberal legitimacy, and merely need to monopolize the information systems in their country, said Samples. This, he said, makes it pretty straightforward to act against these governments and their leaders.
That is not necessarily the case in a democratic nation like the United States. “An average citizen has got to believe in the legitimacy of their leaders—they’ve got to believe in the transparency of the process—if any of those things fail, the democracy fails,” Samples said.
R Street’s Tatyana Bolton said she believed that the Trump ban would be upheld because it checks a couple of the boxes necessary to qualify for a ban. Trump’s behavior in both spreading false information and in inciting violence were against Facebook’s terms of service. Banning him would be consistent with those terms, she said.
Free Speech
Additional Content Moderation for Section 230 Protection Risks Reducing Speech on Platforms: Judge
People will migrate from platforms with too stringent content moderation measures.

WASHINGTON, March 13, 2023 – Requiring companies to moderate more content as a condition of Section 230 legal liability protections runs the risk of alienating users from platforms and discouraging communications, argued a judge of the District of Columbia Court of Appeal last week.
“The criteria for deletion are vague and difficult to parse,” Douglas Ginsburg, a Ronald Reagan appointee, said at a Federalist Society event on Wednesday. “Some of the terms are inherently difficult to define and policing what qualifies as hate speech is often a subjective determination.”
“If content moderation became very rigorous, it is obvious that users would depart from platforms that wouldn’t run their stuff,” Ginsburg added. “And they will try to find more platforms out there that will give them a voice. So, we’ll have more fragmentation and even less communication.”
Ginsburg noted that the large technology platforms already moderate a massive amount of content, adding additional moderation would be fairly challenging.
“Twitter, YouTube and Facebook remove millions of posts and videos based on those criteria alone,” Ginsburg noted. “YouTube gets 500 hours of video uploaded every minute, 3000 minutes of video coming online every minute. So the task of moderating this is obviously very challenging.”
John Samples, a member of Meta’s Oversight Board – which provides direction for the company on content – suggested Thursday that out-of-court dispute institutions for content moderation may become the preferred method of settlement.
The United States may adopt European processes in the future as it takes the lead in moderating big tech, claimed Samples.
“It would largely be a private system,” he said, and could unify and centralize social media moderation across platforms and around the world, referring to the European Union’s Digital Services Act that went into effect in November of 2022, which requires platforms to remove illegal content and ensure that users can contest removal of their content.
Section 230
Section 230 Shuts Down Conversation on First Amendment, Panel Hears
The law prevents discussion on how the first amendment should be applied in a new age of technology, says expert.

WASHINGTON, March 9, 2023 – Section 230 as it is written shuts down the conversation about the first amendment, claimed experts in a debate at Broadband Breakfast’s Big Tech & Speech Summit Thursday.
Matthew Bergman, founder of the Social Media Victims Law Center, suggested that section 230 avoids discussion on the appropriate weighing of costs and benefits that exist in allowing big tech companies litigation immunity in moderation decisions on their platforms.
We need to talk about what level of the first amendment is necessary in a new world of technology, said Bergman. This discussion happens primarily in an open litigation process, he said, which is not now available for those that are caused harm by these products.

Photo of Ron Yokubaitis of Texas.net, Ashley Johnson of Information Technology and Innovation Foundation, Emma Llanso of Center for Democracy and Technology, Matthew Bergman of Social Media Victims Law Center, and Chris Marchese of Netchoice (left to right)
All companies must have reasonable care, Bergman argued. Opening litigation doesn’t mean that all claims are necessarily viable, only that the process should work itself out in the courts of law, he said.
Eliminating section 230 could lead to online services being “over correct” in moderating speech which could lead to suffocating social reform movements organized on those platforms, argued Ashley Johnson of research institution, Information Technology and Innovation Foundation.
Furthermore, the burden of litigation would fall disproportionally on the companies that have fewer resources to defend themselves, she continued.
Bergman responded, “if a social media platform is facing a lot of lawsuits because there are a lot of kids who have been hurt through the negligent design of that platform, why is that a bad thing?” People who are injured have the right by law to seek redress against the entity that caused that injury, Bergman said.
Emma Llanso of the Center for Democracy and Technology suggested that platforms would change the way they fundamentally operate to avoid threat of litigation if section 230 were reformed or abolished, which could threaten freedom of speech for its users.
It is necessary for the protection of the first amendment that the internet consists of many platforms with different content moderation policies to ensure that all people have a voice, she said.
To this, Bergman argued that there is a distinction between algorithms that suggest content that users do not want to see – even that content that exists unbeknownst to the seeker of that information – and ensuring speech is not censored.
It is a question concerning the faulty design of a product and protecting speech, and courts are where this balancing act should take place, said Bergman.
This comes days after law professionals urged Congress to amend the statue to specify that it applies only to free speech, rather than the negligible design of product features that promote harmful speech. The discussion followed a Supreme Court decision to provide immunity to Google for recommending terrorist videos on its video platform YouTube.
To watch the full videos join the Broadband Breakfast Club below. We are currently offering a Free 30-Day Trial: No credit card required!
Free Speech
Creating Institutions for Resolving Content Moderation Disputes Out-of-Court
Private institutions may become primary method for content moderation disputes, says expert.

WASHINGTON, March 9, 2023 – A member of Meta’s oversight board, John Samples, suggested that out-of-court dispute institutions for content moderation may become the preferred method of settlement in Broadband Breakfast’s Big Tech & Speech Summit Thursday.
Meta’s oversight board was created by the company to support free speech by upholding or reversing Facebook’s content moderation decisions. It works independently of the company and hosts 40 members around the world.
The European Union’s Digital Services Act, which came into force in November of 2022, requires platforms to remove illegal content and ensure that users can contest removal of their content. It clarifies that platforms are only liable for users’ unlawful behavior if they are aware of it and fail to remove it.
The Act specifies illegal speech to include speech that does harm to the electoral system, hate speech, and speech that harms fundamental rights. The appeals process allows citizens to go directly to the company, the national courts, or out-of-court dispute resolution institutions, none of which currently exist in Europe.
According to Samples, the Act opens the way for private organizations like the oversight board to play a part in moderation disputes. “Meta has a tremendous advantage here as a first mover,” said Samples, “and the model of the oversight board may well spread to Europe and perhaps other places.”
The United States may adopt European processes in the future as it takes the lead in moderating big tech, claimed Samples. “It would largely be a private system,” he said, and could unify and centralize social media moderation across platforms and around the world.
The private option of self-regulation has worked well, said Samples. “It may well be expanding throughout much of the world. If it goes to Europe, it could go throughout.”
Currently, of the media that Meta reviews for moderation, only one percent is restricted, either by taking down the content or reducing the size of the audience exposed to it, said Samples. The oversight board primarily rules against Meta’s decisions and accepts comments from independent interests.
To watch the full videos join the Broadband Breakfast Club below. We are currently offering a Free 30-Day Trial: No credit card required!
-
Broadband Roundup3 weeks ago
AT&T Floats BEAD in USF Areas, Counties Concerned About FCC Map, Alabama’s $25M for Broadband
-
Big Tech3 weeks ago
Preview the Start of Broadband Breakfast’s Big Tech & Speech Summit
-
Big Tech4 weeks ago
House Innovation, Data, and Commerce Chairman Gus Bilirakis to Keynote Big Tech & Speech Summit
-
Big Tech3 weeks ago
Watch the Webinar of Big Tech & Speech Summit for $9 and Receive Our Breakfast Club Report
-
#broadbandlive2 weeks ago
Broadband Breakfast on March 22, 2023 – Robocalls, STIR/SHAKEN and the Future of Voice Telephony
-
Infrastructure1 week ago
BEAD Build Timelines in Jeopardy if ‘Buy America’ Waivers Not Granted, White House Budget Office Told
-
#broadbandlive3 weeks ago
Broadband Breakfast on March 8: A Status Update on Tribal Broadband
-
Infrastructure4 weeks ago
Nearly 80 Service Providers Engaged Equipment in Secure Networks Blacklist: FCC Report