Anonymous ID: 1d82bb Sept. 23, 2020, 10:43 a.m. No.10756841   🗄️.is 🔗kun   >>6877 >>7012 >>7170 >>7378 >>7518

>>10756805

https://www.justice.gov/ag/department-justice-s-review-section-230-communications-decency-act-1996

The proposal's

 

As part of the President's Executive Order on Preventing Online Censorship, and as a result of the Department's long standing review of Section 230, the Department has put together the following legislative package to reform Section 230. The proposal focuses on the two big areas of concern that were highlighted by victims, businesses, and other stakeholders in the conversations and meetings the Department held to discuss the issue. First, it addresses unclear and inconsistent moderation practices that limit speech and go beyond the text of the existing statute. Second, it addresses the proliferation of illicit and harmful content online that leaves victims without any civil recourse. Taken together, the Department's legislative package provides a clear path forward on modernizing Section 230 to encourage a safer and more open internet.

Anonymous ID: 1d82bb Sept. 23, 2020, 10:44 a.m. No.10756877   🗄️.is 🔗kun   >>6896 >>7012 >>7170 >>7378 >>7518

>>10756841

The Department identified four areas ripe for reform:

 

  1. Incentivizing Online Platforms to Address Illicit Content

The first category of potential reforms is aimed at incentivizing platforms to address the growing amount of illicit content online, while preserving the core of Section 230’s immunity for defamation.

 

a. Bad Samaritan Carve-Out. First, the Department proposes denying Section 230 immunity to truly bad actors. The title of Section 230’s immunity provision—“Protection for ‘Good Samaritan’ Blocking and Screening of Offensive Material”—makes clear that Section 230 immunity is meant to incentivize and protect responsible online platforms. It therefore makes little sense to immunize from civil liability an online platform that purposefully facilitates or solicits third-party content or activity that would violate federal criminal law.

 

b. Carve-Outs for Child Abuse, Terrorism, and Cyber-Stalking. Second, the Department proposes exempting from immunity specific categories of claims that address particularly egregious content, including (1) child exploitation and sexual abuse, (2) terrorism, and (3) cyber-stalking. These targeted carve-outs would halt the over-expansion of Section 230 immunity and enable victims to seek civil redress in causes of action far afield from the original purpose of the statute.

 

c. Case-Specific Carve-outs for Actual Knowledge or Court Judgments. Third, the Department supports reforms to make clear that Section 230 immunity does not apply in a specific case where a platform had actual knowledge or notice that the third party content at issue violated federal criminal law or where the platform was provided with a court judgment that content is unlawful in any respect.

 

  1. Clarifying Federal Government Enforcement Capabilities to Address

Anonymous ID: 1d82bb Sept. 23, 2020, 10:45 a.m. No.10756896   🗄️.is 🔗kun   >>6937 >>7012 >>7170 >>7378 >>7518

>>10756877

  1. Clarifying Federal Government Enforcement Capabilities to Address Unlawful Content

A second category reform would increase the ability of the government to protect citizens from harmful and illicit conduct. These reforms would make clear that the immunity provided by Section 230 does not apply to civil enforcement actions brought by the federal government. Civil enforcement by the federal government is an important complement to criminal prosecution.

 

  1. Promoting Competition

A third reform proposal is to clarify that federal antitrust claims are not covered by Section 230 immunity. Over time, the avenues for engaging in both online commerce and speech have concentrated in the hands of a few key players. It makes little sense to enable large online platforms (particularly dominant ones) to invoke Section 230 immunity in antitrust cases, where liability is based on harm to competition, not on third-party speech.

 

  1. Promoting Open Discourse and Greater Transparency

A fourth category of potential reforms is intended to clarify the text and original purpose of the statute in order to promote free and open discourse online and encourage greater transparency between platforms and users.

 

a. Replace Vague Terminology in (c)(2). First, the Department supports replacing the vague catch-all “otherwise objectionable” language in Section 230(c)(2) with “unlawful” and “promotes terrorism.” This reform would focus the broad blanket immunity for content moderation decisions on the core objective of Section 230—to reduce online content harmful to children—while limiting a platform's ability to remove content arbitrarily or in ways inconsistent with its terms or service simply by deeming it “objectionable.”

 

b. Provide Definition of Good Faith. Second, the Department proposes adding a statutory definition of “good faith,” which would limit immunity for content moderation decisions to those done in accordance with plain and particular terms of service and accompanied by a reasonable explanation, unless such notice would impede law enforcement or risk imminent harm to others. Clarifying the meaning of "good faith" should encourage platforms to be more transparent and accountable to their users, rather than hide behind blanket Section 230 protections.

 

c. Explicitly Overrule Stratton Oakmont to Avoid Moderator’s Dilemma. Third, the Department proposes clarifying that a platform’s removal of content pursuant to Section 230(c)(2) or consistent with its terms of service does not, on its own, render the platform a publisher or speaker for all other content on its service.

Anonymous ID: 1d82bb Sept. 23, 2020, 10:46 a.m. No.10756937   🗄️.is 🔗kun   >>7012 >>7170 >>7378 >>7518

>>10756896

The Department's review of Section 230 arose in the context of our broader review of market-leading online platforms and their practices, announced in July 2019. While competition has been a core part of the Department’s review, we also recognize that not all concerns raised about online platforms (including internet-based businesses and social media platforms) fall squarely within the U.S. antitrust laws. Our review has therefore looked broadly at other legal and policy frameworks applicable to online platforms. One key part of that legal landscape is Section 230, which provides immunity to online platforms from civil liability based on third-party content as well as immunity for removal of content in certain circumstances.

 

Drafted in the early years of internet commerce, Section 230 was enacted in response to a problem that incipient online platforms were facing. In the years leading up to Section 230, courts had held that an online platform that passively hosted third-party content was not liable as a publisher if any of that content was defamatory, but that a platform would be liable as a publisher for all its third-party content if it exercised discretion to remove any third-party material. Platforms therefore faced a dilemma: They could try to moderate third-party content but risk being held liable for any and all content posted by third parties, or choose not to moderate content to avoid liability but risk having their services overrun with obscene or unlawful content. Congress enacted Section 230 in part to resolve this quandary by providing immunity to online platforms both for third-party content on their services or for removal of certain categories of content. The statute was meant to nurture emerging internet businesses while also incentivizing them to regulate harmful online content.

 

The internet has changed dramatically in the 25 years since Section 230’s enactment in ways that no one, including the drafters of Section 230, could have predicted. Several online platforms have transformed into some of the nation’s largest and most valuable companies, and today’s online services bear little resemblance to the rudimentary offerings in 1996. Platforms no longer function as simple forums for posting third-party content, but instead use sophisticated algorithms to promote content and connect users. Platforms also now offer an ever-expanding array of services, playing an increasingly essential role in how Americans communicate, access media, engage in commerce, and generally carry on their everyday lives.

 

These developments have brought enormous benefits to society. But they have also had downsides. Criminals and other wrongdoers are increasingly turning to online platforms to engage in a host of unlawful activities, including child sexual exploitation, selling illicit drugs, cyberstalking, human trafficking, and terrorism. At the same time, courts have interpreted the scope of Section 230 immunity very broadly, diverging from its original purpose. This expansive statutory interpretation, combined with technological developments, has reduced the incentives of online platforms to address illicit activity on their services and, at the same time, left them free to moderate lawful content without transparency or accountability. The time has therefore come to realign the scope of Section 230 with the realities of the modern internet so that it continues to foster innovation and free speech but also provides stronger incentives for online platforms to address illicit material on their services.

 

Much of the modern debate over Section 230 has been at opposite ends of the spectrum. Many have called for an outright repeal of the statute in light of the changed technological landscape and growing online harms. Others, meanwhile, have insisted that Section 230 be left alone and claimed that any reform will crumble the tech industry. Based on our analysis and external engagement, the Department believes there is productive middle ground and has identified a set of measured, yet concrete proposals that address many of the concerns raised about Section 230.

 

A reassessment of America’s laws governing the internet could not be timelier. Citizens are relying on the internet more than ever for commerce, entertainment, education, employment, and public discourse. School closings in light of the COVID-19 pandemic mean that children are spending more time online, at times unsupervised, while more and more criminal activity is moving online. All of these factors make it imperative that we maintain the internet as an open and safe space.