Google non è responsabile per la presenza di app ad uso illecito nel suo play-store, stante il safe harbour ex 230 CDA

Un ex ambasciatore statunitense, di religione ebraica, chiede l’accetameonto di responsabilità di Google perchè permette la presenza sul PlayStore di un social (Telegram) notoriamente usato -anche- da estremisti autori di propaganda antisemita.

In particollare afferma che G. non fa rispetare la propria policy vincolante i creatori di app sullo  Store.

La corte californiana U.S. D C NORTHERN DISTRICT OF CALIFORNIA , SAN JOSE DIVISION, Case No. 21-cv-00570-BLF, Ginsberg c .Google, 18.02.2022, però ,accoglie l’eccezione di safe harbour ex 230 CDA sollevata da Google.

Dei tre requisiti chiesti allo scopo (che sia un service provider; che sia chiamato come Publisher; che si tratti di informazione di terzi), è il secondo quello di solito più litigato.

Ma giustamente la corte lo ravvisa anche in questo caso: <<In the present case, Plaintiffs’ claims are akin to the negligence claim that the Barnes court found to be barred by Section 230. Plaintiffs’ theory is that by creating and publishing guidelines for app developers, Google undertook to enforce those guidelines with due care, and can be liable for failing to do so with respect to Telegram. As in Barnes, however, the undertaking that Google allegedly failed to perform with due care was removing offending content from the Play Store.
But removing content is something publishers do, and to impose liability on the basis of such conduct necessarily involves treating the liable party as a publisher of the content it failed to remove. Barnes, 570 F.3d at 1103. Plaintiffs in the present case do not allege the existence of a contract or indeed any interaction between themselves and Google. Plaintiffs do not allege that Ambassador Ginsberg purchased his smartphone from Google or that he downloaded Telegram or any other app from the Play Store. Thus, the Barnes court’s rationale for finding that Section 230 did not bar Barnes’ promissory estoppel claim is not applicable here.
>>

(notizia a link alla sentenza dal blog del prof. Eric Goldman)

La piattaforma Omegle è titolata al safe harbour ex 230 CDA

la piattaforma Omegle permette comunciazioni on line tra sconosciuti anonimi e casualmente messi in connessione.  Pare abbia discreto successo.

Come immaginabile, si presta a gravi illeciti di sex trafficking coinvolgenti minori.

Nel caso specifico un tribunale della Florida ha però ritenuto che Omegle sia titolata al safe harbour di cui al § 230 CDA, ricorrendone i requisiti: così U.S. DISTRICT COURTMIDDLE DISTRICT OF FLORIDA-TAMPA DIVISION, 8:21-cv-814-VMC-TGW, del 10 gennaio 2022.

Si v. l’analitico ragionamento in proposito condotto sub C), p. 9 segg. , così riassunto alla fine: <<In short, the Court finds that Omegle is entitled to immunity under CDA Section 230 because (1) it is an ICS provider under the CDA, (2) Plaintiffs’ claims seek to treat Omegle as a publisher or speaker, and (3) the information at issue originated from another information provider, John Doe>>, p. 16.

Inoltre la  domanda di violazione del 18 U.S. Code § 1591 – Sex trafficking of children or by force, fraud, or coercion, è rigettata nel merito per carenza di elemento sogettivo (“knowingly”): <<the asserted claims against Omegle are premised upon general, constructive knowledge of past sex trafficking incidents. …  The generalized knowledge of past instances of sex trafficking are not enough to satisfy an exception to immunity.>>

(notizia e link alla sentenza dal blog di Eric Goldman)

Airbnb non perde il safe harbour per avere evidenziato tra i dati della casa in affitto un idromassaggio non a norma

L’Appello dell’Oregon (15.12.2021, n° 316 Or App 378 (2021), Smith c. Airbnb) ritiene che Airbnb , per aver evidenziato tra i dati del cliente la vasca idromassaggi, rivelatasi poi difettosa e fonte di danni alla persona, e per aver creato una categoria di ricerca ad hoc, non diventi CONTENT PROVIDER e dunque non gli precluda di invocare il safe harbour ex § 230 CDa

Cioè tali operazioni non hanno  augmented il contenuto, proveniente dal locatore, fonte di danno e dunque non gli fanno perdere il titolo al safe harbour: <<Airbnb’s provision of a “special search category”allowing users to search for and obtain results based onuser-provided information about hot tubs does not make Airbnb a content provider or developer. Merely providing that search tool did not augment—much less, materially contribute to—Dennis’s indication that the Old Barn hada hot tub.

Moreover, the search function was simply a neutral tool that provided listings of properties with hot tubsto Airbnb users in response to their queries. See Dart, 665F Supp 2d at 969 (“The word-search function is a ‘neutraltool.’”); Gonzalez, 2 F4th at 893 (“A website is not transformed into a content creator or developer by virtue of supplying ‘neutral tools’ that deliver content in response to userinputs.”); La Park La Brea A LLC, 285 F Supp 3d at 1104(Airbnb’s “auto-complete search function” does “not makeAirbnb an information content provider.”); cf. Carafano, 339F3d at 1124 (“[T]he fact that Matchmaker classifies usercharacteristics into discrete categories,” allowing “highlystructured searches,” did “not transform Matchmaker intoa developer of the underlying misinformation.”).

Likewise, by “highlighting” the user-created listings that mention hot tubs—including Dennis’s Old Barnlisting—Airbnb did not materially contribute to any portionof the hot tub information contained in Dennis’s listing and,thus, did not become a content provider. See Dowbenko v.Cite as 316 Or App 378 (2021) 389Google Inc., 582 Fed Appx 801, 805 (11th Cir 2014) (rejecting argument that Google was transformed into a contentprovider when it “manipulated its search results to prominently feature the article at issue”); Ascentive, LLC v. Op.Corp., 842 F Supp 2d 450, 476 (EDNY 2011) (“The fact thatthe defendants *** alter the way [users’] postings are displayed” and caused content to “appear higher in searchengine results list *** do[es] not render [defendant] aninformation content provider.”); cf. Dyroff, 934 F3d at 1096(The defendant’s website “is immune from liability underCDA because its functions, including recommendations andnotifications, were content-neutral tools used to facilitatecommunications.”)>>

La sentenza cita ampjamente il notissimo precedente Fair Hous. Council of San Fernando Valley v. Roommates.Com del 2008.

La sentenza pare esatta.

(notizia e link alla sentenza dal blog di Eric Goldman)

Safe harbour ex 230 CDA per Armslist, piattaforma per vendita on line di armi? Questione dubbia

Due corti statunitensi negano il safe harbour ex 230 CDA alla piattaforma di vendita di armi Armslist  , non trattandosi di azioni in cui son considerati editori/publisher/speaker .

Si trattava di responsabilità consguente ad uccisioni cagionate tramite armi da fuoco acquistate su Armslist: la quale sarebbe stata negligente nel permettere tale commercio incontrollato, avendo implementato  un software inadeguato alla base del proprio marketplace.

 Si tratta di due tribunali del Wisconsin, east. dis.:

1) BAUER and ESTATE OF PAUL BAUER v. ARMSLIST, del 19.11.2021, caso 20-cv-215-pp, sub V.B: <<The court does not mean to imply that §230(c) never can provide protection from liability for entities like Armslist. But that protection is not, as Armslist has argued, a broad grant of immunity. It is a fact-based inquiry. For example, the Seventh Circuit affirmed the district court’s grant of Craigslist’s motion for judgment on the pleadings in Chi. Lawyers’ Comm. The court recounted that “[a]lmost in passing, ” the plaintiff had alleged that Craiglist was liable for violations of the Fair Housing Act because although it had not created the discriminatory posts, it had “caused” the discriminatory third-party posts to be made. Chi. Lawyers’ Comm., 519 F.3d at 671. Emphasizing that Craigslist was not the author of the discriminatory posts, the Seventh Circuit found that the only causal connection between Craigslist and the discriminatory posts was the fact that “no one could post a discriminatory ad if craiglist did not offer a forum.” Id. The court stated that “[n]othing in the service craigslist offers induces anyone to post any particular listing or express a preference for discrimination; for example, craigslist does not offer a lower price to people who include discriminatory statements in their postings.” Id. at 671-72. For that reason, the court concluded that “given § 230(c)(1) [the plaintiff] cannot sue the messenger just because the message reveals a third party’s plan to engage in unlawful discrimination.” Id. at 672.

The plaintiffs in this case have not raised claims of defamation or obscenity or copyright infringement—the types of claims that would require the court to determine whether Armslist is a “publisher” or “speaker” of content, rather than a provider of an interactive computer service that hosts content created by third parties. None of the nine claims in the second amended complaint challenge the content of ads posted on the Armslist.com website—not even Caldwell’s ad. The plaintiffs have alleged that Armslist should have structured the website differently—should have included safeguards and screening/monitoring provisions, should have been aware of the activity of individuals like Caldwell, should have implemented measures that would prevent illegal firearms dealers from using the website to sell guns without a license.

In declining to dismiss the complaint on §230(c) grounds, the court in Webber v. Armslist recently stated that because the plaintiff in that case had alleged “negligence and public nuisance based on Defendants’ affirmative conduct, ” it appeared that “§ 230 is not even relevant to this case.” Webber v. Armslist, No. 20-cv-1526, 2021 WL 5206580, at *6 (E.D. Wis. Nov. 9, 2021). This court agrees. Section 230 does not immunize Armslist from suit and the court will not dismiss the complaint on that basis.>>

2) Webber v. Armslist, del 9 novembre 2021, caso 20-C-1526, più dettagliata sul punto: <<But even if § 230 applies to this type of case, Plaintiff’s claims do not seek to treat Defendants as the “publisher or speaker” of the post in question. Here, Plaintiff seeks to hold Defendants liable for their “role in developing or co-developing [their] own content.” Dkt. No. 13 at 18. Specifically, Plaintiff faults Defendants for failing to prohibit criminals from accessing or buying firearms through Armslist.com; actively encouraging, assisting, and facilitating illegal firearms transactions through their various design decisions; failing to require greater details from users, such as providing credit-card verified evidence of users’ identities; failing to require that sellers certify under oath that they are legal purchasers; and failing to provide regularly updated information regarding applicable firearms laws to its users, among many other things. Compl. at ¶ 165. In essence, the complaint “focuses primarily on Armslist’s own conduct in creating the high-risk gun market and its dangerous features, ” not on the post in question. Dkt. No. 13 at 23. This type of claim, then, does not seek to treat Defendants as the “publisher or speaker” of the post that led to Schmidt’s killer obtaining a firearm; rather, it seeks to hold Defendants liable for their own misconduct in negligently and recklessly creating a service that facilitates the illegal sale of firearms. 47 U.S.C. § 230(c)(1). For these reasons, the Court concludes that § 230 does not immunize Defendants from liability in this case>>.

Viene però osservato dal prof. Eric Goldman (da cui ho tratto notizia e link alle sentenze),  che  la corte suprema del Wisconsin nel 2019 in Daniel v. Armslist aveva invece concesso il safe harbour

Sempre il prof. Goldman con post 16 luglio 2023 ora ci aggiorna su altra decisione relAtiva ad Armslist (Webber v. Armslist)

Ancora su responsabilità per violazione di privacy/right of publicity etc. da parte dei redattori di annuari (yearbooks) e safe harbour ex § 230.

Ancora sul tema in oggetto e ancora (giusto) rigetto dell’eccezine di safe harbour ex § 230 CDA.

Si tratta di filone giudiziario di una certa consistenza , di cui avevo dato conto qui e qui.

1) BONILLA v. ANCESTRY.COM OPERATIONS INC. de 7 dic. 2021 No. 20-C-07390, Colrte ND Illinois, eastern division : <<Ancestry argues that each of Plaintiff’s claims should be dismissed because Ancestry is immune from liability under the Communications Decency Act (“CDA”). The CDA provides that “[n]o provider or user of an interactive computer service shall be treated as the publisher or speaker of any information provided by another information content provider.” 47 U.S.C. § 230(c)(1). The CDA applies to online forums that serve as “a mere passive conduit for disseminating (actionable) statements.” Huon v. Denton, 841 F.3d 733, 742 (7th Cir. 2016). Because affirmative defenses such as CDA immunity frequently turn on facts not before the court at the pleading stage, dismissal is appropriate only when the factual allegations in the complaint unambiguously establish all the elements of the defense. See Siegel v. Zoominfo Techs., LLC, 2021 WL 4306148, at *4 (N.D. Ill. Sept. 22, 2021) citing Hyson USA, Inc. v. Hyson 2U, Ltd., 821 F.3d 935, 939 (7th Cir. 2016)

Here, Plaintiff’s complaint includes allegations that Ancestry created records of Plaintiffs reflecting the information from the yearbook record, and uses that information to entice potential customers to subscribe to its services. Plaintiff also alleges that non-users are shown a limited version of the record (including a low resolution photograph) with a promotional pop-up advertisement promising access to Plaintiff’s identity and likeness if they sign up for a paid subscription. Dkt. 29 at ¶¶38-46. Ancestry recasts Plaintiff’s claims on the “mere reposting” of yearbook records, which is inconsistent with Plaintiff’s allegations that Ancestry uses those records and the likeness/identity to sell subscription services beyond his individual yearbook record.   These allegations, taken as true, do not establish that Ancestry is a “passive conduit” that should receive immunity under the CDA. Plaintiff has alleged that Ancestry collected and organized records and subsequently used Plaintiff’s and the putative class members’ names, likenesses, and identities in these records they curated for commercial gain. See Krause v. Rocketreach, LLC, 2021 WL 4282700 (N.D. Ill. Sept. 21, 2021) (rejecting CDA immunity argument where complaint allegations did not establish affirmative defense); see also Lukis v. Whitepages, Inc., 454 F.Supp.3d 746, 763 (N.D. Ill. 2020)>>.

2) CALLAHAN v. PEOPLECONNECT, INC. Case No. 20-cv-09203-EMC.  del 1.11.2021 , Tribunale del N.D. California. , molto più dettagliata , che valorizza la notevole differenza tra l’annuario in cartaceo a diffusione limitata e la sua messa on line: <<in the instant case, the Court concludes that, at the very least, there is a question of fact as to whether a reasonable person in the position of PeopleConnect (the service provider) would conclude that the yearbook authors/publishers (the information content providers) intended the yearbooks to be published on the internet. As Plaintiffs point out, the yearbooks at issue were published in the 1990s and early 2000s when “[t]he Internet was in its infancy and social media did not exist.” Opp’n at 5. Moreover, there is a difference between publishing a yearbook for a school or local community and publishing a yearbook on the internet where the audience is far broader. Thus, it would be hard to conclude that, as a matter of law, PeopleConnect is a publisher of information provided by another information content provider and is thus entitled to immunity under the CDA.

PeopleConnect’s reliance on Judge Beeler’s Ancestry decisions is unavailing. In Ancestry II, Judge Beeler indicated that, under Batzel, it was reasonable for Ancestry to believe that the yearbooks at issue were being provided to it for publication on the Internet, but Judge Beeler’s ruling appears to turn on her view that an information content provider could be people or entities other than the yearbook author/publisher. See Ancestry II, 2021 U.S. Dist. LEXIS 112036, at *17-18 (stating that, “whether the yearbooks were donated by other former students or obtained from other sources, Ancestry is demonstrably not the content creator and instead is publishing third-party content provided to it for publication”; “[n]othing in Batzel requires the original creator’s permission for publication”). But that view is not consistent with the express definition of “information content provider” under the CDA; an information content provider is one who created or developed the information at issue. In the instant case, the yearbook authors/publishers are the only ones who meet that criteria.

At the hearing, PeopleConnect suggested that a service provider should be allowed to assume that the person or entity who provided the information to the service provider was the creator or developer of the information.6 Such an approach, however, would be contrary to Batzel which focuses on the reasonable perception of the service provider. PeopleConnect fails to explain why a service provider should not be held accountable if, e.g., it is obvious that the person or entity providing information to the service provider is not the creator or developer of the information. In such a situation, if it is obvious that the person or entity providing the information is not the creator or developer of the information, then the service provider “is the one making the affirmative decision to publish, and so . . . contributes materially to [the] allegedly unlawful dissemination” of the information[;] [it] is thus properly deemed a developer and not entitled to CDA immunity.” Roommates.com, 521 F.3d at 1171.

In the instant case, it is obvious that the yearbook users/purchasers were not the creators or developers of the yearbooks. Instead, the yearbook authors/publishers were the content providers. PeopleConnect cannot claim the benefit of CDA immunity, absent a reasonable basis to believe that the yearbook authors/publishers intended for there to be publication on the Internet. This presents a question of fact that cannot be resolved at the 12(b)(6) phase of proceedings>>.

Ancora sulla (al momento impossibile da ottenere) qualificazione delle piattaforme social come State Actors ai fini del Primo Emendamento (libertà di parola)

Altra sentenza (d’appello stavolta) che rigetta la domadna vs. Facebook (rectius, Meta) basata sul fatto che illegalmente filtrerebbe/censurerebbe i post o rimuoverebbe gli account , violando il Primo Emendamento (libertà di parola).

Questo diritto spetta solo verso lo Stato o verso chi agisce in suo nome o assieme ad esso.

Si tratta della sentenza di appello del 9° circuito (su impugnazione di una sentenza californiana confermata) ,  emessa il 22.11.2021, No. 20-17489 , D.C. No. 3:20-cv-05546-RS, Atkinson c. Meta-Zuckerberg.

Sono riproposte dall’utente (e la Corte partitamente rigetta) tutte le consuete e note causae petendi in tema.   Nulla di nuovo ma un utile loro ripasso.

Inoltre la Corte conferma pure l’applicazione del safe harbour ex  230 CDA.

(notizia e link alla sentenza dal blog di Eric Goldman)

Interessante sentenza dagli USA sulla chiusura immotivata da parte di Facebook dell’account di un’utente

Si tratta della corte del nord california 12 noiv. 2021, 21cv04573EMC , King v-. Facebbok (dal blog di Eric Goldman).

Il provveidmento interessa, dato che la chiusura immotivata di account FB pare non sia così rara.

L’attrice avanza varie domande (una basata sul § 230.c.2.A CDA : incomprensibile, visto che , la disposizione esime da responsabilità anzichè comminarla!, p. 4 segg.)

Qui ricordo la domanda sub E, p. 10 ss basata sulla violazione contrattuale ex fide bona e correttezza.

Rigettata quella sulla distruzione di contenuto (sub 1: non condivisibelmente però: se manca obbligo specifico per F. di conservare, quanto meno la buona fede impone di dare congruo preavviso della prossima distruzione), viene accolta quella sulla mancanza di motivazione,. sub 2, p. 12 ss

F. si basa sulla pattuita clausola <<If we determine that you have clearly, seriously or repeatedly breached our Terms or Policies, including in particular our Community Standards, we may suspend or permanently disable access to your account.>> per affermare che aveva piena discrezionalità

Il giudice ha buon gioco però nel dire che non è così: <<Notably, the Terms of Service did not include language providing that Facebook had “sole discretion” to act.  Compare, e.g., Chen v. PayPal, Inc., 61 Cal. App. 5th 559, 570-71 (2021) (noting that contract provisions allowed “PayPal to place a hold on a payment or on a certain amount in a seller’s account when it ‘believes there may be a high level of risk’ associated with a transaction or the account[,] [a]nd per the express terms of the contract, it may do so ‘at its sole discretion’”; although plaintiffs alleged that “‘there was never any high level of risk associated with any of the accounts of any’ appellants, . . . this ignores that the user agreement makes the decision to place a hold PayPal’s decision – and PayPal’s alone”). 

Moreover, by providing a standard by which to evaluate whether an account should be disabled, the Terms of Service suggest that Facebook’s discretion to disable an account is to be guided by the articulated factors and cannot be entirely arbitrary.  Cf. Block v. Cmty. Nutrition Ins., 467 U.S. 340, 349, 351 (1984) (stating that the “presumption favoring judicial review of administrative action . . . may be overcome by specific language or specific legislative history that is a reliable indicator of congressional intent” – i.e., “whenever the congressional intent to preclude judicial review is ‘fairly discernible in the statutory scheme’”). 

At the very least, there is a strong argument that the implied covenant of good faith and fair dealing imposes ome limitation on the exercise of discretion so as to not entirely eviscerate users’ rights>>

Inoltre (sub 3, p. 14) quanto meot una spiegazione era dovuta. (i passaggi sub 2 e il 3 si sovrappontgono)

In breve sono ritenute illegittime la disbilitgazione e la mancanza di motivazione (che si soprappongono, come appena detto: la reciproca distinzione concettuale richiederebbe troppo spazio e tempo)

Da ultimo, l’ovvia eccezione di safe harbour ex § 230.c.1 CDA <Treatment of publisher or speaker> copre la disabilitazione ma non la mancata spiegazione (p. 22).

Sul secondo punto c’è poco da discutere: il giudice ha ragione.

Più difficile rispondere sul primo,  importante nella pratica, dato che qualunque disabilitazione costituirà -dal punto del disabilitato- una violazione di contratto.

Il giudice dà ragione a F.: il fatto che esista un patto, non toglie a F. il safe harbour : <<although Ms. King’s position is not without any merit, she has glossed over the nature of the “promise” that Facebook made in its Terms of Service. In the Terms of Service, Facebook simply stated that it would use its discretion to determine whether an account should be disabled based on certain standards. The Court is not convinced that Facebook’s statement that it would exercise its publishing discretion constitutes a waiver of the CDA immunity based on publishing discretion. In other words, all that Facebook did here was to incorporate into the contract (the Terms of Service) its right to act as a publisher. This by itself is not enough to take Facebook outside of the protection the CDA gives to “‘paradigmatic editorial decisions not to publish particular content.’” Murphy, 60 Cal. App. 5th at 29. Unlike the very specific one-time promise made in Barnes, the promise relied upon here is indistinguishable from “‘paradigmatic editorial decisions not to publish particular content.’” Id. It makes little sense from the perspective of policy underpinning the CDA to strip Facebook of otherwise applicable CDA immunity simply because Facebook stated its discretion as a publisher in its Terms of Service>>.

Decisione forse esatta sul punto specifico, ma servirebbe analisi ulteriore.

Quattro causae petendi relative al First Amendment/libertà di parola per contrastare la sospensione dell’account Youtube, ma nessuna accolta

Interessante sentenza californiana sulla solita questione della libertà di parola  (First Amendement)  asseritamente violata da sospensione dell’account su social media (politicamente di destra) da parte di una state action.

Si tratta della corte distrettuale di S. Josè, Californa, 19 ottobre 2021, Case No. 20-cv-07502-BLF, Doe c. Google,.

Sono azionate quattro causae petendi, tutte rigettate visto che nessuna è applicabile alla censura/content moderation di Youtube:

1) Public function: curiosamente l’attore e la corte invocano in senso reciprocamente opposto il  noto precedente Prager Univ. c. Google del 2020.

2) Compulsion: <<Rep. Adam Schiff and Speaker of the House Nancy Pelosi and an October 2020 House Resolution, which “have pressed Big Tech” into censoring political speech with threats of limiting Section 230 of the Communications Decency Act (“CDA”) and other penalties.>>. Alquanto inverosimile (è però la più lungamente argometnata)

3) joint action: <<Joint action is present where the government has “so far insinuated itself into a position of interdependence with [a private entity] that it must be recognized as a joint participant in the challenged activity.” Gorenc v. Salt River Project Agr. Imp. and Power Dist., 869 F.2d 503, 507 (9th Cir. 1989) (quoting Burton v. Wilmington Parking Authority, 365 U.S. 715, 725 (1961)). Further, a private defendant must be a “willful participant in joint action with the state or its agents.” Dennis v. Sparks, 449 U.S. 24, 27 (1980). Joint action requires a “substantial degree of cooperative action” between private and public actors. Collins v. Womancare, 878 F.2d 1145, 1154 (9th Cir. 1989).>.

Per gli attori la  joint action theory starebbe in un  <<Twitter exchange between Rep. Schiff and YouTube CEO Susan Wojnicki in which Ms. Wojnicki states, “We appreciate your partnership and will continue to consult with Members of Congress as we address the evolving issues around #COVID19.” FAC, Ex. E at 1; Opp. at 10-15. Plaintiffs argue that this Twitter exchange shows Defendants and the federal government were in an “admitted partnership.”>>. Allegazione un pò leggerina.

4) Governmental nexus: ricorre quando c’è << “such a close nexus between the State and the challenged action that the seemingly private behavior may be fairly treated as that of the State itself.” Kirtley v. Rainey, 326 F.3d 1088, 1094-95 (9th Cir. 2003). “The purpose of this requirement is to assure that constitutional standards are invoked only when it can be said that the State is responsible for the specific conduct of which plaintiff complains.” Blum, 457 U.S. at 1004-1005>>. (sembra assai simile alla prcedente).

Non avendo accolto alcuna di quesrta, non affronta il safe harbour ex 230 CDA, p. 12. Curioso l’rodine logico : il criterio della ragine più liquidqa avrebbe potuto a rigttare (nel merito) con tale norma.

(sentenza e link dal blog di Eric Goldman)

Corresponsabilità delle puiattaforme digitali per la strage di Orlando (Florida, USA) del 2016? No

Nella strage di Orlando USA del 2016 Omar Mateen uccise 49 persone e ne ferì 53 con un fucile semiautomatico, inneggiando all’ISIS.

Le vittime proposero domanda giudiziale contro Twitter Google e Facebook sia in base Anti-Terrorism Act, 18 U.S.C. §§ 2333(a) & (d)(2) (è respponsabile chi , by facilitating his access to radical jihadist and ISIS-sponsored content in the months and years leading up to the shooting) sia per legge statale, avendo cagionato  negligent infliction of emotional distress and wrongful death.

La cit. legge ATA imposes civil liability on “any person who aids and abets, by knowingly providing substantial assistance, or who conspires with the person who committed . . . an act of international terrorism,” provided that the “act of international terrorism” is “committed, planned, or authorized” by a designated “foreign terrorist organization.

Nega ogni responsabilità in capo alle piattaforme (confermando la sentenza di primo grado della Florida) la corte d’appello dell’11° circuito 27.09.2021, No. 20-11283 , Colon ed altri c. Twitter-Facebook-Google.

La prima domanda è respinta sia perchè non si trattò di terrorismo internazionale (pur se reclamato dal’lISIS), come richiede la cit legge, sia perchè non fu una foreign terroristic organization a commetterlo (ma un c.d. lupo solitario).

Ma soprattutto è rigettata la seconda domanda (negligenza nel causare danni e decdessi) : gli attori non hanno superato la prova della proximate causation circa il ruolo delle puiattaforme, sub IV.A, p. 21 ss

La corte parla si del nesso di causalità ma in astratto e in base ai precedenti, senza applicarlo al ruolo delle piattaforme nella commissione di delitti.

La corte stranamente non menziona il safe harbour ex 230 CDA che avrebbe potuto essere invocato (cosa che quasi certanente le piattafirme avranno fatto)

(notizia e link dal blog di Eric Goldman)