Supported by
the Luxembourg National Research Fund
Project O19/13946847
ADR |
Alternative Dispute Resolution |
Art |
Article/Articles |
BGB |
Bürgerliches Gesetzbuch (Civil Code) [Germany] |
BGH |
Bundesgerichtshof (Federal Court of Justice) [Germany] |
BVerfG |
Bundesverfassungsgericht (Federal Constitutional Court) [Germany] |
cf |
confer (compare) |
ch |
chapter |
CJEU |
Court of Justice of the European Union |
ECLI |
European Case Law Identifier |
ECtHR |
European Court of Human Rights |
DMCA |
Digital Millennium Copyright Act [USA] |
DSA or DSA Regulation |
Digital Services Act [European Union] |
ECG |
E-Commerce-Gesetz (E-Commerce Act) [Austria] |
ed |
editor/editors |
edn |
edition/editions |
eg |
exempli gratia (for example) |
etc |
et cetera |
EU |
European Union |
EUR |
Euro |
f/ff |
following |
fn |
footnote (external, ie, in other chapters or in citations) |
GSR |
General Secondary Response |
GG |
Grundgesetz (Fundamental Law) [Germany] |
GDPR |
General Data Protection Regulation [European Union] |
GTC |
General terms and conditions |
ibid |
ibidem (in the same place) |
ie |
id est (that is) |
KoPlG |
Kommunikationsplattformengesetz (Communications Platforms Act) [Austria] |
n |
footnote (internal, ie, within the same chapter) |
NetzDG |
Netzwerkdurchsetzungsgesetz (Network Enforcement Act) [Germany] |
no |
number/numbers |
öABGB |
Allgemeines Bürgerliches Gesetzbuch (General Civil Code) [Austria] |
öUrhG |
Urheberrechtsgesetz (Copyright Act) [Austria] |
P2B Regulation |
Regulation (EU) 2019/1150 of 20 June 2019 on promoting fairness and transparency for business users of online intermediary services |
PACT or PACT Act |
Platform Accountability and Consumer Transparency Act [USA] |
para |
paragraph/paragraphs |
pt |
part |
RDG |
Rechtsdienstleistungsgesetz (Legal Services Act) [Germany] |
S.D. Cal. |
District Court for the Southern District of California [USA] |
sec |
Section/Sections |
supp |
supplement/supplements |
TMG |
Telemediengesetz (Telemedia Act) [Germany] |
trans/tr |
translated, translation/translator |
UGC |
User-generated content |
UK |
United Kingdom |
UNIDROIT |
Institut international pour l'unification du droit privé (International Institute for the Unification of Private Law) |
UrhDaG |
Urheberrechts-Diensteanbieter-Gesetz (Act on Copyright Content Sharing Service Providers) [Germany] |
UrhG |
Urheberrechtsgesetz (Copyright Act) [Germany] |
US/USA |
United States of America |
U.S.C. |
United States Code |
USD |
United States Dollar |
v |
versus |
vol |
volume/volumes |
VSBG |
Verbraucherstreitbeilegungsgesetz (Consumer Dispute Resolution Act) [Germany] |
ZPO |
Zivilprozessordnung (Code of Civil Procedure) [Germany] |
*** |
*** |
Charter of Fundamental Rights of the European Union, OJ C 364, 18 December 2000, 1.
Comission and Parliament (EU), Synopsis AI Act, P9_TA(2023)0236, https://www.europarl.europa.eu/doceo/document/TA-9-2023-0236_EN.pdf accessed 31 December 2023.
Commission Recommendation (EU) 2018/334 of 1 March 2018 on measures to effectively tackle illegal content online, OJ L 62, 6 March 2018, 50.
Directive 2000/31/EC of 8 June 2000 on certain legal aspects of information society services, in particular electronic commerce, in the Internal Market (‘Directive on electronic commerce’), OJ L 178, 17 July 2000, 1.
Directive 2001/29/EC of 22 May 2001 on the harmonization of certain aspects of copyright and related rights in the information society, OJ L 167, 22 June 2001, 10.
Directive 2004/48/EC of 29 April 2004 on the enforcement of intellectual property rights, OJ L 157, 30 April 2004, 45.
Directive 2008/52/EC of the European Parliament and of the Council of 21 May 2008 on certain aspects of mediation in civil and commercial matters, OJ L 136, 24 May 2008, 3.
Directive 2013/11/EU of 21 May 2013 on alternative dispute resolution for consumer disputes and amending Regulation (EC) No 2006/2004 and Directive 2009/22/EC (Directive on consumer ADR), OJ L 165, 18 June 2013, 63.
Directive 2019/790/EU of 17 April 2019 on copyright and related rights in the Digital Single Market and amending Directives 96/9/EC and 2001/29/EC, OJ L 130, 17 May 2019, 92.
ICC Arbitration Rules 2021.
Regulation (EU) 2019/1150 of 20 June 2019 on promoting fairness and transparency for business users of online intermediation services, OJ L 186, 11 July 2019, 57.
Regulation (EU) 2021/784 of 29 April 2021 on addressing the dissemination of terrorist content online, OJ L 172, 17 May 2021, 79.
Regulation (EU) 2022/2065 of 19 October 2022 on a Single Market For Digital Services and amending Directive 2000/31/EC (Digital Services Act; hereinafter: DSA Regulation), OJ L 277, 27 October 2022, 1.
Regulation (EU) 2024/1689 of 13 June 2024 laying down harmonised rules on artificial intelligence and amending Regulations (EC) No 300/2008, (EU) No 167/2013, (EU) No 168/2013, (EU) 2018/858, (EU) 2018/1139 and (EU) 2019/2144 and Directives 2014/90/EU, (EU) 2016/797 and (EU) 2020/1828 (Artificial Intelligence Act), OJ L, 12 July 2024.
UNCITRAL Model Law on International Commercial Arbitration (1985).
***
Arbitration Rules of the Netherlands Arbitration Institute (Schiedsordnung Nederlands Arbitrage Instituut).
Austrian Communications Platforms Act (Kommunikationsplattformengesetz: KoPlG).
Austrian Copyright Act (Urheberrechtsgesetz: öUrhG).
Austrian General Civil Code (Allgemeines Bürgerliches Gesetzbuch: öABGB).
Austrian E-Commerce Act (E-Commerce-Gesetz: ECG).
English Arbitration Act 1996.
French Code of Civil Procedure (Code de procédure civile).
French Law of 23 March 2019 (Loi de programmation 2018-2022 et réforme pour la justice) No 2019/222.
German Act on Copyright Content Sharing Service Providers (Urheberrechts-Diensteanbieter-Gesetz: UrhDaG).
German Fundamental Law (Grundgesetz: GG).
German Civil Code (Bürgerliches Gesetzbuch: BGB).
German Code of Civil Procedure (Zivilprozessordnung: ZPO).
German Consumer Dispute Resolution Act (Verbraucherstreitbeilegungsgesetz: VSBG).
German Copyright Act (Urheberrechtsgesetz: UrhG).
German Legal Services Act (Rechtsdienstleistungsgesetz: RDG).
German Network Enforcement Act (Netzwerkdurchsetzungsgesetz: NetzDG).
German Telemedia Act (Telemediengesetz: TMG).
Italian Code of Civil Procedure (Codice di procedura civile: CPC).
Spanish Arbitration Act (Lex 60/2003 de 23 de diciembre, de Arbitraje).
Title 17 of the United States Code – Copyright Law (17 U.S.C.).
Title 47 of the United States Code – Telecommunications (47 U.S.C.).
Turkish International Arbitration Law (Uluslararası Tahkim Kanunu).
United States of America Platform Accountability and Consumer Transparency Act, 116th Cong. § 5(2) (2020) (PACT Act).
United States of America Digital Millennium Copyright Act (DMCA).
***
CJEU, 23 March 2010, C-236/08 and C-238/08 – Google France, ECLI:EU:C:2010:159.
CJEU, 12 July 2011, C‑324/09 – L’Oréal v eBay, ECLI:EU:C:2011:474.
CJEU (Grand Chamber), 24 September 2019, C-507/17 – Google (Spatial scope of delisting), ECLI:EU:C:2019:772.
CJEU, 3 October 2019, C-18/18 – Glawischnig-Piesczek, ECLI:EU:C:2019:821.
CJEU, 26 April 2022, C‑401/19 – Poland v Parliament and Council, ECLI:EU:C:2022:297.
European General Court, 27 September 2023, T-367/23 – Amazon Services Europe v Commission, ECLI:EU:T:2023:589.
Oversight Board, 28 January 2021, 2020-005-FB-UA – Nazi Quote.
Oversight Board, 5 May 2021, 2021-001-FB-FBR – Former President Trump's suspension.
Oversight Board, 8 July 2021, 2021-006-IG-UA – Ocalan's Isolation.
Oversight Board, 14 September 2021, 2021-009-FB-UA – Shared Al Jazeera Post.
Oversight Board, 27 September 2021, 2021-010-FB-UA – Colombia Protests.
Oversight Board, 17 June 2022, 2022-001-FB-UA – Knin Cartoon.
Oversight Board, 15 September 2022, 2022-005-FB-UA – Mention of the Taliban in News Reporting.
Oversight Board, 22 November 2022, 2022-007-IG-MR – UK Drill Music.
Oversight Board, 6 December 2022, 2021-002-FB-PAO – Meta's Cross-Check Program.
Oversight Board, 14 December 2022, 2022-012-IG-MR – India Sexual Harassment Video.
Oversight Board, 14 December 2022, 2022-011-IG-UA – Video after Nigeria Church Attack.
Oversight Board, 9 January 2023, 2022-013-FB-UA – Iran Protest Slogan.
Oversight Board, 9 March 2023, 2022-014-FB-MR – Sri Lanka Pharmaceuticals.
Oversight Board, 18 December 2023, 2023-054-FB-UA, 2023-055-FB-UA, 2023-056-FB-UA, 2023-057-FB-UA – Goebbels Quote.
***
BGH (Germany), 17 August 2011, I ZR 57/09 – Stiftparfüm, BGHZ 191, 19 = (2011) 113(11) GRUR (Gewerblicher Rechtsschutz und Urheberrecht) 1038.
BGH (Germany), 25 October 2011, VI ZR 93/10 – Blog Eintrag, BGHZ 191, 219 = (2012) 65(3) NJW (Neue Juristische Wochenschrift) 148 = (2012) 114(3) GRUR (Gewerblicher Rechtsschutz und Urheberrecht) GRUR 2012, 311.
BGH (Germany), 27 October 2011, I ZR 131/10 – regierung-oberfranken.de, (2012) 65(31) NJW (Neue Juristische Wochenschrift) 2279.
BGH (Germany), 12 July 12, 2012, I ZR 18/11 – Alone in the dark, BGHZ 194, 339 = (2013) 66(11) NJW (Neue Juristische Wochenschrift) 784.
BGH (Germany), 18 June 2015, I ZR 74/1 – Liability for Hyperlink, BGHZ 206, 103 = (2016) 69(11) NJW (Neue Juristische Wochenschrift) 804.
BGH (Germany), 1 March, 2016, VI ZR 34/15 – Ärztebewertungsportal III (jameda.de), BGHZ 209, 139 = (2016) 65(3) NJW (Neue Juristische Wochenschrift) 2106.
BGH (Germany), 22 November 2017, VIII ZR 83/16, (2018) 71(8) NJW (Neue Juristische Wochenschrift) 537.
BGH (Germany), 22 November 2017, VIII ZR 213/16, (2018) 21(3) MMR (Multimedia und Recht) 156.
BGH (Germany), 29 July 2021, III ZR 179/20, (2021) 74(43) NJW (Neue Juristische Wochenschrift) 3179.
BGH (Germany), 29 July 2021, III ZR 192/20, (2021) 25(11) ZUM-RD (Zeitschrift für Urheber- und Medienrecht – Rechtsprechungsdienst) 612.
BVerfG (Germany), 3 February 1959, 2 BvL 10/56, BVerfGE 9, 137 = (1959) 12(21) NJW (Neue Juristische Wochenschrift) 931.
BVerfG (Germany), 8 August 1978, 2 BvL 8/77, BVerfGE 49, 89 = (1979) 32(8) NJW (Neue Juristische Wochenschrift) 359.
BVerfG (Germany), 20 April 1982, 2 BvL 26/81, BVerfGE 60, 253 = (1982) 35(43) NJW (Neue Juristische Wochenschrift) 1982, 2425.
District Court for the Southern District of California (USA), United States v Green, 857 F. Supp. 2d 1015, 1018 (S.D. Cal. 2012).
***
Achleitner R, ‘The Fight against Geo-Blocking – A Never Ending Story? Policy Paper on Geo-Blocking’ https://ssrn.com/abstract=4246896 or http://dx.doi.org/10.2139/ssrn.4246896 accessed 31 December 2023.
Adolphsen J, ‘Der Zivilprozess im Wettbewerb der Methoden’ (2017) 48(4) BRAK Mitteilungen 147.
Althammer C, ‘Alternative Streitbeilegung im Internet’ in F Faust and H-B Schäfer (ed), Zivilrechtliche und rechtsökonomische Probleme des Internet und der künstlichen Intelligenz (Mohr Siebeck 2019) 249.
Ameln F von, ‘Führen und Entscheiden unter Unsicherheit‘ (2021) 52(4) GIO (Gruppe. Interaktion. Organisation. Zeitschrift für Angewandte Organisationspsychologie) 567.
Andrea R, ‘No Safe Harbor: YouTube’s Content ID and Fair Use’ (2020) Boston College Intellectual Property & Technology Forum 1.
Anzinger H M, ‘10 Jahre Modria – KMS und Online-Mediation auf dem Weg zur Digitalisierung der Justiz – Teil 1‘ (2021) 24(2) ZKM (Zeitschrift für Konfliktmanagement) 53.
Anzinger H M, ‘10 Jahre Modria – KMS und Online-Mediation auf dem Weg zur Digitalisierung der Justiz – Teil 2’ (2021) 24(3) ZKM (Zeitschrift für Konfliktmanagement) 84.
AIRBNB, ‘Scoring the user to prevent “suspicious” activity before it occurs: What Does It Mean When Someone’s ID Has Been Checked?’, https://www.airbnb.com/help/article/2356/what-does-it-mean-when-someones-id-has-been-checked accessed 31 December 2023.
Appelman N and Leerssen P, ‘On “Trusted” Flaggers’ (2022) 24 Yale Journal of Law & Technology 452.
Arbitrator Intelligence, ‘Arbitrator Intelligence Database’, https://arbitratorintelligence.vercel.app/ accessed 31 December 2023.
Aschauer C, ‘Automated Decision-Making and Artificial Intelligence (AI) in Arbitration’ in C Leyens, I Eisenberger and R Niemann (ed), Smart Regulation (Mohr Siebeck 2021) 130.
Askani A, Private Rechtsdurchsetzung bei Urheberrechtsverletzungen im Internet (Nomos 2021).
Bambauer D E, ‘Against Jawboning’ (2015) 100(1) MINN. L. REV. 51.
Barocas S and Self A D, ‘Big Data’s Disparate Impact’ (2016) 104(3) California Law Review 671 (2016).
Barona Vilar S, ‘Effizienszsteigerung und Suche nach Beschleunigung von Schiedsverfahren im Spannungsfeld von Mythos. Sublimierung und Vierter Industrieller Revolution (4.0)’ (2018) 23 ZZPInt (Zeitschrift für Zivilprozess International) 295.
Barton H, ‘Rebooting Justice’ (2018) 44(4) Law Practice 32.
Bar-Ziv S and Elkin-Koren N, ‘Behind the Scenes of Online Copyright Enforcement: Empirical Evidence on Notice & Takedown’ (2018) 50(2) Connecticut Law Review 339.
Berberich M and Conrad A, ‘§ 30 Plattformen und KI’ in M Ebers, CA Heinze and B Steinrötter (ed), Künstliche Intelligenz und Robotik (Beck 2020) 930.
Berberich M, ‘§ 5 Sorgfaltspflichten, Moderationsverfahren und prozedurale Fairness’ in B Steinrötter (ed), Europäische Plattformreguliereng (Nomos 2023) 126.
Bizikova L, Hancock P, Jewell D and Sherr I, ‘IA Meets AI – Rise of the Machines’, <https://dailyjus.com/legal-tech/2023/10/ia-meets-ai-rise-of-the-machines> accessed 31 December 2023.
Bloch-Wehba H, ‘Automation in Moderation’ (2020) 53(1) Cornell International Law Journal 41.
Bloch-Wehba H, ‘Global Platform Governance: Private Power in the Shadow of the State’ (2019) 72(1) SMU Law Review 27.
Boll-Kempelmann C, ‘Datenschutz und das Beweisverfahren im Schiedsverfahren’ (2022) 20(5) SchiedsVZ (Zeitschrift für Schiedsverfahren) 241.
Bomhard D and Merkle M, ‘Regulation of Artificial Intelligence’ 2021 (6) EuCML (Journal of European Consumer and Market Law) 257.
Brazil W, ‘Informalism and Formalism in the History of ADR in the United States’ in J Zekoll, M Bälz and I Amelung (ed), Formalisation and Flexibilisation in Dispute Resolution (Brill 2014) 250.
Brechmann B, Legal Tech und das Anwaltsmonopol (Mohr Siebeck 2021).
Breiman L, ‘Random Forests’ (2001) 45(1) Machine Learning 5.
Bridy A, ‘Intellectual Property’ in Keller D (ed), Law, Borders, and Speech: Proceedings and Materials (The Center for Internet and Society 2017) 9.
Bünnau P von, ‘Künstliche Intelligenz im Recht’ in S Breidenbach and F Glatz (ed), Rechtshandbuch Legal Tech (2nd edn, Beck/Manz 2021) 71.
Bull L and Steffek F, ‘The Decoding of Legal Conflicts’ (2018) 21(5) ZKM (Zeitschrift für Konfliktmanagement) 165.
Burk D L, ‘Algorithmic Fair Use’ (2019) 86(2) University of Chicago Law Review 283.
Burrell J, ‘How the machine “thinks”: Understanding opacity in machine learning algorithms’ (2016) (1) Big Data & Society 3, https://doi.org/10.1177/2053951715622512 accessed 31 December 2023.
Busch C, ‘Mehr Fairness und Transparenz in der Plattformökonomie?’ (2019) 121(8) GRUR (Gewerblicher Rechtsschutz und Urheberrecht) 788.
Calliess C and Ruffert M (ed), EUV/AEUV (6th edn, Beck 2022).
Casey B et al, ‘Rethinking Explainable Machines: The GDPR's “Right to Explanation” Debate and the Rise of Algorithmic Audits in Enterprise’ (2019) 34(1) BERKELEY TECH. L.J. 145.
Castets-Renard C, ‘Algorithmic Content Moderation on Social Media in EU Law: Illusion of Perfect Enforcement’ (2020) (2) University of Illinois Journal of Law, Technology & Policy 283.
Cellan-Jones R, ‘The robot lawyers are here’, <https://www.bbc.com/news/technology-41829534> accessed 31 December 2023.
Cervenka A and Schwarz P, ‘Data Protection in Arbitration Proceedings’ (2020) 18(2) SchiedsVZ (Zeitschrift für Schiedsverfahren) 78.
Chang B, ‘From Internet Referral Units to International Agreements: Censorship of the Internet by the UK and EU’ (2018) 49(2) Columbia Human Rights Law Review 114.
Chander A, ‘The Racist Algorithm?’ (2017) 115(6) Michigan Law Review 1023.
Chargeflow, ‘Paypal Dispute automation’, https://www.chargeflow.io/paypal-dispute-automation accessed 31 December 2023.
Citron D K, ‘Technological due process’ (2008) 85(6) Washington University Law Review 1249.
Citron D K and Pasquale F, ‘The Scored Society: due process for automated predictions’ (2014) 89(1) WASH. L. REV. 1.
Civil resolutional tribunal, ‘Civil resolutional tribunal’, https://civilresolutionbc.ca/solution-explorer/ accessed 31 December 2023.
Cohen J E, Between Truth and Power: The Legal Constructions of Informational Capitalism (Oxford University Press 2019).
Cohen P, ‘Bytes and Prejudice’ (2015) 1(1) Journal of Technology in International Arbitration 57.
Conrad A and Nolte G, ‘Schrankenbestimmungen im Anwendungsbereich des UrhDaG’ (2021) 65 ZUM (Zeitschrift für Urheber- und Medienrecht) 111.
Cooper S, Rule C and Del Duca L, ‘From Lex Mercatoria to Online Dispute Resolution’ (2011) 43 Uniform Commercial Code Law Journal, Penn State Legal Studies Research Paper No 09/2011.
Deckenbrock C and Henssler M, Rechtsdienstleistungsgesetz (5th edn, Beck 2021).
Deichsel T, Digitalisierung der Streitbeilegung (Nomos 2022).
Deichsel T, ‘Verbraucherschlichtungsstellen – Ein Anwendungsfeld für Legal Tech?‘ (2020) 35(8) VuR (Verbraucher und Recht) 283.
Denga M, ‘Platform Regulation by European Values: On the Binding of Opinion Platforms to EU Fundamental Rights’ (2021) 56(5) EuR (Europarecht) 569.
Deusch F and Eggendorfer T, ‘IT-Sicherheit’ in J Taeger and J Pohle (ed), Computerrechts-Handbuch (38 th edn, Beck 2023) ch 50.1.
Dewey C, ‘98 personal data points that Facebook uses to target ads to you’, The Washington Post, 19 August 2016, https://www.washingtonpost.com/news/the-intersect/wp/2016/08/19/98-personal-data-points-that-facebook-uses-to-target-ads-to-you/ accessed 31 December 2023.
Dispute Resolution Data, ‘Dispute Resolution Database’, https://www.disputeresolutiondata.com/ accessed 31 December 2023.
Disco, ‘Disco Powerful AI and Analytics’, https://www.csdisco.com/offerings/ediscovery/features-ai accessed 31 December 2023.
Douek E, ‘Content Moderation as Systems Thinking’ (2022) 136(2) Harvard Law Review 528.
Douek E, ‘How Much Power Did Facebook Give Its Oversight Board?’, <https://www.lawfaremedia.org/article/how-much-power-did-facebook-give-its-oversight-board> accessed 31 December 2023.
Douek E, ‘The Oversight Board Moment You Should've Been Waiting For, Lawfare’, <https://www.lawfaremedia.org/article/oversight-board-moment-you-shouldve-been-waiting-facebook-responds-first-set-decisions> accessed 31 December 2023.
Drexl J, ‘Bedrohung der Meinungsvielfalt durch Algorithmen’ (2017) 61(7) ZUM (Zeitschrift für Urheber- und Medienrecht) 529.
Drupal, ‘a2J Author’, https://www.a2jauthor.org/ accessed 31 December 2023.
Dubois E and Blank G, ‘The echo chamber is overstated: the moderating effect of political interest and diverse media’ (2018) 21(5) Information, Communication & Society 729.
Ebrevia, ‘DraftPro’, https://www.dfinsolutions.com/products/ebrevia accessed 31 December 2023.
Eghbariah R and Metwally A, ‘Informal Governance: Internet Referral Units and the Rise of State Inter-pretation of Terms of Service’ (2021) 23 Yale J.L. & Tech. 545.
Eidenmüller H and Wagner G, Law by Algorithm (Mohr Siebeck 2021).
Eifert M et al, Netzwerkdurchsetzungsgesetz in der Bewährung (Nomos 2020).
Elkin-Koren N, ‘After twenty years: revisiting copyright liability of online intermediaries’ in S Frankel and D Gervais (ed), The Evolution and Equilibrium of Copyright in the Digital Age (Cambridge University Press 2014) 29.
Elkin-Koren N, ‘Contesting algorithms: Restoring the public interest in content filtering by artificial intelligence’ (2020) 7(2) Big Data & Society 1, https://journals.sagepub.com/doi/epub/10.1177/2053951720932296 accessed 31 December 2023.
Elkin-Koren N and Perel M, ‘Separation of Functions for AI: Restraining Speech Regulation by Online Platforms’ (2020) 24(3) Lewis & CLARK L. REV. 857.
Engert A, ‘Digitale Plattformen’ (2018) 218(2-4) AcP (Archiv für die civilistische Praxis) 304.
Engstrom E and Feamster N, The Limits of Filtering: A Look at the Functionality & Shortcomings of Content Detection Tools (Engine 2017), https://static1.squarespace.com/static/571681753c44d835a440c8b5/t/58d058712994ca536bbfa47a/1490049138881/FilteringPaperWebsite.pdf accessed 31 December 2023.
European Commission, ‘Staff Working Document Impact Assessment, Accompanying the document Pro-posal for a Regulation of the European Parliament and of the Council on preventing the dissemina-tion of terrorist content online’ (12 September 2018) SWD (2018), 408 final.
European Commission, ‘Tackling online disinformation: Commission proposes an EU-wide Code of Practice’ (26 April 2018), https://ec.europa.eu/commission/presscorner/detail/en/IP_18_3370 accessed 31 December 2023.
Europol, ‘2018 Consolidated Annual Activity 44’ (23 May 2019), https://www.europol.europa.eu/cms/sites/default/files/documents/consolidated_annual_activity_report_2018.pdf accessed 23 December 2023.
Europol, ‘EU Internet Referral Unit – EU IRU’, https://www.europol.europa.eu/about-europol/european-counter-terrorism-centre-ectc/eu-internet-referal-unit-eu-iru accessed 31 December 2023.
Facebook Oversight Board, ‘Oversight Board Bylaws’, https://www.oversightboard.com/wp-content/uploads/2024/03/Oversight-Board-Bylaws.pdf accessed 31 December 2023.
Facebook Oversight Board, ‘Oversight Board Charter’, https://about.fb.com/wp-content/uploads/2019/09/oversight_board_charter.pdf accessed 31 December 2023.
Fiala F and Husovec M, ‘Using Experimental Evidence to Improve Delegated Enforcement’ (2022) 71 International Review of Law and Economics 106079 and (2018) 28 Forthcoming TILEC Discussion Paper no 2018-028.
Fireflies, ‘Fireflies AI’, https://firefliesai/ accessed 31 December 2023.
Fries M, ‘Erfüllung von Geldschulden über eigenwillige Zahlungsdienstleister’ (2018) 33(4) VuR (Verbraucher und Recht) 123.
Fries M, ‘Legal Tech im Schiedsverfahren’ in R Wilhelmi and M Stürner (ed), Mehrparteienschiedsverfahren. Unter besonderer Berücksichtigung gesellschaftsrechtlicher Streitigkeiten (Springer 2021) 85.
Fries M, ‘PayPal Law und Legal Tech – Was macht die Digitalisierung mit dem Privatrecht?’ (2016) 69(39) NJW (Neue Juristische Wochenschrift) 2860.
Fries M, Verbraucherrechtsdurchsetzung (Mohr Siebeck 2016).
Fritz G, Prantl D, Leinwather N and Hofer M, ‘Data Protection in International Arbitration Proceedings’ (2019) 17(6) SchiedsVZ (Zeitschrift für Schiedsverfahren) 301.
Frosio G, ‘Algorithmic Enforcement Online’ in P Torremans (ed), Intellectual Property and Human Rights (4th edn, Kluwer Law International 2020) 24.
Gerdemann S and Spindler G, ‘Das Gesetz über digitale Dienste (Digital Services Act) (Part 2) – Die Regelungen für Online-Plattformen sowie sehr große Online-Plattformen und Suchmaschinen’ (2023) 125(3) GRUR (Gewerblicher Rechtsschutz und Urheberrecht) 115.
Gielen N and Uphues S, ‘Digital Markets Act und Digital Services Act’ (2021) 32(14) EuZW (Europäische Zeitschrift für Wirtschaftsrecht) 627.
Gillespie T, ‘The Relevance of Algorithms’ in T Gillespie, P J Boczkowski and K A Foot (ed), Media Technologies (The MIT Press 2014) 188.
Gläßer U, ‘Mediation und Digitalisierung’ in T Riehm and S Dörr (ed), Digitalisierung und Zivilverfahren (De Gruyter 2023) 529.
Global Arbitration Review (GAR), ‘Database’, https://globalarbitrationreview.com/tools/arbitrator-research-tool accessed 31 December 2023.
Glogowski M, Plattformbedingungen (Mohr Siebeck 2022).
Google, ‘Asset erstellen’, https://support.google.com/youtube/answer/3011552?hl=de&ref_topic=3011550 accessed 31 December 2023.
Google, ‘Transparency Reports’, <https://transparencyreport.google.com/> accessed 31 December 2023.
Google, ‘Transparency Report for YouTube Platform for January to June 2023’ <https://transparencyreport.google.com/netzdg/youtube?hl=de> accessed 31 December 2023.
Google, ‘How Google Fights Piracy’, https://www.youtube.com/watch?v=QuOn93KPIr4 accessed 31 December 2023.
Google, ‘Counter notification’ https://support.google.com/youtube/answer/6005919?hl=de&ref_topic=9282678 accessed 31 December 2023.
Gorwa R, Binns R and Katzenbach C, ‘Algorithmic content moderation’ (2020) 7(1) Big Data & Society 1, https://journals.sagepub.com/doi/epub/10.1177/2053951719897945 accessed 31 December 2023.
Gray J E and Suzor N P, ‘Playing with machines: using machine learning to understand automated copyright enforcement at scale’ (2020) 7(1) Big Data & Society 1, https://journals.sagepub.com/doi/epub/10.1177/2053951720919963 accessed 31 December 2023.
Greger R and Stubbe C, Schiedsgutachten (Beck 2007).
Greger R, ‘Recht der alternativen Konfliktlösung’ in R Greger, H Unberath and F Steffek (ed), Recht der alternativen Konfliktlösung (2nd edn, Beck 2016) 270.
Grimmelmann J, ‘The Virtues of Moderation’ (2015) 17 YALE J.L. & TECH. 42.
Grosse Ruse-Kahn H, ‘Automated Copyright Enforcement Online: From Blocking to Monetization of User-Generated Content’ PIJIP Research Paper Series 51, https://digitalcommons.wcl.american.edu/research/51 accessed 31 December 2023.
Grosse Ruse-Kahn H, ‘Global Content Protection through Automation’ (2018) 49(9) IIC (International Review of Intellectual Property and Competition Law) 1017.
Gsell B, ‘Die Umsetzung der Richtlinie über alternative Streitbeilegung Juristisches Fachwissen der streitbeilegenden Personen und Rechtstreue des Verfahrensergebnisses’ (2015) 128(2) ZZP (Zeitschrift für Zivilprozess) 189.
Gsell B, Krüger W, Lorenz S and Reymann C (ed), Beck-Online Grosskommentar zum BGB (Beck 2023).
Gunning D et al., XAI-Explainable artificial intelligence, 4(37) Science Robotics (2019), https://www.science.org/doi/10.1126/scirobotics.aay7120 accessed 31 December 2023.
Haber E, ‘Privatization of the Judiciary’ (2016) 40(1) Seattle University Law Review 115.
Halis Kasap G, ‘Can Artificial Intelligence (“AI”) Replace Human Arbitrators?’ (2021) (2) Journal of Dispute Resolution 209.
Hartung M, ‘Sonstige Akteure und Rahmenbedingungen’ in M Hartung, M-M Bues and G Halbleib (ed), Legal Tech: Die Digitalisierung des Rechtsmarkts (Beck 2018) 215.
Heetkamp S J and Piroutek C, ‘ChatGPT in Mediation und Schlichtung‘ (2023) 26(3) ZKM (Zeitschrift für Konflikt-management) 80.
Heiss S, ‘Artificial Intelligence Meets European Union Law’ (2021) 10(6) EuCML (Journal of European Consumer and Market Law) 252.
Heldt A, Intensivere Drittwirkung (Mohr Siebeck 2023).
Helfer L and Land M K, ‘The Meta Oversight Board's Human Rights Future’ (2023) 44(106) Cardozo Law Review 2233.
Helmond A, ‘The Platformization of the Web: Making Web Data Platform Ready’ (2015) Social Media and Society 1.
Hess B, Europäisches Zivilprozessrecht (2nd edn, De Gruyter 2021).
Hess B, ‘Prozessuale Mindestgarantien in der Verbraucherschlichtung’ (2015) 70(11) JZ (Juristenzeitung) 548.
Hess T and Waltermann H, ‘Upload-Filter für Content’ (2019) 16(2) MedienWirtschaft 19.
Hoeren T, ‘Sperrpflichten eines Hosting-Anbieters bei rechtswidrigen Informationen sowie wort- und sinngleichen Inhalten’ (2020) (4) LMK (Leitsätze mit Kommentierung) 425949.
Hofmann F, ‘Die neuen Transparenzvorgaben im UWG 2022 im Kontext lauterkeitsrechtlicher Plattformregulierung’ (2022) 124(11) GRUR (Gewerblicher Rechtsschutz und Urheberrecht) 780.
Hofmann F‚ ‘Fünfzehn Thesen zur Plattformhaftung nach Art 17 DSM-RL’ (2019) 121(12) GRUR (Gewerblicher Rechtsschutz und Urheberrecht) 1219.
Hofmann F, ‘Mittelbare Verantwortlichkeit im Internet’ (2017) 57(8) JuS (Juristische Schulung) 713.
Hofmann F, ‘Prozeduralisierung der Haftungsvoraussetzungen im Medienrecht – Vorbild für die Intermediärshaftung’ (2017) 61(2) ZUM (Zeitschrift für Urheber- und Medienrecht) 102.
Hofmann F and Raue B (ed), Digital Services Act (Nomos 2023).
Hofmann F and Specht-Riemenschneider L, ‘Verantwortung von Online-Plattformen (Responsibility of Online Platforms)’ (2021) 13(1) ZGE (Zeitschrift für geistiges Eigentum) 48.
Hofmann F and Sprenger T, ‘Privatization of Enforcement’ (2021) 85(2) UFITA (Archiv für Medienrecht und Medienwissenschaft) 249.
Holznagel D, ‘Melde- und Abhilfeverfahren zur Beanstandung rechtswidrig gehosteter Inhalte nach europäischem und deutschem Recht im Vergleich zu gesetzlich geregelten notice and take-down-Verfahren‘ (2014) 63(2) GRUR Int. (Gewerblicher Rechtsschutz und Urheberrecht Internationaler Teil) 105.
Holznagel D, Notice and Takedown as Part of Provider Liability (Mohr Siebeck 2013).
Holznagel D, ‘Nutzerrechte bei Facebook: Klärung durch den BGH und bevorstehende Irrwege des EU-Gesetzgebers’ (2021) 37(11) CR (Computer und Recht) 733.
Holznagel D, ‘Zu starke Nutzerrechte in Art. 17 und 18 DSA’ (2022) 38(9) CR (Computer und Recht) 594.
Horwitz J, ‘Facebook Says Its Rules Apply to All. Company Documents Reveal a Secret Elite That’s Exempt’, (2021) Wall Street Journal, https://www.wsj.com/articles/facebook-files-xcheck-zuckerberg-elite-rules-11631541353 accessed 31 December 2023.
Hörnle J, Internet Jurisdiction (Oxford University Press 2021).
Jacques S, Garstka K, Hviid M and Street J, ‘An empirical study of the use of automated anti-piracy systems and their consequences for cultural diversity’ (2018) 15(2) Script-Ed 277.
Janal R, ‘Haftung und Verantwortung im Entwurf des Digital Services Acts’ (2021) 29(2) ZEuP (Zeitschrift für Europäisches Privatrecht) 227.
Jensen O, Tribunal secretaries in international arbitration (Oxford University Press 2019).
Jus Mundi, ‘AI-Powered Search for International Law and Arbitration’, https://jusmundi.com/en accessed 31 December 2023.
Jus Mundi, ‘Jus Mundi Introduces Jus AI: A Game-Changing GPT-Powered AI Solution for the Arbitration Community’ https://dailyjus.com/news/2023/06/jus-mundi-introduces-jus-ai-a-game-changing-gpt-powered-ai-solution-for-the-arbitration-community accessed 31 December 2023.
Kadri T and Klonick K, ‘Facebook v. Sullivan: Public Figures and Newsworthiness in Online Speech Online Speech’ (2019) 93(1) Southern California Law Review 37.
Kaesling K, ‘Evolution statt Revolution der Plattformregulierung‘ (2021) 65(3) ZUM (Zeitschrift für Urheber- und Medienrecht) 177.
Kaesling K, ‘Privatising Law Enforcement in Social Networks: A Comparative Model Analysis’ (2018) 11(3) Erasmus Law Review 151.
Kahneman D, Thinking, Fast and Slow (Farrar, Straus and Giroux 2013).
Kalbhenn J-C, ‘Design Specifications for Chatbots, Deepfakes, and Emotion Recognition Systems’ (2021) 65(8/9) ZUM (Zeitschrift für Urheber- und Medienrecht) 663.
Kaminski M E, ‘Binary Governance: Lessons from the GDPR’s Approach to Algorithmic Accountability’ (2019) 92(6) S. CAL. L. REV. 1529.
Katsh E and Rabinovich-Einy O, Digital Justice (Oxford University Press 2017).
Katyal S, ‘Private Accountability in the Age of Artificial Intelligence’ (2019) 66(1) UCLA Law Review 55.
Katzenbach C, ‘The “Alghorithmic turn” in platform governance’ (2022) 74(1 supp) Cologne Journal of Sociology and Social Psychology 283.
Kaulartz M, ‘Smart Contract Dispute Resolution’ in M Fries and B-P Paal (ed), Smart Contracts (Mohr Siebeck 2019) 73.
Kaye D, Speech Police (Columbia Global Reports 2019).
Keller D, ‘Internet Platforms: Observations on Speech, Danger, and Money’ (2018) Hoover Inst. Aegis Paper Series No 1807.
Kindt T, ‘Blockchainbasierte dezentrale Streitbeilegungsverfahren und ihr Verhältnis zur Schiedsgerichtsbarkeit’ (2023) 21(5) SchiedsVZ (Zeitschrift für Schiedsverfahren) 241.
Kleros, ‘Arbitration Service’, https://kleros.io/en/ accessed 31 December 2023.
Klonick K, ‘The New Governors: The People, Rules and Process Governing Online Speech’ (2018) 131(6) Harvard Law Review 1598.
Klonick K, ‘Why the History of Content Moderation Matters’ (2018) TECHDIRT, <https://www.techdirt.com/articles/20180129/21074939116/whyhistory-content-moderation-matters.shtml> accessed 31 December 2023.
Kosseff J, ‘Private Computer Searches and the Fourth Amendment’ (2018) 14(2) I/S A Journal of Law and Policy 187.
Koulu R, ‘Proceduralizing control and discretion: Human oversight in artificial intelligence policy’ (2020) 27(6) Maastricht Journal of European and Comparative Law 720.
Kreis F, ‘KI und ADR-Verfahren’ in Kaulartz M and Braegelmann T (ed), Rechtshandbuch Artificial Intelligence und Machine Learning (Beck and Vahlen 2020) 633.
Krenzler M and Remmertz F R (ed), Rechtsdienstleistungsgesetz (3rd edn, Nomos 2023).
Krüger W and Rauscher T (ed), Münchener Kommentar zur ZPO (6th edn, Beck 2022).
Kumar S and Kumar H, ‘Mediation and Artificial Intelligence’ (2021) 4(4) International Journal of Law Management & Humanities 1472.
Kumkar L, ‘Plattform-Recht revisited: Umgang mit den Marktordnungen digitaler Plattformen de lege lata et ferenda‘ (2022) 30(3) ZEuP (Zeitschrift für Europäisches Privatrecht) 530.
Kuczerawy A, ‘The Good Samaritan that wasn’t: voluntary monitoring under the (draft) Digital Services Act’ (2021) Verfassungsblog, https://verfassungsblog.de/good-samaritan-dsa/ accessed 31 December 2023.
Kraul T (ed), Das neue Recht der digitalen Dienste – Digital Services Act (DSA) (Nomos 2023).
Ladeur K-H, ‘Neue Institutionen für den Daten- und Persönlichkeitsschutz im Internet: „Cyber-Courts“ für die Blogosphere‘ (2012) 36(10) DuD (Datenschutz und Datensicherheit) 711.
Ladeur K-H, ‘Schutz vor Verletzung von Persönlichkeitsrechten und “Desinformation“ in sozialen Medien unter Bedingungen der politischen Polarisierung’ <https://verfassungsblog.de/personlichkeitsrecht-soziale-medien/> accessed 31 December 2023.
Länderarbeitsgruppe, ‘Abschlussbericht der Länderarbeitsgruppe, “Legal Tech: Herausforderungen für die Justiz’’’ (2019), https://www.schleswigholstein.de/DE/landesregierung/ministerienbehoerden/II/Minister/Justizministerkonferenz/Downloads/190605_beschluesse/TOPI_11_Abschlussbericht.pdf?blob=publicationFile&v=1 accessed 31 December 2023.
Land M, ‘The Problem of Platform Law: Pluralistic Legal Ordering on Social Media’ in P SA Berman (ed), The Oxford Handbook of Global Legal Pluralism (Oxford University Press 2020) 974.
Laukemann B, ‘Private law enforcement and intellectual property: Regulatory challenges in a digital era’ in B Hess, E Jayme and H-P Mansel (ed), Europa als Rechts- und Lebensraum: Liber Amicorum für Christian Kohler zum 75. Geburtstag (Gieseking 2018) 357.
Laukemann B, ‘Private Rechtsdurchsetzung zwischen (digitaler Selbsthilfe) und gerichtlichem Rechtsschutz’ (2022) 8(3) ZfPW (Zeitschrift für die gesamte Privatrechtswissenschaft) 357.
Lawlift GmbH, ‘Lawflift’, https://de.lawlift.com/ accessed 31 December 2023.
Ledwich M and Zautsev A, ‘Algorithmic Extremism: Examining YouTube’s Rabbit Hole of Radicalization’, Cornell University (2019), https://arxiv.org/abs/1912.11211 accessed 31 December 2023.
Leeb C-M, Digitalisierung, Legal Technology und Innovation (Duncker & Humboldt 2019).
Legal Grade AI, ‘Luminace’, https://www.luminance.com/overview.html accessed 31 December 2023.
Leenheer Zimmerman D, ‘A Tale of Legislative Abdication’ (2014) 35(1) Pace Law Review 260.
Leerssen P, ‘An end to shadow banning? Transparency rights in the Digital Services Act between content moderation and curation’ (2023) 48 Computer Law & Security Review 6.
Lennartz J and Kraetzig V, ‘Filtering fundamental Rights’ (2022) Verfassungsblog, https://verfassungsblog.de/filtering-fundamental-rights/ accessed 31 December 2023.
Lessig L, Code and Other Laws of Cyberspace (Basic Books 1999).
Lester T and Pachamanova D, ‘The Dilemma of False Positives: Making Content ID Algorithms More Conducive to Fostering Innovative Fair Use in Music Creation’ (2017) 24(1) UCLA Entertainment Law Review 51.
Lev-Aretz Y, ‘Second Level Agreements’ (2012) 45(1) AKRON L. REV. 137.
Lew J, Mistelis L and Kröll S, Comparative International Commercial Arbitration (Wolters Kluwer 2003).
Lex Machina, ‘Lex Machina Legal Analytics’, https://lexmachina.com/ accessed 31 December 2023.
Liesching M (ed), Netzwerkdurchsetzungsgesetz (Nomos 2018).
Lindquist D and Dautaj Y, ‘AI in International Arbitration’ (2021) (1) Journal of Dispute Resolution 39.
Loo R Van, ‘Federal Rules of Platform Procedure’ (2021) 88(4) University of Chicago Law Review 829.
Loo R Van, ‘The Corporation as Courthouse’ (2016) 33(2) Yale Journal on Regulation 547.
Luhmann N, Legtimation durch Verfahren (11th edn, Suhrkamp 2019).
Lüdemann J, ‘Privatisierung der Rechtsdurchsetzung in sozialen Netzwerken? ’ in M Eifert and T Gostomzyk (ed), Netzwerkrecht (2018) 165.
Lüdemann J, ‘Warum und wie reguliert man digitale Informationsintermediäre?’ in J Lüdemann and Y Hermstrüwer (ed), Der Schutz der Meinungsbildung im digitalen Zeitalter (Mohr Siebeck 2021) 15.
Maier H, Remixes on Hosting Platforms (Mohr Siebeck 2018).
Marmont S, ‘Keeping Up with Legal Technology’ (2019) 1(2) ITA in Review 37.
Mast T, ‘AGB-Recht als Regulierungsrecht’ (2023) 78(7) JZ (Juristenzeitung) 287.
Maxwell G and Vannieuwenhuyse G, ‘Robots Replacing Arbitrators: Smart Contract Arbitration’ (2018) (1) ICC Dispute Resolution Bulletin 24.
Mayer M, Soziale Netzwerke im Internet im Lichte des Vertragsrechts (Richard Boorberg 2018).
McColgan P, ‘Das wird man wohl noch löschen dürfen? – Control Standards for Opinion Rules on the Internet‘ (2021) 1(12) RDi (Recht Digital) 605.
Meder S, ‘Die Zukunft der juristischen Methode: Rehabilitierung durch Chat-GPT?‘ (2023) 78(23) JZ (Juristenzeitung) 1041.
Medvedeva M, Vols M and Wieling M, ‘Using machine learning to predict decisions of the European Court of Human Rights’ (2020) 28(2) Artificial Intelligence and Law 237.
Mendelsohn J K, ‘Die “normative Macht“ der Plattformen – Gegenstand der zukünftigen Digitalregulierung’ (2021) 24(11) MMR (Multimedia und Recht) 857.
Menkel-Meadow C, ‘Is ODR ADR? Reflections of an ADR Founder from 15th ODR Conference, the Hague, the Netherlands, 22-23 May 2016’ (2016) 3(1) IJODR (International Journal on Online Dispute Resolution) 4.
Meta, ‘Community Standards Enforcement Report: Child Endangerment: Nudity and Physical Abuse and Child Sexual Exploitation’, https://transparency.fb.com/data/community-standards-enforcement/child-nudity-and-sexual-exploitation/facebook/ accessed December 2023.
Meta, ‘Community Standards Enforcement Report’, https://transparency.fb.com/reports/community-standards-enforcement/ accessed 31 December 2023.
Meta, ‘How Facebook uses super-efficient AI models to detect hate speech, 19 November 2020’, https://ai.facebook.com/blog/how-facebook-uses-super-efficient-ai-models-to-detect-hate-speech/ accessed 31 December 2023.
Meta, ‘How we review content’, 11 August 2020, https://about.fb.com/news/2020/08/how-we-review-content/ accessed 31 December 2023.
Meta, ‘How we review Content – Prioritization’, 11 August 2020, https://about.fb.com/news/2020/08/how-we-review-content/ accessed 31 December 2023.
Meta, ‘Sharing More Details on How We Will Implement the Oversight Board's Decisions, Responding to the Oversight Board’s First Decisions’, 28 January 2021, https://about.fb.com/news/2021/01/responding-to-the-oversight-boards-first-decisions/ accessed 31 December 2023.
Meta, ‘Facebook Transparency Report of January 2023’, <327151920_907084790305794_6193992151844220602_n.pdf> accessed 31 December 2023.
Meta Transparency Center, ‘Global restrictions’, https://transparency.fb.com/data/content-restrictions/ accessed 31 December 2023.
Meta Transparency Center, ‘How technology detects violations’, 18 October 2023, https://transparency.meta.com/de-de/enforcement/detecting-violations/technology-detects-violations/ accessed 31 December 2023.
Meta Transparency Center, ‘How we assess reports of content violating local law’, https://transparency.fb.com/data/content-restrictions/ accessed 31 December 2023.
Meta Transparency Center, ‘Reviewing high-impact content accurately via our cross-check system’, 12 May 2023, https://transparency.fb.com/enforcement/detecting-violations/reviewing-high-visibility-content-accurately/ accessed 31 December 2023.
Metzger A and Senftleben M, ‘Selected Aspects of Implementing Article 17 of the Directive on Copyright in the Digital Single Market into National Law – Comment of the European Copyright Society’ (2020) 11(2) Journal of Intellectual Property, Information Technology and E-Commerce Law 1.
Milano D, Content control: Digital Watermarking and Fingerprinting (Rhozet 2013).
Monroy M, ‘EU-Kommission droht mit “gesetzgeberischen Maßnahmen“ zur Entfernung von Internetinhalten’, https://netzpolitik.org/2018/eu-kommission-droht-mit-gesetzgeberischen-massnahmen-zur-entfernung-von-internetinhalten/ accessed 31 December 2023.
Monroy M, ‘EU-Internetforum: Viele Inhalte zu “Extremismus“ werden mit Künstlicher Intelligenz aufgespürt’, https://netzpolitik.org/2018/eu-kommission-droht-mit-gesetzgeberischen-massnahmen-zur-entfernung-von-internetinhalten/ accessed 31 December 2023.
Mostert F, ‘Free Speech and Internet Regulation’ (2019) 14(8) Journal of Intellectual Property Law & Practice 607.
Müller-Terpitz R and Köhler M (ed), Digital Services Act (Beck 2024).
Muller C, ‘Setting the Record Straight’, https://blog.youtube/news-and-events/setting-record-straight/ accessed 31 December 2023.
Murgia M, Warell H and Bond D, ‘YouTube revenues under threat over ads alongside extremist videos’ (2017) Financial Times, https://www.ft.com/content/04f8bf56-0b12-11e7-97d1-5e720a26771b accessed 31 December 2023.
Musa S and Bendett S, ‘Islamic Radicalization in the United States – New Trends and a Proposed Methodology for Disruption’ (2010) Washington D.C.: National Defense University, Washington DC Center for Technology and National Security Policy, https://apps.dtic.mil/sti/pdfs/ADA532696.pdf accessed 31 December 2023.
Nahmias Y and Perel M, ‘The oversight of content moderation by AI: Impact assessment and their limitations’ (2021) 58(1) Harvard Journal on Legislation 145.
Nathenson I S, ‘The Procedural Foundations of Information Regulation’ (2020) 24(1) Lewis & Clark Law Review 109.
Nink D, Justiz und Algorithmen (Duncker & Humboldt 2021).
Nolte G, ‘Three Theses on the Current Debate on Liability and Distributive Justice in Hosting Services with User-Generated Content (the so-called “Value Gap” Debate)’ (2017) 61(4) ZUM (Zeitschrift für Urheber- und Medienrecht) 304.
Nunziato D C, ‘The Beginning of the End of Internet Freedom’ (2014) 45(2) Georgetown Journal of International Law 383.
Ohly A and Sosnitza O (ed), Gesetz Gegen den unlauteren Wettbewerb: UWG (8th edn, Beck 2023).
Orssich I, ‘Das europäische Konzept für vertrauenswürdige Künstliche Intelligenz’ (2022) 33(6) EuZW (Europäische Zeitschrift für Wirtschaftsrecht) 254.
Otter, ‘Otter AI’, https://otter.ai accessed 31 December 2023.
Paisley K and Sussman E, ‘Artificial Intelligence Challenges and Opportunities for International Arbitration’ (2018) 11(1) New York Dispute Resolution Lawyer 35.
Pasquale F, The Black Box Society (Harvard University Press 2015).
Peifer K-N, ‘Die neuen Transparenzregeln im UWG (Bewertungen, Rankings und Influencer)’ (2021) 123(12) GRUR (Gewerblicher Rechtsschutz und Urheberrecht) 1453.
Perel M and Elkin-Koren N, ‘Accountability in Algorithmic Copyright Enforcement’ (2016) 19(3) Stan. Tech. L. Rev. 473.
Plantin J-C, Lagoze C, Edwards P N and Sandvig C, ‘Infrastructure Studies Meet Platform Studies in the Age of Google and Face-book’ (2016) 20(1) New MEDIA & Society 293.
Polkinghorne M, ‘Different Strokes for Different Folks?’, <https://arbitrationblog.kluwerarbitration.com/2014/05/17/different-strokes-for-different-folks-the-role-of-the-tribunal-secretary-2/> accessed 31 December 2023.
Prütting H, ‘Das neue Verbraucherstreitbeilegungsgesetz: Was sich ändert – und was bleiben wird’ (2016) (3) AnwBl (Anwaltsblatt) 190.
Prütting H, ‘Die rechtliche Stellung des Schiedsrichters’ (2011) 9(5) SchiedsVZ (Zeitschrift für Schiedsverfahren) 233.
Queen Mary University and White & Case, ‘2021 International Arbitration Survey: Adapting Arbitration to a Changing World’, https://arbitration.qmul.ac.uk/research/2021-international-arbitration-survey/ accessed 31 December 2023.
Rabinovich-Einy O and Katsh E, ‘The New New Courts’ (2017) 67(1) Am. U. L. Rev. 165.
Radu R, Negotiating Internet Governance (Oxford University Press 2019).
Rajendra J B and Thuraisingam A S, ‘The deployment of artificial intelligence in alternative dispute resolution: the AI augmented arbitrator’ (2022) 31(2) Information & Communications Technology Law 176.
Raue B, ‘Plattformnutzungsverträge im Lichte der gesteigerten Grundrechtsbindung marktstarker sozialer Netze‘ (2022) 75(2) NJW (Neue Juristische Wochenschrift) 209.
Raue B and Heesen H, ‘Der Digital Services Act‘ (2022) 75(49) NJW (Neue Juristische Wochenschrift) 3537.
Reuter M, ‘Facebook Knew What All Was Going Wrong’ (2021) Netzpolitik.org, https://netzpolitik.org/2021/facebook-files-facebook-wusste-was-alles-schieflaeuft/ accessed 31 December 2023.
Riesenhuber K, ‘§ 10 Die Auslegung’ in K Riesenhuber (ed), Europäische Methodenlehre (De Gruyter 2021) 285.
Richter P and Mendelsohn J, ‘§ 21 Plattformspezifische Vorgaben des Data Acts’ in B Steinrötter (ed), Europäische Plattformregulierung (Nomos 2023) 564.
Rhim Y and Park K, ‘The Artificial Intelligence in International Law’ in E Y J Lee (ed), Revolutionary Approach to international Law: The Role of international Lawyer in Asia (Springer 2023) 215.
Röthemeyer P, ‘Die Schlichtung‘ (2013) 16(2) ZKM (Zeitschrift für Konfliktmanagement) 47.
Ruger T, Kim P, Martin A and Quinn K, ‘The Supreme Court Forecasting Project’ (2004) 104(4) Colum. L. Rev. 1150.
Rule C, ‘Making Peace on eBay’ (2008) ACR Resolution 8.
Rule C, ‘Quantifying the Economic Benefits of Effective Redress: Large ECommerce Data Sets and the Cost-Benefit Case for Investing in Dispute Resolution’ (2012) 34(4) U. ARK. LITTLE ROCK L. REV. 767.
Rühl G, ‘Die Richtlinie über alternative Streitbeilegung: Handlungsperspektiven und Handlungsoptionen’ (2014) 127(1) ZZP (Zeitschrift für Zivilprozess) 61.
Rühl G, ‘Digitale Justiz‘ (2020) 75(17) JZ (Juristenzeitung) 809.
Rühl G, ‘KI in der gerichtlichen Streitbeilegung’ in M Kaulartz and T Braegelmann (ed), Rechtshandbuch Artificial Intelligence und Machine Learning (Beck 2020) 617.
Salter S and Thompson D, ‘Public-Centred Civil Justice Redesign’ (2016-2017) 3 McGill Journal of Dispute Resolution 113.
Saurwein F, ‘Regulierung von Internet-Inhalten: Ombudsstellen als Governance-Option an der Schnittstelle von Recht und Ethik’ in G Marci-Boehncke, M Rath, M Delere and H Höfer (ed), Medien – Demokratie – Bildung (Springer 2022) 47 https://doi.org/10.1007/978-3-658-36446-5_5 accessed 31 December 2023.
Schillmöller J and Doseva S, ‘”Chilling effects” durch YouTubes Content ID?’ (2022) 25(3) MMR (Multimedia und Recht) 181.
Schneiders P, ‘Hate Speech auf Online-Plattformen: Problematization, Regulation and Evaluation against the Background of the Proposal for a Digital Services Act’ (2021) 85(2) UFITA (Archiv für Medienrecht und Medienwissenschaft), https://doi.org/10.5771/2568-9185-2021-2-269 accessed 31 December 2023 accessed 31 December 2023.
Scherer M, ‘Artificial Intelligence and Legal Decision-Making’ (2019) 36(5) Journal of International Arbitration 539.
Scherer M, ‘International Arbitration 3.0. How Artificial Intelligence Will Change Dispute Resolution’ in C Klausegger et al (ed), Austrian Yearbook on International Arbitration (Beck 2019) 503.
Scherer M and Jensen O, ‘Die Digitalisierung der Schiedsgerichtsbarkeit’ in T Riehm and S Dörr (ed), Digitalisierung und Zivilverfahren (De Gruyter 2023), 591.
Schwartz J, ‘Artificial Arbitration?’ in R Wilhelmi and M Stürner (ed), Mehrparteien-Schiedsverfahren: Unter besonderer Berücksichtigung gesellschaftsrechtlicher Streitigkeiten (Springer 2021) 95.
Seetharaman D, Horwitz J and Scheck J, ‘Facebook Says AI Can Enforce Its Rules, but the Company’s Own Engineers Are Doubtful’ (2021) Wall Street Journal, https://www.wsj.com/articles/facebook-ai-enforce-rules-engineers-doubtful-artificial-intelligence-11634338184 accessed 31 December 2023.
Sela A, ‘The Effect of Online Technologies on Dispute Resolution System Design’ (2017) 21(3) Lewis & Clark Law Review 633.
Senftleben M, ‘Institutionalized Algorithmic Enforcement – The Pros and Cons of the EU Approach to UGC Platform Liability’ (2020) 14(2) FIU Law Review 299.
Shaughnessy P and Rogers C, ‘Arbitrator Intelligence - An Interview with its Founder and Director, Professor Catherine Rogers’ (2015) 1(1) Journal on Technology in International Arbitration 87.
Shinn L D, ‘YouTube’s Content ID as a Case Study of Private Copyright Enforcement Systems’ (2015) 43(2/3) AIPLA Quarterly Journal 359.
Silicon Valley Arbitration and Mediation Center, ‘Silicon Valley Arbitration & Mediation Center Guidelines, Draft of August 31, 2023’, https://thearbitration.org/wp-content/uploads/2023/08/SVAMC-AI-Guidelines-CONSULTATION-DRAFT-31-August-2023-1.pdf accessed 31 December 2023.
Sim C, ‘Will Artificial Intelligence Take Over Arbitration?’ (2018) 14(1) Asian International Arbitration Journal 1.
Simshaw D, ‘Access to A.I. Justice: Avoiding an Inequitable Two-Tiered System of Legal Services’ (2022) 24 Yale Journal of Law & Technology 150.
Snijders H, ‘Arbitration and AI, Arbitration 2023’ in H Snijders (ed), Arbitration and AI, Arbitration (Wolters Kluwer 2023) 224.
Solomon L, ‘Fair users or content abusers’ (2015) 44(1) Hofstra L. Rev. 237.
Specht F, ‘Chancen und Risiken einer digitalen Justiz für den Zivilprozess’ (2019) 22(3) MMR (Multimedia und Recht) 153.
Spindler G, ‘Der Vorschlag für ein neues Haftungsregime für Internetprovider – der EU-Digital Services Act (Teil 1)’ (2021) 123(4) GRUR (Gewerblicher Rechtsschutz und Urheberrecht) 545.
Spoerri T, ‘On Upload Filters and other Competitive Advantages for Big Tech Companies under Article 17 of the Directive on Copyright in the Digital Single Market’ (2019) 10(2) Journal of Intellectual Property, Information Technology and Electronic Commerce Law 173.
Staudinger J von von (ed), Kommentar zum BGB, Buch 2 (19th rev edn, De Gruyter 2022).
Stotz R, ‘Die Rechtsprechung des EuGH’ in K Riesenhuber (ed) Europäische Methodenlehre (De Gruyter 2021) 653.
Sunstein C R, Republic – Divided Democracy in the Age of Social Media (Princeton University Press 2017).
Suzor N P, Lawless: The Secret Rules that Govern our Digital Lives (Cambridge University Press 2019).
Suzor N P and Gray J E, ‘Playing with machines: Using machine learning to understand automated copyright enforcement at scale’ (2020) 7(1) Big Data & Society, https://doi.org/10.1177/2053951720919963 accessed 31 December 2023.
Sweeney L, ‘Discrimination in Online Ad Delivery’ May 2013 Comm. ACM 44, http://cacm.acm.org/magazines/2013/5/163753-discrimination-in-online-ad-delivery/ accessed 31 December 2023.
Taeger J and Kremer S, Recht im E-Commerce und Internet (Beck 2021).
Tan V, ‘Online Dispute Resolution for Small Civil Claims in Victoria’ (2019) 24 Deakin Law Review 101.
Titlow J P, ‘Youtube is using AI to police copyright to the tune of $2 billion in payouts’, 31 July 2016, https://www.fastcompany.com/4013603/youtube-is-using-ai-to-police-copyright-to-the-tune-of-2-billion-in-payouts accessed 31 December 2023.
Trint, ‘Trint’, https://trint.com/ accessed 31 December 2023.
Tushnet R, ‘All of this has happened before and all of this will happen again: Innovation in copyright licensing’ (2014) 29(3) Berkeley Technology Law Journal 1447.
Twitter, ‘Network Enforcement Report: January-June 2023’, https://transparency.twitter.com/content/dam/transparency-twitter/country-reports/germany/NetzDG-Jan-Jun-2023.pdf accessed 31 December 2023.
Tyler T R, ‘What is procedural justice?: Criteria used by citizens to assess the fairness of legal procedures’ (1988) 22(1) Law & Society Review 103.
Tyler T R, Why people obey the law (Princeton University Press 2006).
Urban J, Karagani J and Schofield B, ‘Notice and Takedown in Everyday Practice’ (2016) UC Berkeley Public Law Research Paper No 2755628, https://ssrn.com/abstract=2755628 or http://dx.doi.org/10.2139/ssrn.2755628.
Valkanova M, ‘Trainieren von KI-Modellen’ in M Kaulartz and T Braegelmann (ed), Rechtshandbuch Artificial Intelligence und Machine Learning (Beck 2020) 336.
Vannieuwenhuyse G, ‘Arbitration and New Technologies: Mutual Benefits’ (2018) 35(1) Journal of International Arbitration 119.
Voß W, ‘Gerichtsverbundene Online-Streitbeilegung‘ (2020) 84(1) RabelsZ (Rabels Zeitschrift für ausländisches und internationales Privatrecht) 62.
Wagner B, Global Free Expression – Governing the Boundaries of Internet Content (Springer 2016).
Wagner G, ‘Haftung von Plattformen für Rechtsverletzungen (Teil 2)’ (2020) 122(5) GRUR (Gewerblicher Rechts-schutz und Urheberrecht) 447.
Wagner G, ‘Private Law Enforcement and ADR’ in J Zekoll, M Bälz and I Amelung (ed), Formalisation and Flexibilisation in Dispute Resolution (Brill 2014) 369.
Wagner J, Legal Tech und Legal Robots (2nd edn, Springer Gabler 2020).
Walker K, ‘Four ways Google will help to tackle extremism’ (18 June 2017) Financial Times, https://www.ft.com/content/ac7ef18c-52bb-11e7-a1f2-db19572361bb accessed 31 December 2023.
Wall Street Journal, ‘The Facebook Files’ (13 September 2021), https://www.wsj.com/articles/the-facebook-files-11631713039 accessed 31 December 2023.
Welser M von, ‘Die KI-Verordnung – ein Überblick über das weltweit erste Regelwerk für künstliche Intelligenz’ (2024) 16(15) GRUR-Prax (Gewerblicher Rechtsschutz und Urheberrecht in der Praxis) 485.
Wendland M, Mediation und Zivilprozess (Mohr Siebeck 2017).
Wielsch D, ‘Die Ordnungen der Netzwerke, AGB – Code – Community Standards‘ in M Eifert and T Gostomzyk (ed), Netzwerkrecht (Nomos 2018) 61.
Wielsch D, ‘Medienregulierung durch Persönlichkeits- und Datenschutzrechte‘ (2020) 75(3) JZ (Juristenzeitung) 105.
Wilske S, Markert L and Ebert B, ‘Entwicklungen in der internationalen Schiedsgerichtsbarkeit im Jahr 2022 und Ausblick auf 2023’ (2023) 21(3) SchiedsVZ (Zeitschrift für Schiedsverfahren) 121.
Wolfowicz M, Weisburd D and Hasisi B, Examining the interactive effects of the filter bubble and the echo chamber on radicalization, (2023) 19(5) Journal of Experimental Criminology 119.
Wolters Kluwer, ‘Arbitration Database’, https://www.kluwerarbitration.com/ accessed 31 December 2023.
Wolters Kluwer, ‘smartlaw’, https://www.smartlaw.de/ accessed 31 December 2023.
Wu T, ‘Will Artificial Intelligence Eat the Law? The Rise of Hybrid Social-Ordering Systems’ (2019) 119(7) Columbia Law Review 2001.
Yeung K, ‘”Hypernudge”: Big Data as a mode of regulation by design, Information’ (2017) 20(1) Communication & Society 118.
Yablon Y and Landsman-Ross N, ‘Predictive Coding’ (2013) 64(3) South Carolina Law Review 633.
YouTube, ‘Accelerated Content ID & Complaint Process’, https://support.google.com/youtube/thread/171619847 accessed 31 December 2023.
YouTube, ‘Answers to common questions about Copyright claims on YouTube’, https://support.google.com/youtube/thread/1281991 accessed 31 December 2023.
YouTube, ‘Appeal a Content ID claim’, https://support.google.com/youtube/answer/12104471 accessed 31 December 2023.
YouTube, ‘Best practices for claims’, https://support.google.com/youtube/answer/4352063 accessed 31 December 2023.
YouTube, ‘Best practices for references’, <https://support.google.com/youtube/answer/107008> accessed: 31 December 2023.
YouTube, ‘Content eligible for Content ID’, https://support.google.com/youtube/answer/2605065?hl=en accessed 31 December 2023.
Youtube ‘Copyright Strike Basics’ https://support.google.com/youtube/answer/2814000#zippy=%2Cfolgen-einer-urheberrechtsverwarnung accessed 31 December 2023.
YouTube, ‘Copyright Transparency Report H2 2021’, https://storage.googleapis.com/transparencyreport/report-downloads/pdf-report-22_2021-7-1_2021-12-31_en_v1.pdf accessed 31 December 2023.
YouTube, ‘Create an asset’, https://support.google.com/youtube/answer/3011552?hl=en&ref_topic=3011550 accessed 31 December 2023.
YouTube, ‘Deliver content using spreadsheet templates’, https://support.google.com/youtube/answer/6066171 accessed 31 December 2023.
YouTube, ‘Dispute a Content ID claim’, <https://support.google.com/youtube/answer/2797454?hl=en&ref_topic=9282678#zippy=%2Coptionen-f%C3%BCr-den-anspruchsteller> accessed 31 December 2023.
YouTube, ‘Fix reference overlaps’, https://support.google.com/youtube/answer/3022604?hl=en&ref_topic=3013248 accessed 31 December 2023.
YouTube, ‘Frequently asked questions about copyright’, https://support.google.com/youtube/answer/2797449?hl=en accessed 31 December 2023.
YouTube, ‘Frequently asked questions about fair use’, https://support.google.com/youtube/answer/6396261#zippy=%2Ci-posted-a-disclaimer-on-my-video%2Ci-gave-credit-to-the-copyright-owner%2Cim-using-the-content-for-entertainment-or-non-profit-uses%2Cwhen-does-fair-use-apply%2Cwhat-constitutes-fair-use%2Chow-does-fair-use-work%2Chow-does-content-id-work-with-fair-use accessed 31 December 2023.
YouTube, ‘How policies are applied’, https://support.google.com/youtube/answer/3369929 accessed 31 December 2023.
YouTube, ‘Monetization during Content ID disputes’, https://support.google.com/youtube/answer/7000961?hl=en&ref_topic=9282678 accessed 31 December 2023.
YouTube, ‘Qualify for Content ID’, https://support.google.com/youtube/answer/1311402 accessed 31 December 2023.
YouTube, ‘Requirements for counter notifications’, https://support.google.com/youtube/answer/6005919?hl=en&ref_topic=9282678 accessed 31 December 2023.
YouTube, ‘Review potentially invalid references’, https://support.google.com/youtube/answer/6013183 accessed 31 December 2023.
YouTube, ‘Update: Improving Content ID for creators’, https://blog.youtube/news-and-events/update-improving-content-id-for-creators/ accessed 31 December 2023.
YouTube, ‘Use Content ID matching on live streams’, https://support.google.com/youtube/answer/9896248?hl=en accessed 31 December 2023.
YouTube, ‘Using Content ID’, https://support.google.com/youtube/answer/3244015?hl=en accessed 31 December 2023.
YouTube, ‘Using the YouTube DDEX feed’, https://support.google.com/youtube/topic/3505247 accessed 31 December 2023.
YouTube, ‘What are policies?’, https://support.google.com/youtube/answer/107383?hl=en&ref_topic=24332 accessed 31 December 2023.
YouTube, ‘What Does Fair Use Mean’, https://support.google.com/youtube/answer/9783148?hl=de accessed 31 December 2023.
YouTube, ‘What is copyright?’, https://support.google.com/youtube/answer/2797466#refrained&zippy=%2Cmissverst%C3%A4ndnis-nr-if-you-angive-that-your-content-does-not-serve-commercial-purposes-you-can-use-any-content accessed 31 December 2023.
YouTube, ‘What is a reference?’ https://support.google.com/youtube/answer/107004?hl=en accessed 31 December 2023.
YouTube, ‘Where can I get more information about copyright outside the U.S.?’ https://support.google.com/youtube/answer/2797449?hl=de&ref_topic=2778546#zippy=%2Cwo-erhalte-ich-weitere-informationen-zum-urheberrecht-außerhalb-der-usa accessed 31 December 2023.
YouTube, ‘YouTube Partner Program overview & eligibility’ https://support.google.com/youtube/answer/72851 accessed 31 December 2023.
Zawada K L, ‘The Emergence and Development of Content ID in Light of User-generated Law, in How Deep is your Law?’ (2017) 5th International Conference of PhD Students and Young Researchers Conference Papers 483.
Zekos G, Advanced Artificial Intelligence and Robo-Justice (Springer 2022).
Zhou T, Postmorten: ‘Every Frame a Painting (2017)’ https://perma.cc/U5WU-M6ZZ accessed 31 December 2023.
Zorrilla E, ‘Towards a Credible Future’ (2018) 16(2) SchiedsVZ (Zeitschrift für Schiedsverfahren) 106.
***
[1] The document is up to date as at 31 December 2023 with regard to sources and the legal situation. However, where relevant, changes resulting from the European Union’s AI Act have been incorporated.
[2] Prof Dr Björn Laukemann (Maîtr en droit Aix-en-Provence) holds the Chair of Civil Law, German and International Law of Civil Procedure at the Eberhard Karls University of Tübingen (Germany).
[3] H Bloch-Wehba, ‘Automation in Moderation’ (2020) 53(1) Cornell International Law Journal 41, 74 f.
[4] On the particular market dominance of the platform and its relevance to public discourse see M Perel and N Elkin-Koren, ‘Accountability in Algorithmic Copyright Enforcement’ (2016) 19(3) Stan. Tech. L. Rev. 473, 497.
[5] This is why Google, YouTube’s parent company, markets the Content ID System as not merely an upload filter, but even a ‘copyright management system’, see Google, ‘How Google Fights Piracy’ (2018) 24 https://kstatic.googleusercontent.com/files/2bc15c350e6d8ba6363594195712a3c2528e56502c41c8a8a431746afce40adb9956ff837f9e54887c0277b413bceb8d79adc02ddae97c24969b55a30c70d836 accessed 31 December 2023; J Schillmöller and S Doseva, ‘”Chilling effects” durch YouTubes Content ID?’ (2022) 25(3) MMR (Multimedia und Recht) 181, 182; J E Gray and N P Suzor, ‘Playing with machines: using machine learning to understand automated copyright enforcement at scale’ (2020) 7(1) Big Data & Society 1, 2 https://doi.org/10.1177/2053951720919963 accessed 31 December 2023. – For general information on the Content ID procedure, see Perel and Elkin-Koren (n 4), (2016) 19(3) Stan. Tech. L. Rev. 473, 497; H Grosse Ruse-Kahn, ‘Automated Copyright Enforcement Online: From Blocking to Monetization of User-Generated Content’ (2020) PIJIP Research Paper Series 51 https://digitalcommons.wcl.american.edu/research/51 accessed 31 December 2023.
[6] For an overview of the technical functioning see Perel and Elkin-Koren (n 4), (2020) 19(3) Stan. Tech. L. Rev. 473 at fn 210-211 with further references; Gray and Suzor (n 5), (2020) 7(1) Big Data & Society 1, 2.
[7] YouTube, ‘What are policies?’ https://support.google.com/youtube/answer/107383?hl=en&ref_topic=24332, https://support.google.com/youtube/answer/107383?hl=de&ref_topic=24332 accessed 31 December 2023.
[8] Blocking can thus take place on the basis of so-called content ID claims as well as on the basis of an infringement of state copyright law.
[9] Schillmöller and Doseva (n 5), (2022) 25(3) MMR (Multimedia und Recht) 181 f.
[10] T Hess and H Waltermann, ‘Upload-Filter für Content’ (2019) 16(2) MedienWirtschaft 16, 19 f https://web.archive.org/web/20220225020059id_/https://www.beck-elibrary.de/10.15358/1613-0669-2019-2-16.pdf accessed 31 December 2023; L Solomon, ‘Fair users or content abusers’ (2015) 44(1) Hofstra L. Rev. 237, 256. Here, the rightsholder has the possibility to view data on the use of the video. There is no further information about the procedure. Google denies a legitimate reason for objection if the video is not monetized (YouTube, ‘Dispute a Content ID claim’ https://support.google.com/youtube/answer/2797454?hl=en&ref_topic=9282678#zippy=%2Coptionen-f%C3%BCr-den-anspruchsteller accessed 31 December 2023). However, the link to this notice makes it clear that the monetization of the uploaded video is at issue here, but not a possible restriction of the right to object in the case of monitoring: YouTube, ‘What is copyright?’ https://support.google.com/youtube/answer/2797466#refrained&zippy=%2Cmissverst%C3%A4ndnis-nr-wenn-du-angibst-dass-deine-inhalte-nicht-kommerziellen-zwecken-dienen-kannst-du-jegliche-inhalte-verwenden accessed 31 December 2023.
[11] This is specifically done by concluding a copyright licensing agreement, see Perel and Elkin-Koren (n 4), (2016) 19(3) Stan. Tech. L. Rev. 473, 512; Y Lev-Aretz, ‘Second Level Agreements’ (2012) 45(1) AKRON L. REV. 137, 152.
[12] The ability to monitor and monetize infringement redresses the ‘value gap’ between what YouTube pays for monetized content and what services such as Spotify or Pandora, which license content directly from rightsholders, pay, see Bloch-Wehba (n 3), (2020) 53(1) Cornell International Law Journal 41, 64. – Over the years, it has become clear that monetization is the preferred method: in the second half of 2021, monetization was selected in 90% of cases for a total volume of 759,540,199 Content ID claims In less than one percent of the cases in which a Content ID claim was made, an objection (dispute) was raised at all, see YouTube, ‘Copyright Transparency Report H2 2021’ 3, 10 f https://storage.googleapis.com/transparencyreport/report-downloads/pdf-report-22_2021-7-1_2021-12-31_en_v1.pdf accessed 31 December 2023. In 2017 for example, monetization was chosen in 90% of all cases – in the music industry even in 95%. This has led to payments on the part of YouTube amounting to around 3 billion dollars, cf: Grosse Ruse-Kahn (n 5), (2020) PIJIP Research Paper Series 51, 1, 4.
[13] Schillmöller and Doseva (n 5), (2022) 25(3) MMR (Multimedia und Recht) 181, 183 fn 25; YouTube, ‘Monetization during Content ID disputes’ https://support.google.com/youtube/lanswer/7000961?hl=en&ref_topic=9282678 accessed 31 December 2023. If the uploader remains inactive for five days, the revenue is paid to the rightsholder.
[14] YouTube (n 13), ‘Monetization during Content ID disputes’; YouTube, ‘Appeal a Content ID claim’ https://support.google.com/youtube/answer/12104471 accessed 31 December 2023.
[15] This is subject to strong criticism with regard to the fair design of the procedure (see below para 95-98). Instructive on the problem: Schillmöller and Doseva (n 5), (2022) 25(3) MMR (Multimedia und Recht) 181 f.
[16] Accordingly, YouTube states: ‘The initial objection and complaint will be reviewed by the claimant, as YouTube cannot make ownership decisions We do not know what content is properly licensed and therefore cannot determine when copyright exceptions such as fair use or fair dealing apply’, YouTube, ‘Dispute a Content ID claim’ https://support.google.com/youtube/answer/2797454 accessed 31 December 2023.
[17] ‘The whole process is dictated by the Digital Millennium Copyright Act. […] YouTube also has Content ID, an automated copyright management system. It exists in parallel to the copyright takedown process and allows copyright owners to manage their content at scale on YouTube’, YouTube, ‘Frequently asked questions about copyright’ https://support.google.com/youtube/answer/2797449?hl=en accessed 31 December 2023. The Notice and Takedown procedure must always be provided by YouTube due to other liability, K L Zawada ‘The Emergence and Development of Content ID in Light of User-generated Law’ (2017) How Deep is your Law?, 5th International Conference of PhD Students and Young Researchers Conference Papers 438.
[18] Such a deactivation request leads to a copyright warning, Youtube ‘copyright strike basics’ https://support.google.com/youtube/answer/2814000#zippy=%2Cfolgen-einer-urheberrechtsverwarnung accessed 31 December 2023.
[19] Grosse Ruse-Kahn (n 5), (2020) PIJIP Research Paper Series 51, 1, 4 at fn 7, according to which Content ID claims have outnumbered copyright takedowns by a ratio of 50 to 1 since 2014. In 2017, over 98% of copyright infringements were claimed via Content ID instead of notice-and-takedown: Google (n 5), ‘How Google Fights Piracy’ 23 f.
[20] See YouTube, ‘Requirements for counter notifications’ https://support.google.com/youtube/answer/6005919?hl=en&ref_topic=9282678 accessed 31 December 2023.
[21] Schillmöller and Doseva (n 5), (2022) 25(3) MMR (Multimedia und Recht) 181, 183.
[22] YouTube (n 12), ‘Copyright Transparency Report H2 2021’, 1.
[23] Regarding Contend ID, YouTube describes the amount of automation as ‘high’. According to the platform, 98% of all copyright actions on YouTube are handled through content ID, see YouTube (n 12), ‘Copyright Transparency Report H2 2021’, 1.
[24] This level, for instance, addresses creators in the YouTube’s partner program and ‘any channel that’s filled out the copyright management tools application and shown a need for an advanced rights management tool’: YouTube (n 12), ‘Copyright Transparency Report H2 2021’, 2.
[25] This is the case since October 2021. Previously, the ‘Copyright match Tool’ was only available to the users of YouTube’s partner program, see YouTube (n 12), ‘Copyright Transparency Report H2 2021’, 5.
[26] As far as the ‘Copyright Match Tool’ is concerned, YouTube describes the level of automation as ‘medium’: YouTube (n 12), ‘Copyright Transparency Report H2 2021’, 1. However, the meaning of this classification as well as the precise difference between the technical capabilities of the Copyright Match Tool, on the one hand, and Content ID, on the other, remains opaque.
[27] YouTube (n 12), ‘Copyright Transparency Report H2 2021’, 5.
[28] As of July 2015, more than 8,000 ‘partners’ were using the Content ID tool, Zawada (n 17), (2017) How Deep is your Law?, 5th International Conference of PhD Students and Young Researchers Conference Papers 438, 445; Google (n 5), ‘How Google Fights Piracy’ 18. As of November 2018, more than 9.000 ‘partners’ were using Content ID, Google (n 5), ‘How Google Fights Piracy’ 13.
[29] As of November 2018, there were already more than 80 million reference files on Google's servers: Google (n 5), ‘How Google Fights Piracy’ 25.
[30] This is information for managing copyrights The information consists of the reference file, metadata, ownership notices, and established policies: YouTube, ‘Create an asset’ https://support.google.com/youtube/answer/3011552?hl=en&ref_topic=3011550 accessed 31 December 2023. In addition, the information has to indicate the (local) scope of the exclusive rights: YouTube, ‘Using Content ID’ https://support.google.com/youtube/answer/3244015?hl=en accessed 31 December 2023; cf Schillmöller and Doseva (n 5), (2022) 25(3) MMR (Multimedia und Recht) 181, 182.
[31] To increase the number of hits, YouTube recommends ‘full length’ references: YouTube, ‘What is a reference?’ https://support.google.com/youtube/answer/107004?hl=en accessed 31 December 2023; YouTube, ‘Best practices for references’ https://support.google.com/youtube/answer/107008 accessed 31 December 2023. – For more information on using a CSV template: see YouTube, ‘Deliver content using spreadsheet templates’ https://support.google.com/youtube/answer/6066171 accessed 31 December 2023; for the DDEX feed, see YouTube, ‘Using the YouTube DDEX feed’ https://support.google.com/youtube/topic/3505247 accessed 31 December 2023.
[32] YouTube, ‘Best practices for references’ https://support.google.com/youtube/answer/107008 accessed 31 December 2023.
[33] For this, a user must have at least 1,000 subscribers, the channel (among other requirements) must have a playback time of more than 4,000 hours in the last 12 months: YouTube, ‘YouTube Partner Program overview & eligibility’ https://support.google.com/youtube/answer/72851 accessed 31 December 2023; YouTube, ‘Qualify for Content ID’ https://support.google.com/youtube/answer/1311402 accessed 31 December 2023. Otherwise, content creators must rely on other programs such as the ‘Copyright Match Tool’, the ‘Content Verification Tool’, or even the ‘Copyright Complaint Web Form’. – Cf also Grosse Ruse-Kahn (n 5), (2020) PIJIP Research Paper Series 51, 1, 8 with further references.
[34] YouTube (n 30), ‘Using Content ID’; referring to YouTube (n 33), ‘Qualify for Content ID’.
[35] YouTube (n 33), ‘Qualify for Content ID’.
[36] YouTube, ‘Content eligible for Content ID’ https://support.google.com/youtube/answer/2605065?hl=en accessed 31 December 2023. However, YouTube notes that appeals are monitored ‘continuously’, see YouTube (n 30), ‘Using Content ID’.
[37] Details about the Content ID matching process have been kept secret by Google so far; more precise statements about how the algorithm works are therefore difficult to make, Perel and Elkin-Koren (n 4), (2016) 19(3) Stan. Tech. L. Rev. 473, 514.
[38] Hess and Waltermann (n 10), (2019) 16(2) MedienWirtschaft 16, 18.
[39] E Engstrom and N Feamster, ‘The Limits of Filtering: A Look at the Functionality & Shortcomings of Content Detection Tools’ (2017) Engine, 12 https://static1.squarespace.com/static/571681753c44d835a440c8b5/t/58d058712994ca536bbfa47a/1490049138881/FilteringPaperWebsite.pdf accessed 31 December 2023.
[40] Hess and Waltermann (n 10), (2019) 16(2) MedienWirtschaft 16, 18 f.
[41] T Lester and D Pachamanova, ‘The Dilemma of False Positives: Making Content ID Algorithms More Conducive to Fostering Innovative Fair Use in Music Creation’ (2017) 24(1) UCLA Entertainment Law Review 51, 62 f https://escholarship.org/content/qt1x38s0hj/qt1x38s0hj.pdf?t=ovwl6c accessed 31 December 2023.
[42] Hess and Waltermann (n 10), (2019) 16(2) MedienWirtschaft 16, 19 f. – Examples of hashing algorithms are: (i) terrorist content screening databases. One of them was developed by the EU Internet Forum (including Microsoft, Google, Facebook and Twitter) together with Europol. The program identifies terrorist propaganda content and reports it. The content is then reviewed by staff. This is a staged process based on the ‘human-in-the-loop principle’, see M Monroy, ‘EU-Kommission droht mit „gesetzgeberischen Maßnahmen“ zur Entfernung von Internetinhalten’ (2018) Netzpolitik.Org https://netzpolitik.org/2018/eu-kommission-droht-mit-gesetzgeberischen-massnahmen-zur-entfernung-von-internetinhalten/#netzpolitik-pw.However, the filter can only identify uploads that have already been uploaded once and subsequently deleted, M Monroy, ‘“EU-Internetforum”: Viele Inhalte zu “Extremismus” werden mit Künstlicher Intelligenz aufgespürt’ (2017) Netzpolitik.Org https://netzpolitik.org/2017/eu-internetforum-viele-inhalte-zu-extremismus-werden-mit-kuenstlicher-intelligenz-aufgespuert/ accessed 31 December 2023 – (ii) PhotoDNA (from Microsoft): used by Google, Facebook, and Twitter, among others, to identify child pornography content.
[43] Lester and Pachamanova (n 41), (2017) 24(1) UCLA Entertainment Law Review 51, 63 f.
[44] Solomon (n 10), (2015) 44(1) Hofstra L. Rev. 237, 238 (2015). – On the distinction of fingerprinting from so-called watermarking, see D Milano, Content control: Digital Watermarking and Fingerprinting (Rhozet 2013) 2 ff; Jacques S, Garstka K, Hviid M and Street J, ‘An empirical study of the use of automated anti-piracy systems and their consequences for cultural diversity’ (2018) 15(2) Script-Ed 277, 287 at fn 28.
[45] Hess and Waltermann (n 10), (2019) 16(2) MedienWirtschaft 16, 19; Solomon (n 10), 44(1) Hofstra L. Rev. 237, 256 (2015).
[46] E Engstrom and N Feamster (n 39) 13.
[47] Cf Google (n 5), ‘How Google Fights Piracy’ 25; H Grosse Ruse-Kahn, ‘Global Content Protection through Automation’ (2018) 49(9) IIC (International Review of Intellectual Property and Competition Law) 1017, 1018 f; G Nolte, ‘Three Theses on the Current Debate on Liability and Distributive Justice in Hosting Services with User-Generated Content (the so-called “Value Gap” Debate)’ (2017) 61(4) ZUM (Zeitschrift für Urheber- und Medienrecht) 304, 309; Schillmöller and Doseva (n 5), (2022) 25(3) MMR (Multimedia und Recht) 181, 182.
[48] Hess and Waltermann (n 10), (2019) 16(2) MedienWirtschaft 16, 19; Solomon (n 10), (2015) 44(1) Hofstra L. Rev. 237, 256.
[49] Engstrom and Feamster (n 39) 14.
[50] Ibid.
[51] The scope of Content ID technology is nevertheless limited in the context of live streaming: For example, the content must be ‘time-sensitive live content’ (such as a sporting event) that guarantees a ‘high probability that users will live stream copies of your content’. Also, the rights ownership must be global and exclusive. In the case of live streaming, the sanctioning is naturally different from the so-called standard matching of classic uploads: A warning message is first displayed to the streamer. The live stream is then replaced by a standard image without sound and finally interrupted, see YouTube, ‘Use Content ID matching on live streams’ https://support.google.com/youtube/answer/9896248?hl=en accessed 31 December 2023.
[52] Cf Lester and Pachamanova (n 41), (2017) 24(1) UCLA Entertainment Law Review 51, 65.
[53] Lester and Pachamanova (n 41), (2017) 24(1) UCLA Entertainment Law Review 51, 65.
[54] ‘The True Positive Rate’ (TPR) (‘sensitivity’) describes the ratio of ‘true positives’ to ‘true positives’ plus ‘false negatives’ and thus the probability that the algorithm will find infringing content. Accordingly, the ‘sensitivity’ expresses the percentage of all infringing cases that the algorithm identifies as infringing.
‘The True Negative Rate’ (TNR) (‘specificity’) indicates the ratio of ‘true negatives’ to ‘true negatives’ plus ‘false positives’ and, therefore, the probability that the algorithm identifies non-infringing content as non-infringing. The ‘specificity’ stands for the percentage of all non-infringing cases that the algorithm identifies as non-infringing.
‘The Positive Predictive Value’ (PPV) (‘precision’) describes the ratio of ‘true positives’ to ‘true positives’ plus ‘false positives’ and, therefore, the probability that content classified as infringing is actually infringing. ‘Precision’ indicates the percentage of all content identified as infringing that is actually infringing.
‘The Negative Predictive Value’ (NPV) expresses the ratio of true negatives to true negatives plus false negatives, and, consequently, the probability that content classified as non-infringing is actually non-infringing. This paraphrases a percentage of all content identified as non-infringing that is actually non-infringing, Lester and Pachamanova (n 41), (2017) 24(1) UCLA Entertainment Law Review 51, 65 f.
[55] This applies to Content ID claims, see C Muller, ‘Setting the Record Straight’ https://blog.youtube/news-and-events/setting-record-straight/ accessed 31 December 2023; referencing Nolte (n 47), (2017) 61(4) ZUM (Zeitschrift für Urheber- und Medienrecht) 304, 309.
[56] Cf the study by Gray and Suzor (n 5), (2020) 7(1) Big Data & Society 1, 4.
[57] Google (n 5), ‘How Google Fights Piracy’ 27. In addition, melodies and compositions can now be recognized, J P Titlow, ‘Youtube is using AI to police copyright to the tune of $2 billion in payouts’ https://www.fastcompany.com/4013603/youtube-is-using-ai-to-police-copyright-to-the-tune-of-2-billion-in-payouts accessed 31 December 2023; Nolte (n 47), (2017) 61(4) ZUM (Zeitschrift für Urheber- und Medienrecht) 304, 309.
[58] R Andrea, ‘No Safe Harbor: YouTube’s Content ID and Fair Use’ (2020) Boston College Intellectual Property & Technology Forum 1, 5.
[59] Assets are composed of the reference file, metadata, ownership information, and established policies, Google, ‘Asset erstellen’ https://support.google.com/youtube/answer/3011552?hl=de&ref_topic=3011550 accessed 31 December 2023.
[60] YouTube, ‘Fix reference overlaps’ https://support.google.com/youtube/answer/3022604?hl=en&ref_topic=3013248 accessed 31 December 2023. This is the case ‘when two reference files have segments that collect audio, video, or audiovisual content’.
[61] Ibid.
[62] Ibid. – For the rest, YouTube refers to a mutually agreeable solution (‘If an asset has two or more rightsholders, you must resolve the conflict together with other holders’): YouTube (n 60), ‘Fix reference overlaps’.
[63] YouTube has determined that the least restrictive is when none of the policies (monetize, watch, block, or disable) are in place, followed by monetize. More restrictive is watching, then blocking, and finally disabling: YouTube, ‘How policies are applied’ https://support.google.com/youtube/answer/3369929 accessed 31 December 2023.
[64] This is the case, for example, when no rights information has been provided in a particular country, YouTube, ‘How policies are applied’ https://support.google.com/youtube/answer/3369929 accessed 31 December 2023.
[65] In the latter case, if both rightsholders have chosen the same legal consequence, revenue is withheld for music assets until an appeal has been decided; if it is not a music asset, the legal consequence of monitoring applies to both rightsholders, albeit only after the conflict has been resolved: YouTube (n 64), ‘How policies are applied’.
[66] Google (n 5), ‘How Google Fights Piracy’ 28.
[67] YouTube, ‘Update: Improving Content ID for creators’ https://blog.youtube/news-and-events/update-improving-content-id-for-creators/ accessed 31 December 2023.
[68] Gray and Suzor (n 5), (2020) 7(1) Big Data & Society 1, 4. Post-publication means that the video is ‘uploaded’ to the YouTube site but not retrievable. In contrast, blocking can also occur before ‘upload’ (publication). According to Gray and Suzor, however, there is no data on this from YouTube.
[69] By its own account, YouTube/Google invested more than $100 million in 2018 in the development of Google (n 5), ‘How Google Fights Piracy’ 27. In addition, there are maintenance costs: D L Burk, ‘Algorithmic Fair Use’ (2019) 86(2) University of Chicago Law Review 283, 289. Conversely, Google retains 45% of advertising revenue, see Zawada (n 17), (2017) How Deep is your Law?, 5th International Conference of PhD Students and Young Researchers Conference Papers 438, 442.
[70] See already above para 22 fn 67.
[71] For example, with respect to reducing the response time to a complaint in the Content ID process, or with respect to the possibility of a ‘direct complaint’: YouTube, ‘Accelerated Content ID & Complaint Process’ https://support.google.com/youtube/thread/171619847 accessed 31 December 2023.
[72] This aspect applies whether or not the accusation of uploading copyrighted material is true: Burk (n 69), (2019) 86(2) University of Chicago Law Review 283, 289.
[73] See B Laukemann, ‘Private Rechtsdurchsetzung zwischen (digitaler Selbsthilfe) und gerichtlichem Rechtsschutz’ (2022) 8(3) ZfPW (Zeitschrift für die gesamte Privatrechtswissenschaft) 357, 380.
[74] Cf Burk (n 69), (2019) 86(2) University of Chicago Law Review 283, 289. Burk also believes that, in principle, costs do not disappear, but are only ever redistributed; ibid 293.
[75] YouTube, ‘What are policies?’ https://support.google.com/youtube/answer/107383?hl=en&ref_topic=24332 accessed 31 December 2023.
[76] Ibid.
[77] Grosse Ruse-Kahn (n 5), (2020) PIJIP Research Paper Series 51, 1, 10 at fn 44.
[78] Cf Bloch-Wehba (n 3), (2020) 53(1) Cornell International Law Journal 41, 42, 51; see also Y Nahmias and M Perel, ‘The oversight of content moderation by AI: Impact assessment and their limitations’ (2021) 58(1) Harvard Journal on Legislation 145, 171: ‘[…] the organized practice of screening online content based on the characteristics of the website, its targeted audience, and jurisdictions of user-generated content to determine whether such content is appropriate’.
[79] J Grimmelmann, ‘The Virtues of Moderation’ (2015) 17 YALE J.L. & TECH. 42, 47, differentiating between hard and soft moderation, defining ‘moderation’ as ‘the governance mechanisms that structure participation in a community to facilitate cooperation and prevent abuse’.
[80] Vividly expressed by Bloch-Wehba (n 3), (2020) 53(1) Cornell International Law Journal 41, 78: ‘[…] the new wave of Internet regulation and the emergence of “voluntary” filtering illustrates the risk that governments will informally pressure platforms to adopt limitations on speech’.
[81] With respect to copyright enforcement: J E Cohen, Between Truth and Power: The Legal Constructions of Informational Capitalism (Oxford University Press 2019) 123 f. As a reaction to unsatisfying copyright enforcement based on notice and takedown, commercial copyright owners and providers of user-generated content (UGC) services entered, in 1997, into the ‘UGC Principles’, a non-binding set of principles calling, inter alia, for to use of ‘effective content identification technology’ – as, for example, was the case, in 2007, with YouTube’s fingerprinting technology (Content ID). Correspondingly, large rightsholders developed automated mechanisms to detect, track, and report online infringement and generate takedown requests, see Bloch-Wehba (n 3), (2020) 53(1) Cornell International Law Journal 41, 63 f, referring also to N P Suzor, Lawless: The Secret Rules that Govern our Digital Lives (Cambridge University Press 2019) 76-78.
[82] See Art 17(8) of the EU Directive 2019/790 of 17 April 2019 on copyright and related rights in the Digital Single Market and amending Directives 96/9/EC and 2001/29/EC, Official Journal L 130/92: ‘The application of this Article shall not lead to any general monitoring obligation’. In its recent judgment regarding the annulment of Art 17(4) lit b) and c) of the DSM Directive, the CJEU stated that ‘a filtering system which might not distinguish adequately between unlawful content and lawful content, with the result that its introduction could lead to the blocking of lawful communications, would be incompatible with the right to freedom of expression and information […] and would not respect the fair balance between that right and the right to intellectual property’. Regarding the prohibition of a general monitoring obligation under Art 17 DSM Directive, the service providers ‘cannot be required to prevent the uploading and making available to the public of content which, in order to be found unlawful, would require an independent assessment of the content by them in the light of the information provided by the rightholders and of any exceptions and limitations to copyright’: CJEU, 26 April 2022, C‑401/19 – Poland v Parliament and Council, ECLI:EU:C:2022:297, para 86, 90.
[83] See Art 15(1) of the Directive 2000/31/EC of 8 June 2000 on certain legal aspects of information society services, in particular electronic commerce, in the Internal Market, Official Journal L 178/1.
[84] Regulation on a Single Market for Digital Services (Digital Services Act) and amending Directive 2000/31/EC, COM (2020) 825 final. While Art 14(1) and 15 of the E-Commerce Directive prohibit general monitoring, Member States may oblige host providers to ‘detect and prevent certain types of illegal activities’ (recital 48) and impose ‘monitoring obligations in a specific case[s]’ (recital 47). To this, see K Kaesling, ‘Privatising Law Enforcement in Social Networks: A Comparative Model Analysis’ (2018) (3) Erasmus Law Review 151, 154 f. In sharp contrast to 47 U.S.C. § 230, there is no Good Samaritan Privilege under the E-Commerce Directive. With regard to the CJEU ruling in L’Oréal, denying the liability privilege to be applied to platforms when an ‘active role’ (CJEU, 12 July 2011, C‑324/09 – L’Oréal v eBay, ECLI:EU:C:2011:474, para 113 and CJEU, 23 March 2010, C-236/08 und C-238/08 – Google France v Luis Vuitton Malletier, ECLI:EU:C:2010:159, para 120) and Art 14(1) E-Commerce Directive referring to the threshold of actual and constructive knowledge, platforms run the serious risk of liability for user generated content. – Art 6 DSA Regulation now introduces a Good Samaritan Clause. This provision applies to measures taken in accordance with EU law as well as to a platform’s own terms and conditions. Therefore, Art 6 DSA Regulation might ‘not oblige platforms to monitor but rather invite them to do so’, N Gielen and S Uphues, ‘Digital Markets Act und Digital Services Act’ (2021) 32(14) EuZW (Europäische Zeitschrift für Wirtschaftsrecht) 627, 632, thus incentivizing platforms to perform more removals on their own initiative, see A Kuczerawy, ‘The Good Samaritan that wasn’t: voluntary monitoring under the (draft) Digital Services Act’ https://verfassungsblog.de/good-samaritan-dsa/ accessed 31 December 2023.
[85] See 17 U.S.C. § 512(m) DMCA.
[86] Cf also Bloch-Wehba (n 3), (2020) 53(1) Cornell International Law Journal 41, 63 with further references.
[87] Neither the German Netzwerkdurchsetzungsgesetz (NetzDG) nor the EU Code of Conduct on Countering Illegal Hate Speech (this Code of Conduct is a self-regulatory, joint act of internet service companies initiated by the European Union: European Commission, ‘Tackling online disinformation: Commission proposes an EU-wide Code of Practice’ https://ec.europa.eu/commission/presscorner/detail/en/IP_18_3370 accessed 31 December 2023) explicitly demand proactive automated measures. In doing so, the short time frames (less than 24 hours) and the mere scale of affected content prompt platforms to take such measures. Similar see Bloch-Wehba (n 3), (2020) 53(1) Cornell International Law Journal 41, 70-72.
[88] See, eg, CJEU, 3 October 2019, C-18/18 – Glawischnig-Piesczek, ECLI:EU:C:2019:821, para 46, imposing the obligation to prevent defamatory content of equivalent nature and explicitly stating that this is not ‘[…] an excessive obligation being imposed on the host provider, in so far as the monitoring of and search for information […] does not require the host provider to carry out an independent assessment, since the latter has recourse to automated search tools and technologies’.
[89] See Commission Recommendation (EU) 2018/334 on measures to effectively tackle illegal content online, 1 March 2018, declaring that ‘in addition to notice-and-action mechanisms, proportionate and specific proactive measures taken voluntarily by hosting service providers, including by using automated means in certain cases, can also be an important element in tackling illegal content online, without prejudice to Article 15(1) of Directive 2000/31/EC’, C [2018] 1177 final https://eur-lex.europa.eu/legal-content/EN/TXT/PDF/?uri=CELEX:32018H0334&from=EN.
[90] Cf Art 14(1) of the E-Commerce Directive 2000/31/EC; Art 3-5 DSA Regulation, and the 47 U.S.C. § 230. – Art 14(1) of the E-Commerce Directive is described as the ‘European equivalent’ to 47 U.S.C. § 230, see Gielen and Uphues (n 84), (2021) 32(14) EuZW (Europäische Zeitschrift für Wirtschaftsrecht) 627, 632.
[91] K Klonick, ‘Why the History of Content Moderation Matters’ https://www.techdirt.com/articles/20180129/21074939116/whyhistory-content-moderation-mattersshtml accessed 31 December 2023; Bloch-Wehba (n 3), (2020) 53(1) Cornell International Law Journal 41, 51.
[92] See S.D. Cal. (USA), 25 April 2012, 11cr0938 JM – United States v Green, 857 F. Supp. 2d 1015, quoting testimony of Don Colcolough, AOL’s Director of Investigations and Cyber Security.
[93] See Bloch-Wehba (n 3), (2020) 53(1) Cornell International Law Journal 41, 58, referring to the PhotoDNA technique developed by Microsoft. This tool, which is licensed for free to technology companies and law enforcement, can match the hash values of photos or videos uploaded by individual users against a database of hash values of other photos or videos containing illegal images of child sexual abuse. Cf also J Kosseff, ‘Private Computer Searches and the Fourth Amendment’ (2018) 14(2) I/S: A Journal of Law and Policy 187, 209.
[94] K Klonick, ‘The New Governors: The People, Rules and Process Governing Online Speech’ (2018) 131(6) Harvard Law Review 1598, 1635; R Gorwa, R Binns and C Katzenbach, ‘Algorithmic content moderation’, (2020) 7(1) Big Data & Society 1, 7, tabulating the use of human moderators in the automated moderation systems of various platforms.
[95] N Elkin-Koren and M Perel, ‘Separation of Functions for AI: Restraining Speech Regulation by Online Platforms’ (2020) 24(3) Lewis & CLARK L. REV. 857, 878.
[96] Gorwa, Binns and Katzenbach (n 94), (2020) 7(1) Big Data & Society 1, 6; Meta, ‘How we review content’ https://about.fb.com/news/2020/08/how-we-review-content/ accessed 31 December 2023. To Wikipedia’s ‘Huggle’ bot, human Wikipedia moderators by prioritizing suspicious content for human content review: Content with the highest likelihood of abusive editing is thus reviewed first, see E Katsh and O Rabinovich-Einy, Digital Justice: Technology and the Internet of Disputes (Oxford University Press 2017) 125.
[97] Google’s Perspective technology calculates a score about the ‘impact a comment might have on a conversation’. This score could be used by content moderators or provide real-time ‘feedback’ to the posting user about their content; Gorwa, Binns and Katzenbach (n 94), (2020) 7(1) Big Data & Society 1, 9.
[98] Meta (n 96), ‘How we review Content’: ‘Our AI systems automate decisions for certain areas where content is highly likely to be violating’.
[99] Elkin-Koren and Perel (n 95), (2020) 24(3) Lewis & CLARK L. REV. 857, 879.
[100] Differentiation according to Klonick (n 94), (2018) 131(6) Harvard Law Review 1598, 1635 f; similarly: Katsh and Rabinovich-Einy (n 96) 52; on Facebook: see Meta, ‘How Facebook uses super-efficient AI models to detect hate speech’ https://ai.facebook.com/blog/how-facebook-uses-super-efficient-ai-models-to-detect-hate-speech/ accessed 31 December 2023; on Airbnb: AIRBNB, ‘Scoring the user to prevent “suspicious” activity before it occurs: What Does It Mean When Someone’s ID Has Been Checked?’ https://www.airbnb.com/help/article/2356/what-does-it-mean-when-someones-id-has-been-checked accessed 31 December 2023; cf R Van Loo, ‘Federal Rules of Platform Procedure’ (2021) 88(4) University of Chicago Law Review 829, 845.
[101] It is by no means mandatory to focus on the time of publication, nor is it primarily descriptive in nature. It is also conceivable, for example, to differentiate between action by the platform before the infringed party becomes aware of the content (proactive) or only in response to a report (such as flagging) of the incriminated content by users or other actors (reactive). Nevertheless, in the case of content relevant to the law of expression, the time of publication (as the initial possibility of third parties taking notice) is sensitive to fundamental rights and thus particularly relevant, see Klonick (n 94), (2018) 131(6) Harvard Law Review 1598, 1635 f.
[102] As regards the technical functioning of fingerprinting, see already above para 13-15.
[103] Gorwa, Binns and Katzenbach (n 94), (2020) 7(1) Big Data & Society 1, 5: ‘There are also systems which blur the lines between the two. For instance, a series of photos taken milliseconds apart might be something that a matching system ought to classify as similar, even though the underlying images are different and therefore technically not matches Facial recognition technologies may serve the dual purpose of inducing patterns from many faces and matching particular faces belonging to the same person. In these cases, the distinction between identity-matching and classification is a matter of degree’.
[104] Gorwa, Binns and Katzenbach (n 94), (2020) 7(1) Big Data & Society 1, 4 f; Elkin-Koren and Perel (n 95), (2020) 24(3) Lewis & CLARK L. REV. 857, 885 f.
[105] N Elkin-Koren, ‘Contesting algorithms: Restoring the public interest in content filtering by artificial intelligence’, (2020) 7(2) Big Data & Society 1, 5.
[106] Gorwa, Binns and Katzenbach (n 94), (2020) 7(1) Big Data & Society 1, 4 f.
[107] Elkin-Koren (n 105), (2020) 7(2) Big Data & Society 1, 5; Gorwa, Binns and Katzenbach (n 94), (2020) 7(1) Big Data & Society 1, 4 f.
[108] N Elkin-Koren (n 105), (2020) 7(2) Big Data & Society 1, 55 f.
[109] In this regard: Elkin-Koren and Perel (n 95), (2020) 24(3) Lewis & CLARK L. REV. 857, 885 f.
[110] On AI and data protection, see M Valkanova, ‘Trainieren von KI-Modellen’ in M Kaulartz and T Braegelmann (ed), Rechtshandbuch Artificial Intelligence und Machine Learning (2020) 336 ff.
[111] Elkin-Koren and Perel (n 95), (2020) 24(3) Lewis & CLARK L. REV. 857, 862, 876.
[112] For example, Facebook incorporates data from its users who identify themselves to other services using Facebook accounts into the Facebook Social Graph, Elkin-Koren and Perel (n 95), (2020) 24(3) Lewis & CLARK L. REV. 857, 862, 876; J-C Plantin, C Lagoze, P N Edwards and C Sandvig, ‘Infrastructure Studies Meet Platform Studies in the Age of Google and Facebook’ (2018) 20(1) New Media & Society 293, 304; A Helmond, ‘The Platformization of the Web: Making Web Data Platform Ready’ (2015) 1(2) Social Media and Society 1 ff. Airbnb has patented an AI technology that screens users’ online activities outside the platform to rank individual users in a ‘trustworthiness score’ or a ‘compatibility score’. This is done based on user behavior and ‘personality trait metrics’ on the basis of a scoring system. Deploying these tools serves to prevent ‘suspicious’ activities before they actually occur, see Van Loo (n 100), (2021) 88(4) University of Chicago Law Review 829, 844.
[113] Elkin-Koren and Perel (n 95), (2020) 24(3) Lewis & CLARK L. REV. 857, 862, 886.
[114] This part (1.1.2.4) Excursus: The Content Moderation of the Meta Group as a Prototype of Individualized Process Design) was written independently by Helena Müller, former research assistant at the chair of Prof. Dr. Björn Laukemann.
[115] According to Dave Willners, author of the first draft of the Facebook Community Standards, when he joined the company in 2009, all content moderation was based on one page of internal rules These were applied globally. There was also very limited guidance on content moderation, see Klonick (n 94), (2018) 131(6) Harvard Law Review 1598, 1630 f; see also A Heldt, Intensivere Drittwirkung (Mohr Siebeck 2023) 198 f.
[116] Meta Transparency Center, ‘How technology detects violations’ https://transparency.meta.com/de-de/enforcement/detecting-violations/technology-detects-violations/ accessed 31 December 2023.
[117] The cross-check system received significant public attention in the wake of the Facebook paper scandal. The cross-check system was recently the subject of a Policy Advisory Opinion by Meta’s Oversight Board, see Oversight Board, 6 December 2022, 2021-002-FB-PAO – Meta’s Cross-Check Program. The process also goes by the name ‘X-Check’, see J Horwitz, ‘Facebook Says Its Rules Apply to All. Company Documents Reveal a Secret Elite That’s Exempt’ https://www.wsj.com/articles/facebook-files-xcheck-zuckerberg-elite-rules-11631541353 accessed 31 December 2023.
[118] Oversight Board, 6 December 2022, 2021-002-FB-PAO – Meta’s Cross-Check Program, Part 1, 4.
[119] This was also harshly criticized by the Oversight Board (decision of 6 December 2022, 2021-002-FB-PAO – Meta’s Cross-Check Program, Part 1, 3): ‘While Meta told the Board that cross-check aims to advance Meta’s human rights commitments, we found that the program appears more directly structured to satisfy business concerns. The Board understands that Meta is a business, but by providing extra protection to certain users selected largely according to business interests, cross-check allows content which would otherwise be removed quickly to remain up for a longer period, potentially causing harm’.
[120] Oversight Board, 6 December 2022, 2021-002-FB-PAO – Meta’s Cross-Check Program, para 12, 18; Meta Transparency Center, ‘Reviewing high-impact content accurately via our cross-check system, 12 May 2023’ https://transparency.fb.com/enforcement/detecting-violations/reviewing-high-visibility-content-accurately/ accessed 31 December 2023.
[121] Oversight Board, 6 December 2022, 2021-002-FB-PAO – Meta’s Cross-Check Program, para 12.
[122] Oversight Board, 6 December 2022, 2021-002-FB-PAO – Meta’s Cross-Check Program, para 11-13.
[123] Oversight Board, 6 December 2022, 2021-002-FB-PAO – Meta’s Cross-Check Program, para 16.
[124] The ‘Spirit of Policy Allowance’ is a well-known example. Structurally, this equates to a teleological interpretation of the Community Standards Based on these teleological considerations, a dispensation from the regular moderation decision can be granted. In the regular moderation process (‘at scale’), this possibility does not exist. Regular moderators must decide strictly on the basis of the wording of the policies (‘letter of the policy’); Oversight Board, 17 June 2022, 2022-001-FB-UA – Knin Cartoon, Part 8.1, 17.
[125] In this context, the so-called ‘Newsworthiness Allowance’ has been the subject of much discussion. According to press reports, the soccer player Neymar is considered to be a beneficiary of this exemption, see Horwitz (n 117), ‘Facebook Says Its Rules Apply to All. Company Documents Reveal a Secret Elite That’s Exempt’.
[126] In light of this, Meta’s Oversight Board criticizes the divergence in content between publicly available policies and non-publicly available internal moderation standards and policies, see Oversight Board, 28 January 2021, 2020-005-FB-UA – Nazi Quote, Key Findings and Part 8.1.
[127] Oversight Board, 6 December 2022, 2021-002-FB-PAO – Meta’s Cross-Check Program, para 18.
[128] For example, the ‘Early Response Team’, see Oversight Board, 6 December 2022, 2021-002-FB-PAO – Meta’s Cross-Check Program, charts at 14, 21.
[129] Under the content-based General Secondary Response (GSR) process, this is only the case under certain conditions; see below para 40-44 for both procedural modalities specific to Meta’s General Secondary Review System: Oversight Board, 15 September 2022, 2022-005-FB-UA – Mention of the Taliban in News Reporting, Part 6, 9.
[130] Oversight Board, 9 January 2023, 2022-013-FB-UA – Iran Protest Slogan, Part 8.1 II., 14: ‘[...] Cross-check [...] enable[s] users' content to be reviewed on escalation prior to removal’. See Oversight Board, 6 December 2022, 2021-002-FB-PAO – Meta's Cross-Check Program, Part 1, 4. This applies without limitation in terms of timing only to the ERSR process: Oversight Board, 6 December 2022, 2021-002-FB-PAO – Meta's Cross-Check Program, para 47; see on this point below para 50-52.
[131] Oversight Board, 6 December 2022, 2021-002-FB-PAO – Meta's Cross-Check Program, para 19, 24-39. In technical terms, this is done by tagging the data subjects, see para 24.
[132] See below para 73 for more information.
[133] The cross-check procedure only provides information about the internal categorization of the Meta Group. The exact identity of the beneficiary users is unclear. Exemplary named are: Users associated with ‘significant world events’, members of groups of people disproportionately affected by overenforcement, and media organizations, see: Oversight Board, 6 December 2022, 2021-002-FB-PAO – Meta's Cross-Check Program, para 24 f. For example, the Al Jazeera news network and Donald Trump are part of the cross-check system: Oversight Board, 14 September 2021, 2021-009-FB-UA – Shared Al Jazeera Post, Part 6; also Oversight Board, 5 May 2021, 2021-001-FB-FBR – Former President Trump's suspension, Part 2. In addition, The Oversight Board failed to gain access to the company’s internal cross-check lists in the course of the cross-check process, despite multiple requests to the Meta Group, Horwitz (n 117), ‘Facebook Says Its Rules Apply to All. Company Documents Reveal a Secret Elite That's Exempt’.
[134] Explicitly Oversight Board, 6 December 2022, 2021-002-FB-PAO – Meta's Cross-Check Program, para 24. The categorization of the groups of persons already suggests this.
[135] Oversight Board, 6 December 2022, 2021-002-FB-PAO – Meta's Cross-Check Program, para 20, 40-55.
[136] Oversight Board, 6 December 2022, 2021-002-FB-PAO – Meta's Cross-Check Program, Part I, 3.
[137] The scandal originated from internal documents of the Meta Group (‘Facebook Files’), which were leaked to the Wall Street Journal and the US Senate by whistleblower Frances Haugen. The documents reveal, among other things, that the Meta Group was aware as early as 2019, following an internal investigation, that the platform’s recommendation system was significantly encouraging the spread of hate and disinformation. The documents also made the existence of the cross-check program public knowledge, see the Wall Street Journal’s article series ‘The Facebook Files’ https://www.wsj.com/articles/the-facebook-files-11631713039 accessed 31 December 2023; esp D Seetharaman , J Horwitz and J Scheck, ‘Facebook Says AI Can Enforce Its Rules, but the Company’s Own Engineers Are Doubtful’ https://www.wsj.com/articles/facebook-ai-enforce-rules-engineers-doubtful-artificial-intelligence-11634338184 accessed 31 December 2023. Specific to the cross-check program: M Reuter, ‘Facebook Knew What All Was Going Wrong’ https://netzpolitik.org/2021/facebook-files-facebook-wusste-was-alles-schieflaeuft/ accessed 31 December 2023.
[138] Oversight Board, 6 December 2022, 2021-002-FB-PAO – Meta's Cross-Check Program, para 40.
[139] Oversight Board, 6 December 2022, 2021-002-FB-PAO – Meta's Cross-Check Program, para 43.
[140] Oversight Board, 6 December 2022, 2021-002-FB-PAO – Meta's Cross-Check Program, para 43.
[141] Oversight Board, 6 December 2022, 2021-002-FB-PAO – Meta's Cross-Check Program, para 42 f.
[142] Oversight Board, 6 December 2022, 2021-002-FB-PAO – Meta's Cross-Check Program, para 46.
[143] Oversight Board, 15 September 2022, 2022-005-FB-UA – Mention of the Taliban in News Reporting, Part 8.1, II, 12.
[144] Oversight Board, 6 December 2022, 2021-002-FB-PAO – Meta's Cross-Check Program, para 44 f.
[145] The Oversight Board cites a two to four-day time frame here: Oversight Board, 6 December 2022, 2021-002-FB-PAO – Meta's Cross-Check Program, para 46.
[146] Oversight Board, 6 December 2022, 2021-002-FB-PAO – Meta's Cross-Check Program, para 48.
[147] Oversight Board, 6 December 2022, 2021-002-FB-PAO – Meta's Cross-Check Program, para 46.
[148] Oversight Board, 6 December 2022, 2021-002-FB-PAO – Meta's Cross-Check Program, para 47.
[149] For example, Meta states that approximately 35% of the content that went through the cross-check system had no ‘legal recourse’ to Meta’s Oversight Board, see decision from 6 December 2022, 2021-002-FB-PAO – Meta's Cross-Check Program, para 174.
[150] Oversight Board, 6 December 2022, 2021-002-FB-PAO – Meta's Cross-Check Program, para 174.
[151] Oversight Board, 6 December 2022, 2021-002-FB-PAO – Meta's Cross-Check Program, para 114.
[152] Oversight Board, 6 December 2022, 2021-002-FB-PAO – Meta's Cross-Check Program, para 114.
[153] Oversight Board, 15 September 2022, 2022-005-FB-UA – Mention of the Taliban in News Reporting, Part 2, 5.
[154] Oversight Board, 15 September 2022, 2022-005-FB-UA – Mention of the Taliban in News Reporting, Part 6, 9.
[155] Cf also recently: Oversight Board, 18 December 2023, 2023-054-FB-UA, 2023-055-FB-UA, 2023-056-FB-UA, 2023-057-FB-UA – Goebbels Quote.
[156] Ibid, 9.
[157] According to the board, users who regularly report on the activities of dangerous organizations or individuals are exposed to an increased risk of sanctions. This results from the policy of deleting content related to such dangerous organizations or persons in case of doubt, if it is not quite clear that the content in question is merely a factual report about the events. This leads to such content being removed disproportionately often. In the facts underlying the ’Board's decision, the HIPO ranker did not recognize the weightiness of the content. The subject of the decision was a post by an Indian magazine that reported on the Taliban’s school closures in Afghanistan. The post was blocked by the Meta Group on the basis of its ‘Dangerous Individuals and Organizations policy’. This policy forbids ‘praising’ such ‘dangerous entities’ as terrorist organizations. A human moderator also classified the content as a violation. Against this, the Indian newspaper filed a user appeal, whereupon the post in question was added to the queue of the HIPO proceedings. The content was classified as non-priority by the HIPO ranker. In addition, there was a lack of capacity of Urdu-speaking HIPO moderators. For these reasons, the content fell out of the HIPO system and was adjudicated ‘at scale’. In its decision, the Board criticized (in addition to insufficient staff capacity in the HIPO process) the HIPO ranker’s lack of sensitivity to press coverage under the Dangerous Individuals and Organizations policy. Due to the high relevance of press coverage for freedom of expression, such content was particularly weighty. See: Oversight Board, 15 September 2022, 2022-005-FB-UA – Mention of the Taliban in News Reporting, Part 8.3, III, 16.
[158] Oversight Board, 15 September 2022, 2022-005-FB-UA – Mention of the Taliban in News Reporting, Part 6, III, 10.
[159] Oversight Board, 15 September 2022, 2022-005-FB-UA – Mention of the Taliban in News Reporting, Part 6, III, 9.
[160] Oversight Board, 15 September 2022, 2022-005-FB-UA – Mention of the Taliban in News Reporting, Part 6, III, 9.
[161] Oversight Board, 15 September 2022, 2022-005-FB-UA – Mention of the Taliban in News Reporting, Part 6, III, 9.
[162] Oversight Board, 6 December 2022, 2021-002-FB-PAO – Meta's Cross-Check Program, para 181 f.
[163] Oversight Board, 6 December 2022, 2021-002-FB-PAO – Meta's Cross-Check Program, para 181 f.
[164] Oversight Board, 22 November 2022, 2022-007-IG-MR – UK Drill Music, Part 6, 16.
[165] Oversight Board, 22 November 2022, 2022-007-IG-MR – UK Drill Music, Part 6, 16.
[166] Oversight Board, 6 December 2022, 2021-002-FB-PAO – Meta's Cross-Check Program, para 56; Meta Transparency Center, ‘How we assess reports of content violating local law’ https://transparency.fb.com/data/content-restrictions/ accessed 31 December 2023. If ordered by the state, (potentially) infringing content is also blocked worldwide. According to Meta, this was the case in 14 cases in the period from January to June 2022. Meta generally refers to such orders as ‘extraterritorial jurisdiction’, Meta Transparency Center, ‘Global restrictions’ https://transparency.fb.com/data/content-restrictions/ accessed 31 December 2023.
[167] For instructions on how this works, see J Hörnle, Internet Jurisdiction. Law and Practice (Oxford University Press 2021) 448-450.
[168] Oversight Board, 22 November 2022, 2022-007-IG-MR – UK Drill Music, Part 1, 6. – The Oversight Board had already requested the Meta Group on several occasions to formalize the procedure for handling government requests and to list their number in transparency reports: Oversight Board, 14 September 2021, 2021-009-FB-UA – Shared Al Jazeera Post, Part 10.
[169] Meta justifies this by saying that decisions made in the Escalation process are not made by content moderators.
[170] Oversight Board, 22 November 2022, 2022-007-IG-MR – UK Drill Music, Part 6, 16. This had already been criticized by the Board in an earlier decision, cf Oversight Board, 8 July 2021, 2021-006-IG-UA – Ocalan's Isolation, Part 10.
[171] In the facts underlying the Oversight Board’s decision in Drill Music, the ‘veiled threats analysis’ was applied. In the context of this analysis, it is again considered whether the report is made by the state. Thus, in situations such as this one, where government entities report content that falls within the scope of this exception, there is a de facto double consideration of the government identity of the reporter: Oversight Board, 22 November 2022, 2022-007-IG-MR – UK Drill Music, Part 6, 15.
[172] Oversight Board, 22 November 2022, 2022-007-IG-MR – UK Drill Music, Part 8.1, II b, 26.
[173] Oversight Board, 6 December 2022, 2021-002-FB-PAO – Meta's Cross-Check Program, para 56 f.
[174] See in detail below para 71-73.
[175] The so-called reporter appeals serve this purpose, see: Oversight Board, 6 December 2022, 2021-002-F-PAO – Meta's Cross-Check Program, para 181 f.
[176] Oversight Board, 6 December 2022, 2021-002-F-PAO – Meta's Cross-Check Program, para 16.
[177] Introducing the legal economic foundations of digital platforms: A Engert, ‘Digitale Plattformen‘ (2018) 218(2-4) AcP (Archiv für die civilistische Praxis) 218, 304; J K Mendelsohn, ‘Die “normative Macht” der Plattformen – Gegenstand der zukünftigen Digitalregulierung’ (2021) 24(11) MMR (Multimedia und Recht) 857, 858.
[178] Engert (n 178), (2018) 218(2-4) AcP (Archiv für die civilistische Praxis) 218, 304, 307.
[179] Engert (n 178), (2018) 218(2-4) AcP (Archiv für die civilistische Praxis) 218, 304, 307 f.
[180] Klonick (n 94), (2018) 131(6) Harvard Law Review 1598, 1627.
[181] Elkin-Koren and Perel (n 95), (2020) 24(3) Lewis & CLARK L. REV. 857, 875-879.
[182] Instructive on this point: R Van Loo, (2016) ‘The Corporation as Courthouse’ 33(2) Yale Journal on Regulation 547.
[183] H Askani, Private Rechtsdurchsetzung bei Urheberrechtsverletzungen im Internet (Nomos 2021) 162. For more details, see below para 78-82.
[184] For more details, see: Engert (n 178), (2018) 218(2-4) AcP (Archiv für die civilistische Praxis) 218, 304, 307.
[185] Mendelsohn (n 178), (2021) 24(11) MMR (Multimedia und Recht) 857, 858.
[186] See in detail M Glogowski, Plattformbedingungen (Mohr Siebeck 2022) 3-5.
[187] Furthermore: D Wielsch, ‘Die Ordnungen der Netzwerke. AGB – Code – Community Standards’ in M Eifert and T Gostomzyk (ed), Netzwerkrecht (Nomos 2018) 61 f. Cf also Askani (n 184) 173; Elkin-Koren and Perel (n 95), (2020) 24(3) Lewis & CLARK L. REV. 857, 871.
[188] L Lessig, Code and Other Laws of Cyberspace (Basic Books 1999), 382. Yeung criticizes the fact that the relevance of the design or code of technical environments generally receives too little attention in legal research, and criticizes the fact that too much thought is still given to the classical structure of legal prohibitions and their enforcement (‘command and control’), K Yeung, ‘”Hypernudge”: Big Data as a mode of regulation by design’, (2017) 20(1) Information, Communication & Society 118, 120; see also B Wagner, Global Free Expression – Governing the Boundaries of Internet Content (Springer 2016), 130.
[189] Mendelsohn (n 178), (2021) 24(11) MMR (Multimedia und Recht) 857, 859.
[190] S Katyal, ‘Private Accountability in the Age of Artificial Intelligence’ (2019) 66(1) UCLA Law Review 55, 93 f.
[191] On the technical functioning of microtargeting and the implications of this technology for private autonomy, see M Ebers, ‘§ 3 Regulierung von KI und Robotik’ in M Ebers, C Heinze and B Steinrötter (ed), Künstliche Intelligenz und Robotik (Beck 2020) 75 para 101 ff. For a list of data that Facebook uses to generate personalized ads, see: C Dewey, ‘98 personal data points that Facebook uses to target ads to you’ https://www.washingtonpost.com/news/the-intersect/wp/2016/08/19/98-personal-data-points-that-facebook-uses-to-target-ads-to-you/ accessed 31 December 2023.
[192] Van Loo (n 183), (2016) 33(2) Yale Journal on Regulation 547, 602. On the individualization of the customer relationship: J Taeger and S Kremer, Recht im E-Commerce und Internet (Fachmedien Recht und Wirtschaft 2021) 10.
[193] As a consequence, this score determines whether the customer is served by an internal ‘executive customer relations’ department of the bank or by an external call center: Van Loo (n 183), (2016) 33(2) Yale Journal on Regulation 547, 564 f.
[194] On Amazon, see: H Eidenmüller and G Wagner, Law by Algorithm (Mohr Siebeck 2021) 239 ff; Van Loo (n 183), (2016) 33(2) Yale Journal on Regulation 547, 564-566.
[195] For more details on the procedure, see above para 39-66.
[196] In 2020, the Meta Group stated that this prioritization process was used to automatically decide on content that obviously violates standards. The goal is to create capacities to use human moderators primarily for deciding complex or context-dependent constellations of facts, see: Meta, ‘How we review Content – Prioritization’ https://about.fb.com/news/2020/08/how-we-review-content/ accessed 31 December 2023.
[197] In detail below para 83-89.
[198] For the technical procedure, see above para 1-22.
[199] For the concept and operation of monetization, see above para 2-6.
[200] Also in that regard: Schillmöller and Doseva (n 5), (2022) 25(3) MMR (Multimedia und Recht) 181 f.
[201] G Frosio, ‘Algorithmic Enforcement Online’ in P Torremans (ed), Intellectual Property and Human Rights (4th edn, Kluwer Law International 2020) 24; J Lennartz and V Kraetzig, ‘Filtering fundamental Rights’ https://verfassungsblog.de/filtering-fundamental-rights/ accessed 31 December 2023. On the DSM Directive’s incentives for the use of proactive filtering technologies, see: Elkin-Koren and Perel (n 95), (2020) 24(3) Lewis & CLARK L. REV. 857, 883.
[202] For this, see below para 99-102.
[203] This can be seen not least in Audible Magic’s lobbying for the enshrinement of an obligation to use filtering technologies within the framework of Art 17 DSM Directive, see: Bloch-Wehba (n 3), (2020) 53(1) Cornell International Law Journal 41, 85 f.
[204] Grosse Ruse-Kahn (n 5), (2020) PIJIP Research Paper Series 51, 1, 7.
[205] Mendelsohn (n 178), (2021) 24(11) MMR (Multimedia und Recht) 857, 858.
[206] Elkin-Koren and Perel (n 95), (2020) 24(3) Lewis & CLARK L. REV. 857, 875 f.
[207] Ebers (n 192), 75 para 112.
[208] On echo chambers, see C R Sunstein, #Republic – Divided Democracy in the Age of Social Media (Princeton University Press 2017) 122-124. – The studies on the existence of both phenomena are not clear, on this Ebers (n 192), 75 para 112-114. Critically also J Lüdemann, ‘Warum und wie reguliert man digitale Informationsintermediäre?’ in J Lüdemann and Y Hermstrüwer (ed), Schutz der Meinungsbildung im digitalen Zeitalter (Mohr Siebeck 2021) 15-19. However, studies suggest that the influence of personalized media offerings on opinion formation may be less than previously assumed, see: E Dubois and G Blank, ‘The echo chamber is overstated: the moderating effect of political interest and diverse media’ (2018) 21(5) Information, Communication & Society 729.
[209] Elkin-Koren and Perel (n 95), (2020) 24(3) Lewis & CLARK L. REV. 857, 889. Affirmatively on the influence of personalization algorithms on radicalization processes: S Musa and S Bendett, ‘Islamic Radicalization in the United States – New Trends and a Proposed Methodology for Disruption’ (2010), National Defense University, Washington DC Center for Technology and National Security Policy 17 ff https://appsdtic.mil/sti/pdfs/ADA532696.pdf accessed 31 December 2023. In particular, recent studies on the YouTube algorithm assume that it is even capable of counteracting radicalization, see M Ledwich and A Zautsev, ‘Algorithmic Extremism: Examining YouTube’s Rabbit Hole of Radicalization’, https://arxiv.org/abs/1912.11211 accessed 31 December 2023. M Wolfowicz, D Weisburd and B Hasisi, ‘Examining the interactive effects of the filter bubble and the echo chamber on radicalization’ (2023) 19(1) Journal of Experimental Criminology, 119.
[210] T Gillespie vividly refers to this algorithmic separation and structuring process as ‘calculated publics’: ‘The Relevance of Algorithms’ in T Gillespie, P Boczkowski and K Foot (ed), Media Technologies (The MIT Press 2014) 188.
[211] Ebers (n 192) 75 para 111.
[212] Engert (n 178), (2018) 218(2-4) AcP (Archiv für die civilistische Praxis) 218, 304, 307 f.
[213] On the individualization of matching on platforms: M Berberich and A Conrad, ‘§ 30 Plattformen und KI’ in M Ebers, C Heinze and B Steinrötter (ed), Künstliche Intelligenz und Robotik (Beck 2020) para 28.
[214] Specifically on Airbnb, see Katsh and Rabinovich-Einy (n 96) 68 ff.
[215] Hörnle (n 167) 450.
[216] Detailed on the phenomenon of privatization of the judiciary, including in the context of the right to be forgotten: E Haber, ‘Privatization of the Judiciary’, (2016) 40(1) Seattle University Law Review 115, 120 ff; Hörnle (n 167) 450. Specifically, in the context of expression and copyright law, see Elkin-Koren and Perel (n 95), (2020) 24(3) Lewis & CLARK L. REV. 857, 871; Askani (n 184) 177 ff, 251 f, 269.
[217] E Haber (n 217), (2016) 40(1) Seattle University Law Review 115, 118 ff.
[218] Hörnle (n 167) 450. On the causes of the increase in private enforcement in copyright law, see Askani (n 184) 174-176.
[219] Elkin-Koren and Perel (n 95), (2020) 24(3) Lewis & CLARK L. REV. 857, 871.
[220] F Hofmann, ‘Prozeduralisierung der Haftungsvoraussetzungen im Medienrecht – Vorbild für die Intermediärshaftung’ (2017) 61(2) ZUM (Zeitschrift für Urheber- und Medienrecht) 102, 104 f.
[221] According to Askani, the adaptation of intermediaries can be understood as a reaction to the existing legal system that the development took place, so to speak, ‘in the shadow of the law’, Askani (n 184) 170; see also Frosio (n 202) 6; Grosse Ruse-Kahn (n 47), (2018) 49(9) IIC (International Review of Intellectual Property and Competition Law) 1017, 1018.
[222] Lennartz and Kraetzig (n 202), Filtering fundamental Rights.
[223] Wielsch (n 188), 61, 65; Klonick (n 94), (2018) 131(6) Harvard Law Review 1598, 1603 f, 1669 f.
[224] See only BGH (Germany), 29 July 2021, III ZR 179/20, BGHZ 230, 347 = (2021) 24(11) MMR (Multimedia und Recht) 903 para 78 (deletion of posts and account blocking by Facebook in the case of hate speech).
[225] As a result of the incident, major advertisers, including the British government and L'Oréal Group, withdrew their ads, also on YouTube, see: M Murgia, H Warell and D Bond, ‘YouTube revenues under threat over ads alongside extremist videos’ https://www.ft.com/content/04f8bf56-0b12-11e7-97d1-5e720a26771b accessed 31 December 2023; K Walker, ‘Four ways Google will help to tackle extremism’ https://www.ft.com/content/ac7ef18c-52bb-11e7-a1f2-db19572361bb accessed 31 December 2023; Kent Walker is Google’s senior vice-president and general counsel.
[226] Murgia, Warell and Bond (n 226), ‘YouTube revenues under threat over ads alongside extremist videos’.
[227] Critically: J Lüdemann, ‘Privatisierung der Rechtsdurchsetzung in sozialen Netzwerken?’ in M Eifert and T Gostomzyk (ed), Netzwerkrecht (2018) 165.
[228] See Report of the German Federal Government on the Evaluation of the Network Enforcement Act, Bundestag-Drucksache 19/22610, 8: ‘In this context, the specifications contain implementation leeway for the providers of the social networks with regard to the implementation of the specifications’.
[229] Report of the German Federal Government on the Evaluation of the Network Enforcement Act, Bundestag-Drucksache 19/22610, 10, 29, 86. However, the provision of § 2(2) no 2 NetzDG imposes transparency obligations on network operators to provide information on procedures for automated content recognition, see J-C Kalbhenn, ‘Design Specifications for Chatbots, Deepfakes, and Emotion Recognition Systems’ (2021) 65(8/9) ZUM (Zeitschrift für Urheber- und Medienrecht) 663, 672.
[230] See also Elkin-Koren and Perel (n 95), (2020) 24(3) Lewis & CLARK L. REV. 857, 871. – On the copyright liability model: Askani (n 184) 192-194.
[231] § 4 NetzDG (Network Enforcement Act).
[232] Also: Bloch-Wehba (n 3), (2020) 53(1) Cornell International Law Journal 41, 86.
[233] Bloch-Wehba (n 3), (2020) 53(1) Cornell International Law Journal 41, 86.
[234] Hörnle (n 167) 38.
[235] CJEU, 3 October 2019, C-18/18 – Glawischnig-Piesczek, ECLI:EU:C:2019:821, para 46.
[236] D Kaye, Speech Police (Columbia Global Reports 2019) 79.
[237] H Bloch-Wehba, ‘Global Platform Governance: Private Power in the Shadow of the State’ (2019) 72(1) SMU Law Review 27, 63; Bloch-Wehba (n 3), (2020) 53(1) Cornell International Law Journal 41, 78; D E Bambauer, ‘Against Jawboning’ (2015) 100(1) MINN. L. REV. 51, 57-58.
[238] Subdivisions of (also national) police forces are often affected: for example, the British Counter-Terrorism Internet Referral Unit (CTIRU) was established by Scotland Yard, see Kaye (n 237) 79. The European equivalent is located at Europol, see Europol, EU Internet Referral Unit – EU IRU https://www.europol.europa.eu/about-europol/european-counter-terrorism-centre-ectc/eu-internet-referal-unit-eu-iru accessed 31 December 2023. On the genesis of the EU Internet Referral Unit, see R Eghbariah and A Metwally, ‘Informal Governance: Internet Referral Units and the Rise of State Interpretation of Terms of Service’ (2021) 23 Yale J.L. & Tech. 545, 574 f.
[239] According to the Meta Group, sanctioning of content reported for potential violation of state law regularly occurs only locally, for example through geo-blocking. In contrast, blocking of content violating community standards has a global effect: Oversight Board, 6 December 2022, 2021-002-FB-PAO – Meta’s Cross-Check Program, para 56; Meta Transparency Center, ‘How we assess reports of content violating local law’ https://transparency.meta.com/reports/content-restrictions/content-violating-local-law/ accessed 31 December 2023. – Instructive on how geo-blocking works: Hörnle (n 167) 448-450. Critical of this form of extraterritorial governance: Eghbariah and Metwally (n 239), (2021) 23 Yale J.L. & Tech. 545, 599.
[240] For example, Europol emphasizes: ‘The decision to remove the referred content is taken by the concerned service provider in accordance with their policies and terms of service’, in EU Internet Referral Unit Transparency Report (2021) 3 https://www.europol.europa.eu/cms/sites/default/files/documents/EU_IRU_Transparency_Report_2021.pdf; see also B Chang, ‘From Internet Referral Units to International Agreements: Censorship of the Internet by the UK and EU’ (2018) 49(2) Columbia Human Rights Law Review 114, 135; see also L Helfer and M K Land, ‘The Meta Oversight Board's Human Rights Future’ (2023) 44(6) Cardozo Law Review 2233, 2275 ff.
[241] Eghbariah and Metwally (n 239), (2021) 23 Yale J.L. & Tech. 545, 592, 601-606; Kaye (n 237) 81.
[242] Kaye (n 237) 82.
[243] Moderators in the ‘escalation process’, for instance, may apply special policies and exceptions not available to the public. Such moderators have special expertise, and undertake an in-depth, contextual review of the content in question, see: Oversight Board, 22 November 2022, 2022-007-IG-MR – UK Drill Music, Part 6, 15 f.
[244] Oversight Board, 22 November 2022, 2022-007-IG-MR – UK Drill Music, Key Findings, 3. For more details on the Government Request process, see above para 61-64.
[245] Chang (n 241), (2018) 49(2) Columbia Human Rights Law Review 114, 122, 135; Bloch-Wehba (n 3), (2020) 53(1) Cornell International Law Journal 41, 61; Bloch-Wehba (n 238), (2019) 72(1) SMU Law Review 27, 45 f, 62 f.
[246] The EU Internet Referral Unit, for instance, states: ‘The EU IRU participated in the EU Internet Forum Senior Officials meetings [...] and provided relevant contents to feed the database of hashes’, Europol, 2018 Consolidated Annual Activity 44 https://www.europol.europa.eu/cms/sites/default/files/documents/consolidated_annual_activity_report_2018.pdf accessed 31 December 2023; see also Eghbariah and Metwally (n 239), (2021) 23 Yale J.L. & Tech. 545, 604.
[247] Wagner (n 189) 6.
[248] See above para 33-38; Elkin-Koren and Perel (n 95), (2020) 24(3) Lewis & CLARK L. REV. 857, 887.
[249] A Bridy, ‘Intellectual Property’ in D Keller (ed), Law, Borders, and Speech: Proceedings and Materials (2017) 13; Bloch-Wehba (n 3), (2020) 53(1) Cornell International Law Journal 41, 86; D Holznagel, Notice and Take-Down-Verfahren als Teil der Providerhaftung (Mohr Siebeck 2013) 125 (stating that the notice-and-takedown process in Germany is purely self-regulatory); Askani (n 184) 169.
[250] A Conrad and G Nolte, ‘Schrankenbestimmungen im Anwendungsbereich des UrhDaG’ (2021) 65(2) ZUM (Zeitschrift für Urheber- und Medienrecht) 111, 118, referring to the statement of Google/YouTube from 8 November 2020 on the draft bill of the German Federal Ministry of Justice and Consumer Protection (BMJV) for a law to adapt copyright law to the requirements of the digital single market 15 https://www.bmj.de/SharedDocs/Downloads/DE/Gesetzgebung/Stellungnahmen/2020/110820_Stellungnahme_Google_RefE_Urheberrecht-ges.pdf?__blob=publicationFile&v=3 accessed 31 December 2023.
[251] Nahmias and Perel (n 78), (2021) 58(1) Harvard Journal on Legislation 145, 176 ff, with respect to ‘offensive language’; further Engstrom and Feamster (n 39) 18: ‘[...] such technologies are not sufficient to consistently identify infringements with accuracy, as they can only indicate whether a file's contents match protected content, not whether a particular use of an identified file is an infringement in light of the context within which the media was being used’. Suggesting design improvement for the Content ID process: L D Shinn, ‘YouTube’s Content ID as a Case Study of Private Copyright Enforcement Systems’ (2015) 43(2/3) AIPLA Quarterly Journal 359, 386 ff.
[252] Conrad and Nolte (n 251), (2021) 65(2) ZUM (Zeitschrift für Urheber- und Medienrecht) 111, 118 (In the foreseeable future, it is not to be expected that, for example, quotations or parodies will be correctly included in a legal assessment of AI).
[253] Engstrom and Feamster (n 39), (2017) The Limits of Filtering: A Look at the Functionality & Shortcomings of Content Detection Tools 18: ‘It is often permissible to excerpt or otherwise refer to copyrighted content in contexts that are permitted by fair use [...]. Although an automated algorithm could determine whether the content (or excerpt) matched known copyrighted content, such an algorithm would not be able to determine whether the particular use of a given file is infringing or not’. See also Shinn (n 252), (2015) 43(2/3) AIPLA Quarterly Journal 359, 364; Bloch-Wehba (n 3), (2020) 53(1) Cornell International Law Journal 41, 65. – YouTube itself recognizes that Content ID cannot decide ‘fair use’: YouTube, ‘Frequently asked questions about fair use’ https://support.google.com/youtube/answer/6396261#zippy=%2Ci-posted-a-disclaimer-on-my-video%2Ci-gave-credit-to-the-copyright-owner%2Cim-using-the-content-for-entertainment-or-non-profit-uses%2Cwhen-does-fair-use-apply%2Cwhat-constitutes-fair-use%2Chow-does-fair-use-work%2Chow-does-content-id-work-with-fair-use accessed 31 December 2023.
[254] Bloch-Wehba (n 3), (2020) 53(1) Cornell International Law Journal 41, 65. See further S Jacques, K Garstka, M Hviid and J Street, ‘An empirical study of the use of automated anti-piracy systems and their consequences for cultural diversity’, (2018) 15(2) Script-Ed 277, 298, which found that in a sample of 1.839 parodies, videos were five times more likely to be blocked by Content ID than by a DMCA proceeding; Gray and Suzor (n 5), (2020) 7(1) Big Data & Society 1, 6; A Metzger and M Senftleben, ‘Selected Aspects of Implementing Article 17 of the Directive on Copyright in the Digital Single Market into National Law – Comment of the European Copyright Society’ (20 April 2020) 1, 16 https://ssrn.com/abstract=3589323 or http://dx.doi.org/10.2139/ssrn.3589323 accessed 31 December 2023.
[255] Nolte (n 47), (2017) 61(4) ZUM (Zeitschrift für Urheber- und Medienrecht) 304, 310.
[256] Grosse Ruse-Kahn (n 5), (2020) PIJIP Research Paper Series 51, 1, 2; M Becker, ‘Von der Freiheit, rechtswidrig handeln zu können‘ (2019) 63(8/9) ZUM (Zeitschrift für Urheber- und Medienrecht) 636, 644.
[257] Cf Burk (n 69), (2019) 86(2) University of Chicago Law Review 283, 297 f: ‘The common law evolves, whether from purely judicial reasoning or from judicial riffing off of legislative enactments’ (at 298).
[258] Bloch-Wehba (n 3), (2020) 53(1) Cornell International Law Journal 41, 65; D Keller, ‘Internet Platforms: Observations on Speech, Danger, and Money’ (2018) Hoover Inst. Aegis Paper Series no 1807 6, 7: ‘an ISIS video looks the same, whether used in recruiting or in news reporting’.
[259] Grosse Ruse-Kahn (n 5), (2020) PIJIP Research Paper Series 51, 1, 9.
[260] YouTube, ‘What Does Fair Use Mean’ https://support.google.com/youtube/answer/9783148?hl=de accessed 31 December 2023. With respect to deactivations for copyright infringement, YouTube makes reference to the US DMCA (cf the reference to the formal counter notification requirements of 17 USC § 512(g)(3), which requires submission to US jurisdiction in 17 USC § 512(g)(3)(D) as a condition of a counter notification, see https://support.google.com/youtube/answer/6005919?hl=de&ref_topic=9282678 accessed 31 December 2023. With respect to other jurisdictions, the YouTube website contains only a reference to where ‘useful information on copyright outside the U.S.’ can be found and refers to the websites of the European Commission and WIPO in this regard. Here, it is expressly clarified that these references serve only ‘informational purposes’ and do not constitute a ‘binding recommendation’ by YouTube. No reference contains specific information. The link to the WIPO’s website refers to a list of National IP Offices, see YouTube, ‘Where can I get more information about copyright outside the U.S.?’, https://support.google.com/youtube/answer/2797449?hl=de&ref_topic=2778546#zippy=%2Cwo-erhalte-ich-weitere-informationen-zum-urheberrecht-außerhalb-der-usa accessed 31 December 2023.
[261] To this aspect see Nahmias and Perel (n 78), (2021) 58(1) Harvard Journal on Legislation 145, 178, referring to R Radu, Negotiating Internet Governance (1st edn, Oxford University Press 2019) 179.
[262] I S Nathenson, ‘The Procedural Foundations of Information Regulation’ (2020) 24(1) Lewis & Clark Law Review 109, 129; Gray and Suzor (n 5), (2020) 7(1) Big Data & Society 1, 2; D K Citron, ‘Technological due process’ (2008) 85(6) Washington University Law Review 1249, 1250, 1254; N Elkin-Koren, ‘After twenty years: revisiting copyright liability of online intermediaries’ in S Frankel and D Gervais (ed), The Evolution and Equilibrium of Copyright in the Digital Age (1st edn, Cambridge University Press 2014) 29, 47; Burk (n 69), (2019) 86(2) University of Chicago Law Review 283, 301; C Castets-Renard, ‘Algorithmic Content Moderation on Social Media in EU Law: Illusion of Perfect Enforcement’ (2020) (2)2 University of Illinois Journal of Law, Technology & Policy 283, 308.
[263] Bloch-Wehba (n 3), (2020) 53(1) Cornell International Law Journal 41, 81 ff: ‘But overconfidence in technical solutions can have damaging effects. Far from serving as a neutral arbiter, the algorithms that Internet intermediaries use to rank and prioritize content often reflect and encode social bias’.
[264] Cf the example of T Zhou, ‘Postmortem: Every Frame a Painting’ (2 December 2017) https://perma.cc/U5WU-M6ZZ accessed 31 December 2023. – On so-called reverse engineering, see Burk (n 69), (2019) 86(2) University of Chicago Law Review 283, 303; Bloch-Wehba (n 3), (2020) 53(1) Cornell International Law Journal 41, 83.
[265] See only Burk (n 69), (2019) 86(2) University of Chicago Law Review 283, 296: ‘Algorithms do not make judgments; they are rather the products of human judgment’.
[266] Art 13 of the EU AI Act provides for transparency obligations for so-called high-risk systems in the form of users being able to ‘appropriately interpret and use the results of the system’. The problem of lacking comprehensibility of AI decisions is the starting point of the research field of so-called ‘Explainable AI’, see D Bomhard and M Merkle, ‘Regulation of Artificial Intelligence’ (2021) 10(6) EuCML (Journal of European Consumer and Market Law) 257, 260; S Heiss, ‘Artificial Intelligence Meets European Union Law’ (2021) 10(6) EuCML (Journal of European Consumer and Market Law) 252, 258; D Gunning et al, ‘XAI-Explainable artificial intelligence’ (2019) 4(37) Science Robotics DOI: 10.1126/scirobotics.aay7120 accessed 31 December 2023.
[267] Nathenson (n 263), (2020) 24(1) Lewis & Clark Law Review 109, 129 f. Consequently, even experts in the field might not understand the formulas if the algorithm were fully disclosed. For a detailed discussion of the black box problem, see F Pasquale, The Black Box Society (1st edn, Harvard University Press 2015) 3 ff.
[268] Even full disclosure of an algorithm would be insufficient to the extent that AI-based results may also depend on the algorithm’s technical infrastructure (hardware and other software): Burk (n 69), (2019) 86(2) University of Chicago Law Review 283, 302.
[269] J Burrell, ‘How the machine thinks: Understanding opacity in machine learning algorithms’ (1/2016) 3(1) Big Data & Society 3 ff DOI: 10.1177/2053951715622512 accessed 31 December 2023.
[270] Schillmöller and Doseva (n 5), (2022) 25(3) MMR (Multimedia und Recht) 181, 185; Perel and Elkin-Koren (n 4), (2016) 19(3) Stanford Technology Law Review 473, 483; S Bar-Ziv and N Elkin-Koren, ‘Behind the Scenes of Online Copyright Enforcement: Empirical Evidence on Notice & Takedown’ (2018) 50(2) Connecticut Law Review 339, 382; D Leenheer Zimmerman, ‘A Tale of Legislative Abdication’ (2014) 35(1) Pace Law Review 260, 273 f.
[271] See already B Laukemann, ‘Private law enforcement and intellectual property: Regulatory challenges in a digital era’ in B Hess, E Jayme and H-P Mansel (ed), Europa als Rechts- und Lebensraum: Liber Amicorum für Christian Kohler zum 75. Geburtstag (Gieseking 2018) 269, 276 f; cf also: Bloch-Wehba (n 3), (2020) 53(1) Cornell International Law Journal 41, 46, with respect to content moderation (‘Rather, content moderation rules – and the technologies that apply them – reflect corporate, social, and legal values’); Pasquale (n 268) 61.
[272] Cf Bloch-Wehba (n 3), (2020) 53(1) Cornell International Law Journal 41, 65.
[273] Bloch-Wehba (n 3), (2020) 53(1) Cornell International Law Journal 41, 83.
[274] A study of US law by J Urban, J Karagani and B Schofield found that about 30% of notifications were probably unfounded. According to another study, which looked at Google Images, this affected 70% of notices: ‘Notice and Takedown in Everyday Practice’ (2016) UC Berkeley Public Law Research Paper No 2755628 https://ssrn.com/abstract=2755628 or http://dx.doi.org/10.2139/ssrn.2755628, 11 f; D C Nunziato, ‘The Beginning of the End of Internet Freedom’ (2014) 45 Georgetown Journal of International Law 383, 383: ‘[...] such Internet filtering regimes [...] inevitably lead to overblocking of harmless Internet content’; M Senftleben, ‘Institutionalized Algorithmic Enforcement – The Pros and Cons of the EU Approach to UGC Platform Liability’ (2020) 14(2) FIU Law Review 299, 312: ‘Filtering more than necessary is less risky than filtering only clear-cut cases of infringement’; Gray and Suzor (n 5), (2020) 7(1) Big Data & Society 1, 2; Bar-Ziv and Elkin-Koren (n 271), (2018) 50(2) Connecticut Law Review 339: ‘Analysis of the data reveals that the N&TD procedure has been extensively used to remove non-infringing materials’; H Maier, Remixes on Hosting Platforms (Mohr Siebeck 2018) 152; C Katzenbach, ‘The “Alghorithmic turn” in platform governance’ (2020) 74(1 supp) Cologne Journal of Sociology and Social Psychology 283, 297: ‘For the copyright field, the few existing studies point to clear overblocking’; Gorwa, Binns and Katzenbach (n 94), (2020) 7(1) Big Data & Society 1, 5.
[275] For example, only one minute of a total 15-minute video, as exemplified by Grosse Ruse-Kahn (n 5), (2020) PIJIP Research Paper Series 51, 1, 5.
[276] This appears to be YouTube’s current approach, Grosse Ruse-Kahn (n 5), (2020) PIJIP Research Paper Series 51, 1, 5 at fn 19. YouTube, for its part, cites case law on so-called fair use, according to which there is no minimum time (of exploitation of others’ works) that would be allowed under copyright law (concerning a few seconds of a sample): YouTube, ‘Answers to common questions about Copyright claims on YouTube’ https://support.google.com/youtube/thread/1281991 accessed 31 December 2023.
[277] Schillmöller and Doseva (n 5), (2022) 25(3) MMR (Multimedia und Recht) 181, 186 f.
[278] Nahmias and Perel (n 78), (2021) 58(1) Harvard Journal on Legislation 145, 173; R Tushnet, ‘All of this has happened before and all of this will happen again: Innovation in copyright licensing’ (2014) 29(3) Berkeley Technology Law Journal 1447, 1460; T Spoerri, ‘On Upload Filters and other Competitive Advantages for Big Tech Companies under Article 17 of the Directive on Copyright in the Digital Single Market’ (2019) 10(2) Journal of Intellectual Property, Information Technology and Electronic Commerce Law 173, 176; F Mostert, ‘Free Speech and Internet Regulation’ (2019) 14(8) Journal of Intellectual Property Law & Practice 607, 612: ‘Over-blocking and excessive filtering could too easily lead to censorship’.
[279] Gray and Suzor (n 5), (2020) 7(1) Big Data & Society 1, 6 f: When content related to video games is removed, it is usually because of a music rightsholder’s demand.
[280] YouTube states: ‘For example, we may disable certain reference files or segments and remove associated claims entirely. Manual review is also required for certain reference categories. In cases of serious infringement, we may revoke access to Content ID or terminate the partnership between YouTube and the copyright owner’: YouTube, ‘Content eligible for Content ID’ https://support.google.com/youtube/answer/2605065#zippy accessed 31 December 2023.
[281] YouTube, ‘Best practices for claims’ https://support.google.com/youtube/answer/4352063 accessed 31 December 2023.
[282] In this respect the assumption of Grosse Ruse-Kahn (n 5), (2020) PIJIP Research Paper Series 51, 1, 11, referring to YouTube, ‘Review potentially invalid references’ https://support.google.com/youtube/answer/6013183 accessed 31 December 2023.
[283] See in more detail above para 68-77 and below para 167-170. Further: Askani (n 184) 178 f.
[284] Bloch-Wehba (n 3), (2020) 53(1) Cornell International Law Journal 41, 82. For example, evaluating a user's online activities enables accurate predictions about ethnicity, partisan political views, religion, substance use, sexual orientation, extraversion, intelligence, or emotional stability.
[285] Ibid 41, 76.
[286] Ibid 41, 77.
[287] In the Google Ads algorithm, this learning process occurs by assigning weights or statistical probabilities based on the call history of ads, see L Sweeney, ‘Discrimination in Online Ad Delivery’ (2013) 56(5) Comm. ACM 44, http://cacm.acm.org/magazines/2013/5/163753-discrimination-in-online-ad-delivery/ accessed 31 December 2023; A Chander, ‘The Racist Algorithm?’ (2017) 115(6) Michigan Law Review 1023, 1037.
[288] Ebers (n 192) 75 para 162-168. In-depth on the technical causes of algorithmic discrimination also S Barocas and A D Self, ‘Big Data’s Disparate Impact’ (2016) 104(3) California Law Review 671, 680 f.
[289] More generally on the technical workings of feedback loops, see Yeung (n 189), (2017) 20(1) Communication & Society 118, 121 f.
[290] See already above para 68-70.
[291] See already above para 95-98. Bloch-Wehba (n 3), (2020) 53(1) Cornell International Law Journal 41, 83; Chander (n 288), (2017) 15(6) Michigan Law Review 1023, 1037.
[292] Nahmias and Perel (n 78), (2021) 58(1) Harvard Journal on Legislation 145, 181 ff.
[293] See M E Kaminski, ‘Binary Governance: Lessons from the GDPR’s Approach to Algorithmic Accountability’ (2019) 92(6) S. CAL. L. REV. 1529, 1580-1582; Nahmias and Perel (n 78), (2021) 58(1) Harvard Journal on Legislation 145, 182 with further references at fn 231.
[294] See Nahmias and Perel (n 78), (2021) 58(1) Harvard Journal on Legislation 145, 178 with further references at para 235; Van Loo (n 100), (2021) 88(4) University of Chicago Law Review 829, 830, 863.
[295] Facebook (Meta), ‘Community Standards Enforcement Report: Child Endangerment: Nudity and Physical Abuse and Child Sexual Exploitation’ https://transparency.fb.com/data/community-standards-enforcement/child-nudity-and-sexual-exploitation/facebook/ accessed 31 December 2023.
[296] See, for example, in German law the complaint and counter-proposal procedure in §§ 3 ff NetzDG (Network Enforcement Act).
[297] BGH (Germany), 29 July 2021, III ZR 179/20, BGHZ 230, 347 = (2021) 65(11) ZUM (Zeitschrift für Urheber- und Medienrecht) 953; BGH (Germany), 29 July 2021, III ZR 192/20, (2021) 25 (11) ZUM-RD (Zeitschrift für Urheber- und Medienrecht – Rechtsprechungsdienst) 612.
[298] The BGH (Germany) argued for a competence of network operators to prohibit forms of ‘hate speech’ on the basis of their terms and conditions also on this side of punishable or right-infringing expressions of opinion, see only BGH (Germany), 29 July 2021, III ZR 192/20, (2021) 25(11) ZUM-RD (Zeitschrift für Urheber- und Medienrecht – Rechtsprechungsdienst) 612 para 91.
[299] BGH (Germany), 29 July 2021, III ZR 192/20, (2021) 25(11) ZUM-RD (Zeitschrift für Urheber- und Medienrecht – Rechtsprechungsdienst) 612 para 96.
[300] BGH (Germany), 29 July 2021, III ZR 192/20, (2021) 25(11) ZUM-RD (Zeitschrift für Urheber- und Medienrecht – Rechtsprechungsdienst) 612 para 66, 77, 80. In this light, the BGH does not develop the procedural rights on the basis of ‘contractual types’, ie in relation to the – in any case difficult to identify – legal model of usage agreements between platform and user, cf also D Holznagel, ‘Nutzerrechte bei Facebook: Klärung durch den BGH und bevorstehende Irrwege des EU-Gesetzgebers’ (2021) 37(11) CR (Computer und Recht) 733, 735.
[301] Directive (EU) 2019/790 of 17 April 2019 on copyright and related rights in the digital single market and amending Directives 96/9/EC and 2001/29/EC.
[302] On the internal complaints procedure pursuant to Art 11 Regulation (EU) 2019/1150 of 20 June 2019 on promoting fairness and transparency for business users of online intermediary services (hereinafter: P2B Regulation).
[303] Cf Art 16(6), 14(4), 23(3) and recitals 24 s 3, 26 s 2 of the Regulation (EU) 2022/2065 of 19 October 2022 on a Single Market For Digital Services and amending Directive 2000/31/EC (Digital Services Act; hereinafter: DSA Regulation), OJ L 277, 27 October 2022, 1-102.
[304] Directive 2000/31/EC of 8 June 2000 on certain legal aspects of information society services, in particular electronic commerce, in the Internal Market (‘Directive on electronic commerce’).
[305] CJEU, 3 October 2019, C-18/18 – Glawischnig-Piesczek, ECLI:EU:C:2019:821, para 48 f. The case was submitted by the Austrian Supreme Court. In casu, the issue was the assessment of defamatory statements in accordance with § 78 Austrian Copyright Act (öUrhG), § 1330 Austrian General Civil Code (öABGB). – Critically: T Hoeren, ‘Sperrpflichten eines Hosting-Anbieters bei rechtswidrigen Informationen sowie wort- und sinngleichen Inhalten’ (2020) LMK (Leitsätze mit Kommentierung) 425949. – In contrast, the CJEU had ruled for search engines that they are obliged to remove links from result lists only on a Europe-wide basis when assessed from a data protection perspective, see CJEU (Grand Chamber), 24 September 2019, C-507/17 – Google (Spatial scope of delisting), ECLI:EU:C:2019:772, para 44 f.
[306] Platform Accountability and Consumer Transparency Act, 116th Cong. § 5(2) (2020) (‘PACT Act’); On 3 June 2023, the revised Constitution was introduced in the U.S. Senate, see Platform Accountability and Consumer Transparency Act, 118th Congress.
[307] H.B. 20 (Tx. 2021); S.B. 7072 (Fl. 2021); see also E Douek, ‘Content Moderation as Systems Thinking’ (2022) 136(2) Harvard Law Review 528, 566 f.
[308] Examples include: (i) EU law, Art 11 s 2 of the Enforcement Directive (2004/48/EC of April 29, 2004 on the enforcement of intellectual property rights) or Art 8(3) of the InfoSoc Directive (2001/29/EC of 22 May 2001 on the harmonization of certain aspects of copyright and related rights in the information society). – (ii) In German law, so-called ‘Störerhaftung’ applies in part; in copyright law, platforms have recently even been held liable as perpetrators: Art 17 of the DSM Directive, for example, imposes perpetrator liability (‘täterschaftliche Haftung’) on the platform operators concerned if certain traffic duties are violated: F Hofmann, ‘Fünfzehn Thesen zur Plattformhaftung nach Art 17 DSM-RL’ (2019) 121(12) GRUR (Gewerblicher Rechtsschutz und Urheberrecht) 1219. An overview of German law is provided, for example, by F Hofmann, ‘Mittelbare Verantwortlichkeit im Internet’ (2017) 57(8) JuS (Juristische Schulung) 713 ff.
[309] § 7(2) of the German Telemedia Act (Telemediengesetz: TMG) states: ‘Service providers within the meaning of Sections 8 to 10 are not obliged to monitor the information they transmit or store or to investigate circumstances that indicate illegal activity’.
[310] On permitted ‘specific monitoring obligations’ see also recital 47 Directive 2000/31/EC as well as CJEU, 3 October 2019, C-18/18 – Glawischnig-Piesczek, ECLI:EU:C:2019:821, para 31 ff., 34. In Austria, the provision was implemented in § 18 E-Commerce Act.
[311] The situation is different with regard to liability for setting hyperlinks: In this regard, the Federal Court of Justice (Bundesgerichtshof) allows a simple reference to the infringement to suffice: BGH (Germany), 18 June 2015, I ZR 74/1 – Liability for Hyperlink, BGHZ 206, 103 = (2016) 69(11) NJW (Neue Juristische Wochenschrift) 804 para 27.
[312] See, for example, Art 16(1), (3), Art 6(1) lit b) DSA Regulation. Furthermore, BGH (Germany), 17 August 2011, I ZR 57/09 – Stiftparfüm, BGHZ 191, 19 = (2011) 113(11) GRUR (Gewerblicher Rechtsschutz und Urheberrecht) 1038 para 21 ff, 26 (‘Störerhaftung’ for Internet auction houses); BGH (Germany), 12 July 2012, I ZR 18/11 – Alone in the dark, BGHZ 194, 339 = (2013) 66(11) NJW (Neue Juristische Wochenschrift) 784 para 28 (‘Störerhaftung’ of file hosting services). – An infringement of rights that gives rise to liability (and thus indicates a risk of repetition) only exists if the intermediary does not comply with a justified request for deletion. Only with this breach of duty do the costs of a notice of infringement become recoverable, see § 97a(3) s 1 German Copyright Act (UrhG): BGH (Germany), 17 August 2011, I ZR 57/09 – Stiftparfüm, BGHZ 191, 19 = (2011) 113(11) GRUR (Gewerblicher Rechtsschutz und Urheberrecht) 1038 para 39; F Hofmann (n 309), (2017) 57(8) JuS (Juristische Schulung) 713, 715.
[313] The EU legislator also exempts service providers from a general obligation to monitor or actively investigate under Art 8 of the DSA Regulation. However, Art 4(3), 5(2) and 6(4) DSA Regulation allow service providers to be required by Member State law to ‘cease or prevent infringements’ in response to judicial or administrative orders (stay down). In this respect, questions of demarcation also arise in the context of the DSA Regulation with regard to the prohibition of general monitoring and the duty to prevent future infringements, cf on this (with regard to the DSA draft): R Janal, ‘Haftung und Verantwortung im Entwurf des Digital Services Acts’ (2021) 29(2) ZEuP (Zeitschrift für Europäisches Privatrecht) 227, 248-252.
[314] See recently under the Digital Services Act: Art 16(2) s 1 (‘sufficiently precise and adequately substantiated notification’), s 2 lit a) (explanation of the basis for classifying information as unlawful) and s 2 lit d) (confirmation of the accuracy and completeness of the notification). – The obligation to substantiate also varies depending on the law concerned: For example, the registry for ‘.de-domains’ (DENIC) is only liable for the breach of obligations of conduct in the case of domain registrations that infringe the law if the infringement is readily apparent. For this to be the case, the German Federal Court of Justice requires that either DENIC has a legally enforceable title or that the infringement is so clear that it must impose itself on [DENIC]: BGH (Germany), 27 October 2011, I ZR 131/10 – regierung-oberfranken.de (2012) 65(31) NJW (Neue Juristische Wochenschrift) 2279 para 26, on § 12 BGB (German Civil Code); clear abuse.
[315] On the identification function of the notice, see F Hofmann (n 221), (2017) 61(2) ZUM (Zeitschrift für Urheber- und Medienrecht) 102, 104 f.
[316] This is also accompanied by a (considerable) reduction of due diligence costs, see G Wagner, ‘Haftung von Plattformen für Rechtsverletzungen (Teil 2)’ (2020) 122(5) GRUR (Gewerblicher Rechtsschutz und Urheberrecht) 447, 448.
[317] For more details, see below para 155-160.
[318] This idea is also echoed by Wagner (n 317), (2020) 122(5) GRUR (Gewerblicher Rechtsschutz und Urheberrecht) 474, 455, when he formulates that effective law enforcement by state courts is ‘illusory’ under the conditions of the Internet. Cf from a U.S. perspective: Van Loo (n 100), (2021) 88(4) University of Chicago Law Review 829, 830, 889.
[319] In addition, if – reinforced by increasingly strict ex ante due diligence and auditing obligations of platforms – there is at least a de facto liability differential, because the preventive control function of liability law is less pronounced in the relationship of the intermediary to the enforcement addressee than vis-à-vis the affected holder of the right to be enforced, this in turn creates additional incentives for excessive enforcement of private rights.
[320] On corresponding practices of the sharing platform Airbnb: Van Loo (n 100), (2021) 88(4) University of Chicago Law Review 829, 844 f, 860, 879.
[321] Cf BGH (Germany), 29 July 2021, III ZR 179/20, BGHZ 230, 347 = (2021) 65(11) ZUM (Zeitschrift für Urheber- und Medienrecht) 953 para 66; also Van Loo (n 100), (2021) 88(4) University of Chicago Law Review 829, 863 (2021).
[322] See also CJEU, 3 October 2019, C-18/18 – Glawischnig-Piesczek, ECLI:EU:C:2019:821, para 28, 36; also EU Commission, ‘Staff Working Document Impact Assessment, Accompanying the document Proposal for a Regulation of the European Parliament and of the Council on preventing the dissemination of terrorist content online’ SWD (2018), 408 final, at 2.4.3 (‘Generally speaking, the longer the content is able to survive online, the more views it may receive, and the more harm it may cause’).
[323] As a result, especially in the case of personality rights, there is a risk of deepening or even irreversible damage; Wagner (n 317), (2020) 122(5) GRUR (Gewerblicher Rechtsschutz und Urheberrecht) 447, 455, speaks of ‘Schadensvertiefung durch Zeitablauf’.
[324] Cf generally on the contribution of procedural structures to the increased legitimacy of the results produced by be: T R Tyler, Why people obey the law (Princeton University Press 2006) 5, 9; further N Luhmann, Legtimation durch Verfahren (11th edn, Suhrkamp 2019) 55 ff.
[325] Van Loo (n 100), (2021) 88(4) University of Chicago Law Review 829, 865; see also C Rule, ‘Quantifying the Economic Benefits of Effective Redress: Large ECommerce Data Sets and the Cost-Benefit Case for Investing in Dispute Resolution’ (2012) 34(4) U ARK LITTLE ROCK L REV 767, 776, finding that buyers who reached amicable dispute resolutions were more likely to return than buyers who simply achieved a full refund in their dispute.
[326] Thus, for the notification and redress procedure vis-à-vis ‘all data subjects’: recital 52 s 1, 2 DSA Regulation; for general terms and conditions of providers of intermediary services vis-à-vis users: Art 14(4) DSA Regulation.
[327] For more details, see below para 150-151.
[328] On this point, H-J Blanke in C Callies and M Ruffert (ed), EUV/AEUV (6th edn, Beck 2022) Art 47 EU-GRCh para 15 f.
[329] Cf on this in extrajudicial dispute resolution: B Hess, ‘Prozessuale Mindestgarantien in der Verbraucherschlichtung’ (2015) 70(11) JZ (Juristenzeitung) 548 ff.
[330] For more details, see below para 129-132.
[331] Cf, for example, the instruments of provisional legal protection under the German Code of Civil Procedure (§§ 916 ff ZPO), such as measures to protect against unjustified claims and an abusive use of state legal protection, for example by means of prima facie evidence, §§ 920(2), 294 ZPO. Furthermore, the provision of security by the claimant (§ 921 s 2 ZPO); effective downstream legal protection (order to bring an action, § 926 ZPO; legal remedies, §§ 924, 927 ZPO) as well as protection in the event of unjustified recourse to interim legal protection (for example, through compensatory damages under § 945 ZPO).
[332] For more details, see below para 155-160.
[333] Art 41(3) lit e) DSA Regulation with regard to providers of very large online platforms – On the classification of Amazon Services Europe Sàrl as a ‘very large online platform’ within the meaning of the Digital Services Act, see most recently the Order of the European General Court, 27 September 2023, T-367/23 – Amazon Services Europe v Commission, ECLI:EU:T:2023:589.
[334] Art 37(1) lit a) DSA Regulation in relation to providers of very large online platforms
[335] See on the sanctions regime of the Digital Services Act: Art 52(1), Art 74(1) lit a).
[336] On the whole, T Mast, ‘AGB-Recht als Regulierungsrecht’ (2023) 78(7) JZ (Juristenzeitung) 287, 291.
[337] These are the programmatic functional descriptions of Art 16(1) and 20(4) of the DSA Regulation. – For the principle of transparency in out-of-court dispute resolution, see Art 7 Directive 2013/11/EU of the European Parliament and of the Council of 21 May 2013 on alternative dispute resolution for consumer disputes and amending Regulation (EC) No 2006/2004 and Directive 2009/22/EC (Directive on consumer ADR), OJ L 165, 18 June 2013, 63.
[338] For example, the deadline for a decision under the German UrhDaG is one week after the complaint has been filed, § 14(3) no 3 UrhDaG. – The German NetzDG establishes a graduated time limit model that is essentially based on the complexity of the decision: In the case of complaints involving ‘obviously illegal content’, blocking must take place within 24 hours; in the case of merely ‘illegal content’, it must generally occur within seven days, § 3(2) NetzDG. The subsequently introduced counternotification procedure within the meaning of § 3b NetzDG does not stipulate a time limit requirement, nor does the Digital Services Act, see Art 20(4): ‘zeitnah’/’timely’; cf also G Spindler, ‘Der Vorschlag für ein neues Haftungsregime für Internetprovider – der EU-Digital Services Act (Teil 1)’ (2021) 123(4) GRUR (Gewerblicher Rechtsschutz und Urheberrecht) 545, 553. – Pursuant to § 89b(5) no 5 Austrian UrhG (Copyright Act), complaints must generally be concluded within two weeks.
[339] Cf the German § 14(1) UrhDaG (internal complaints procedure): ‘The service provider must provide users and rightsholders with an effective, free of charge and expeditious complaints procedure about blocking and about the communication to the public of protected works’ Similarly, the internal complaints procedure for commercial users under Art 11 P2B Regulation; also §§ 3-3b NetzDG. – Pursuant to § 3(1) of the Austrian Communications Platforms Act (KoPlG), service providers must establish an ‘effective and transparent procedure for dealing with and settling reports of allegedly illegal content available on the communications platform’. Apart from this, § 3(4) of the KoPlG stipulates an ‘effective and transparent’ counternotification procedure.
[340] See also Art 14(1), Art 16(1) DSA Regulation.
[341] See Art 11 P2B Regulation.
[342] Texas: H.B. 20 (Tx. 2021); Florida: S.B. 7072, Subchapter C (Fl. 2021).
[343] §§ 14, 15 UrhDaG; §§ 3-3b NetzDG.
[344] German: Aufklärungsverwortung.
[345] In that regard, the so-called ‘shuttle procedure’ of the VI Civil Senate of the German Federal Court of Justice: BGH (Germany), 25 October 2011, VI ZR 93/10 – Blog Eintrag, BGHZ 191, 219 = (2012) 65(3) NJW (Neue Juristische Wochenschrift) 148 para 27; also: BGH (Germany), 1 March 2016, VI ZR 34/15 – Ärztebewertungsportal III (jameda.de), BGHZ 209, 139 = (2016) 69(29) NJW (Neue Juristische Wochenschrift) 2106 para 21 ff, 37 ff, 41 ff.
[346] As a consequence, failure to respond on the part of the parties to the dispute leads to disadvantages in enforcing the law: If, for example, a blogger remains silent in response to a request for comments, this upholds the prior deletion of his (presumably incriminated) blog entry; conversely, a corresponding failure to act on the part of the rightsholder precludes a (final) deletion of this content: BGH (Germany), 25 October 2011, VI ZR 93/10 – Blog-Eintrag, BGHZ 191, 219 = (2012) 65(3) NJW (Neue Juristische Wochenschrift) 148, para 27; also: BGH (Germany), 1 March 2016, VI ZR 34/15 – Ärztebewertungsportal III (jameda.de), BGHZ 209, 139 = (2016) 69(29) NJW (Neue Juristische Wochenschrift) 2106 para 21 ff, 37 ff, 41 ff. – Cf in German law furthermore the counternotification procedure under § 3b NetzDG.
[347] After a decision has been made pursuant to § 3 NetzDG, the complainant or, in a mirror image, the user must file an application for review of the decision pursuant to § 3b of the NetzDG. This application must be substantiated (without making ‘too high demands’ on this, see Bundestags-Drucksache 19/18972, 47). In the event of a planned remedy on the part of the platform, the other party is, in turn, granted an opportunity for a countermotion, on this also F Hofmann and L Specht-Riemenschneider, ‘Verantwortung von Online-Plattformen (Responsibility of Online Platforms)’ (2021) 13(1) ZGE (Zeitschrift für geistiges Eigentum) 48, 99, who refer to these mechanisms as ‘notice-and-negotiation procedures’ and assume that §§ 3, 3b NetzDG map the procedure of blog entry jurisdiction. – For the most part, the affected third party is not involved in the decision under § 3 NetzDG. The possibility of involving the user does exist in the context of the review of ‘illegal’ content pursuant to § 3(2) no 3 lit a) NetzDG. However, according to the wording, this review is merely optional and is not actually carried out by the platforms, see M Eifert et al, Netzwerkdurchsetzungsgesetz in der Bewährung (Nomos 2020) 80 f. – In contrast to the procedure under § 3(4) Austrian KoPlG, the counternotification procedure pursuant to § 3b KoPlG is much more comprehensive.
[348] In German law, for example, §§ 138(3), 142, 144 ZPO (Code of civil procedure).
[349] Similarly as with the state regime on provisional legal protection the objective of optimized clarification of facts recedes with special platform-typical endangerment situations behind the request to protect a realization of the endangered right effectively, by enforcing platforms this immediately, at least on basis of a smaller measure of conviction, provisionally.
[350] Keller (n 259), (2018) Hoover Inst. Aegis Paper Series no 1807 6, 7, speaks of ‘context-blindness’; cf further Bloch-Wehba (n 3), (2020) 53(1) Cornell International Law Journal 41, 65 with reference to the difficulty of detecting ‘hate speech’ offenses using fingerprinting technology.
[351] This is evident, for example, in the German NetzDG proceedings: the figures for ‘obviously unlawful’ and ‘unlawful’ content differ greatly depending on the platform. While the rate of ‘obviously illegal’ content was 95% for one provider, it was 8,69% (!) for another, see Eifert et al (n 348), (2020) 69 f. To make matters worse, the examination in the NetzDG counternotification proceedings is subsequently not limited to the reasons stated in the counternotification, but rather takes place ‘under all legal aspects that come into consideration’, see Bundestags-Drucksache 19/18972, 47.
[352] In German: Modell der prozeduralen Handlungsverantworung.
[353] Cf under German law: § 9(2) UrhDaG, which for clearly defined facts of user-generated content – eg, content that contains less than half of a work of a third party or several works of third parties (no 1) or that is marked as legally permitted pursuant to § 11 UrhDaG (no 3) – rebuttably presumes that its use is legally permitted according to § 5 UrhDaG (so-called presumed permitted uses). – In such a case, the service provider shall immediately inform the rightsholder of the public communication and of the right to file a complaint pursuant to § 14 UrhDaG in order to have the presumption reviewed pursuant to § 9(2) UrhDaG.
[354] Cf in German law §§ 920(2), 936, 294 ZPO (Code of civil procedure).
[355] The Digital Services Act has recently moved in this direction in Art 16(3), as a liability standard for content-related inspection obligations: ‘Notices referred to in this Article shall be considered to give rise to actual knowledge or awareness for the purposes of Article 6 in respect of the specific item of information concerned where they allow a diligent provider of hosting services to identify the illegality of the relevant activity or information without a detailed legal examination’.
[356] For more details, see below para 141-143.
[357] Recently, the German legislator has formulated a comparable obligation to respond to internal complaint procedures of the platforms (service providers) in § 14(4) UrhDaG, thereby implementing the DSM Directive on copyrights in the digital single market, although only for so-called trustworthy rightsholders – The trustworthiness of the rightsholder is to be assessed by the service provider and can result, for example, from the scope of the valuable repertoire deposited with the service provider, the associated deployment of particularly qualified personnel or also from the successful completion of quite a few complaint procedures in the past. In the event of disputes about the trustworthiness of a rightsholder (and thus at the same time about access to the ‘red button’), this question can also be clarified in court, see Bundestag-Drucksache 19/27426, 144. – A repeated, clearly incorrect blocking request under § 14(4) UrhDaG leads to exclusion from the procedure under § 18(3) UrhDaG.
[358] See Art 16 DSA Regulation. The provision applies only to illegal content, not to community standards.
[359] Art 16(2) s 1, 2 lit a) DSA Regulation.
[360] Art 16(2) s 2 lit b) DSA Regulation.
[361] Art 16(2) s 2 lit c) DSA Regulation.
[362] Art 16(2) s 2 lit d) DSA Regulation. The DSA thus waives the requirement for an affidavit. – Pursuant to Art 6(1) lit b) of the DSA Regulation, such notifications trigger an obligation on the part of the service provider to act promptly after becoming aware of them in order to block access to the illegal content or to remove it. If the providers duly comply with this obligation, they are no longer liable for the stored information.
[363] Art 15 of the DSA Regulation recently introduced transparency reporting obligations for providers of moderation services. The provision obliges the latter to make publicly available at least once a year, in a machine-readable format and in an easily accessible manner, clear, easily understandable reports on the content moderation they have carried out during the period in question (para 1, sentence 1). – The transparency reporting obligations under Art 24, 42 DSA Regulation also apply to providers of online platforms – § 2 of the German NetzDG also stipulates a corresponding reporting obligation. – Cf also the voluntarily prepared transparency reports of major online platforms such as: Meta, ‘Community Standards Enforcement Report’ https://transparency.fb.com/reports/community-standards-enforcement/ accessed 31 December 2023 or: Google, ‘Transparency Report’ https://transparencyreport.google.com/ accessed 31 December 2023.
[364] Art 40(4) DSA Regulation vis-à-vis providers of very large online platforms.
[365] Criticism of the ‘information overload’, the confidence-building effect of such information obligations, and the high information costs of the platforms caused by this: N Mamaar, ‘Sorgfalstpflichten der Anbieter von Vermittlungsdiensten’ in Kraul (ed), Das neue Recht der digitalen Dienste: Digital Services Act (Nomos 2023) § 4 para 51-53.
[366] By compensating for disturbed contractual parity between platforms and their innumerable customers, the law of general terms and conditions aims to prevent incentives on the part of the clause user to exploit the clause opponents’ lack of motivation and capacity to examine the general terms and conditions in detail, cf Mast (n 337), (2023) 78(7) JZ (Juristenzeitung) 287 with further evidence.
[367] See Art 14(5) DSA Regulation with its obligation of providers of very large online platforms and of very large online search engines to provide recipients of services with a machine-readable summary of the terms and conditions, including the available remedies and redress mechanisms, in clear and unambiguous language, aimed at reducing complexity.
[368] Questioning the usefulness of transparency obligations: Gielen and Uphues (n 84), (2021) 32(14) EuZW (Europäische Zeitschrift für Wirtschaftsrecht) 627, 636. – On the feasibility of transparency obligations: J Drexl, ‘Bedrohung der Meinungsvielfalt durch Algorithmen’ (2017) 61(7) ZUM (Zeitschrift für Urheber- und Medienrecht) 529, 542; cf also K-N Peifer, ‘Die neuen Transparenzregeln im UWG (Bewertungen, Rankings und Influencer)’ (2021) GRUR (Gewerblicher Rechtsschutz und Urheberrecht) 1453, 1454; F Hofmann, ‘Die neuen Transparenzvorgaben im UWG 2022 im Kontext lauterkeitsrechtlicher Plattformregulierung’ (2022) 124(11) GRUR (Gewerblicher Rechtsschutz und Urheberrecht) 780, 785.
[369] See already above para 95-98.
[370] Art 5 P2B Regulation obliges providers of online intermediary services to present in their general terms and conditions the main parameters determining the ranking and the reasons for the relative weighting of these main parameters compared to other parameters.
[371] According to Art 14(1) s 2, recital 45 s 2 of the DSA Regulation, providers of intermediary services shall include in their general terms and conditions, among other things, information on all guidelines (including permissible content and the manner in which it is presented; requirements for restrictions), procedures (for the restriction of content as well as for internal complaint management), measures and their and tools used for content moderation (such as blocking of access, restriction of visibility or demotion or removal of content), including algorithmic decision-making. Art 27 DSA Regulation requires the disclosure of the parameters of recommendation systems within the meaning of Art 3 lit s) of the Regulation.
[372] Cf, inter alia, Art 15(1) lit b) and c) DSA Regulation.
[373] For example, Art 27, Art 3 lit s) DSA Regulation stipulates an obligation to specify the parameters regarding recommendation systems of online platforms.
[374] Cf Art 15(1) lit a) DSA Regulation.
[375] See Art 14(1) DSA Regulation; Art 11(3), (4) P2B Regulation; Art 11(1) TCO Regulation.
[376] Details: Art 15(1) lit d) DSA Regulation.
[377] Cf under the Digital Services Act: Art 24(5) (although not vis-à-vis decisions to the detriment of rightsholders) as in Germany § 3(2) no 4 NetzDG with a too short storage period of only 10 weeks.
[378] Regulation (EU) 2024/1689 of the European Parliament and of the Council of 13 June 2024 laying down harmonised rules on artificial intelligence and amending Regulations (EC) No 300/2008, (EU) No 167/2013, (EU) No 168/2013, (EU) 2018/858, (EU) 2018/1139 and (EU) 2019/2144 and Directives 2014/90/EU, (EU) 2016/797 and (EU) 2020/1828 (Artificial Intelligence Act, hereinafter: AI Act). Pursuant to Art 113(2) AI Act, the Regulation applies mainly from 2 August 2026. The AI Act was published on 12 July 2024 and enters into force on the twentieth day following its publication in accordance with Art 113(1).
[379] Art 3(4) AI Act.
[380] Art 13(1) s 1 AI Act.
[381] Compare this with regard to the DSA Regulation: B Raue in F Hofmann and B Raue (ed), Digital Services Act (Nomos 2023) Art 14 para 3.
[382] A corresponding updating obligation provides recital 45 s 2 DSA Regulation. – For the whole, see P Leerssen, ‘An end to shadow banning? Transparency rights in the Digital Services Act between content moderation and curation’ (2023) 48 Computer Law & Security Review 105790, 6.
[383] Cf Art 16(6) vis-à-vis the whistleblower (‘provide information’); Art 17(3) lit c) DSA Regulation vis-à-vis the users concerned (but only ‘where applicable’).
[384] It should be criticized that the Digital Services Act in Art 17(1) lit a) does not stipulate an obligation to give reasons for decisions to affected rightsholders – nor to store them in a publicly accessible database pursuant to Art 24(5) – and thus treats decisions to the detriment of the right of personality differently from those to the detriment of freedom of expression. Similarly: K-H Ladeur, ‘Schutz vor Verletzung von Persönlichkeitsrechten und “Desinformation” in sozialen Medien unter Bedingungen der politischen Polarisierung’ in Verfassungsblog https://verfassungsblog.de/personlichkeitsrecht-soziale-medien/ accessed 31 December 2023.
[385] For details, see Art 17(1), (3) DSA Regulation; Art 4(1), (2), (5) P2B Regulation. See further already BGH (Germany), 29 July 2021, III ZR 179/20, (2021) 74(43) NJW (Neue Juristische Wochenschrift) 3179 para 88 f; BGH (Germany), 29 July 2021, III ZR 192/20, (2021) 25(11) ZUM-RD (Zeitschrift für Urheber- und Medienrecht – Rechtsprechungsdienst) 612 para 97 f. Compare to the whole: Nahmias and Perel (n 78), (2021) 58(1) Harvard Journal on Legislation 145, 167 ff; Van Loo (n 100), (2021) 88(4) University of Chicago LRev 829, 841. – See also the right to access meaningful information according to Art 13-25 GDPR: B Casey et al, ‘Rethinking Explainable Machines: The GDPR's “Right to Explanation” Debate and the Rise of Algorithmic Audits in Enterprise’ (2019) 34(1) BERKELEY TECH LJ 145, 158-162.
[386] At the level of the initial decision of a platform, in view of its large number (and corresponding follow-up costs), this does not, a priori, exclude the use of not comprehensively individualized forms of justification.
[387] According to its clear wording, this is also the case in Art 17(1) DSA Regulation.
[388] D K Citron and F Pasquale, ‘The Scored Society: due process for automated predictions’ (2014) 89(1) WASH L REV 1, 20; Nahmias and Perel (n 78), (2021) 58(1) Harvard Journal on Legislation 145, 167 f; cf also Art 22(3) GDPR.
[389] It should therefore be criticized that Art 16(5) DSA Regulation only requires notification of ‘the decision’ and the legal remedies available to the whistleblower, whereas not a statement of the reasons for the decision, which would enable the whistleblower to effectively lodge an appeal under Art 20 DSA Regulation: Principle of effective legal protection based on Art 17(3) lit b), d) and e) DSA Regulation. – In addition, information must be provided on whether a platform has changed its original decision in response to a successful complaint by the user concerned. Similarly: S Gerdemann and G Spindler, ‘Das Gesetz über digitale Dienste (Digital Services Act) (Part 2) – Die Regelungen für Online-Plattformen sowie sehr große Online-Plattformen und -Suchmaschinen’ (2023) 125(3) GRUR (Gewerblicher Rechtsschutz und Urheberrecht) 115, 116.
[390] Cf only Art 6(1) ECHR and Art 47(1), (3) EU-GRCh. In general: H-J Blanke in C Calliess and M Ruffert (ed), EUV/AEUV (6th edn, Beck 2022) Art 47 EU-GRCh para 9; B Hess, Europäisches Zivilprozessrecht (2nd edn, Walter de Gruyter 2021) para 3.67 with further references.
[391] See Art 16(1) s 1 DSA Regulation.
[392] See Art 16(1) s 2 DSA Regulation.
[393] See Art 16(6), recital 52 s 1 DSA Regulation.
[394] See recital 52 s 3 DSA Regulation.
[395] See Art 22 DSA Regulation.
[396] Recital 62, subsec. 2, s 3, 4 DSA Regulation.
[397] For this purpose, already above para 78-89.
[398] B Raue in F Hofmann and B Raue (ed), Digital Services Act (Nomos 2023) Art 22 para 2. The representation of marginalized groups in society is emphasized by N Appelman and P Leerssen, ‘On “Trusted” Flaggers’ (2022) 24 Yale Journal of Law & Technology 452, 469 f.
[399] More on this below para 167-170.
[400] On the acceleration aspect: recital 61 s 1 DSA Regulation.
[401] Art 22(2) DSA Regulation. The digital services coordinator is equally competent to revoke this status in accordance with paragraph 7. – For more details, see below para 144-147.
[402] In this sense, the Digital Services Act in recital 61 s 3, which denies trusted whistleblower status to individuals in principle (‘should not be granted to individuals’). – On self-regulatory cooperation of platforms with trusted whistleblowers (such as YouTube’s Trusted Flaggers program or TikTok Safety Partners) as well as co-regulatory models (such as flagging based on the EU Code of Conduct on Hate Speech), see in more detail Appelman and Leerssen (n 399), (2022) 24 Yale Journal of Law & Technology 452, 454 ff.
[403] Accordingly, the Digital Services Act in its final version, notwithstanding recital 61 s 3 (exclusion of individuals). In contrast, § 14(4) of the German UrhDaG explicitly applies to individual, so-called trusted right holders.
[404] Thus, recital 61 s 4 DSA Regulation mentions Europol in the field of law enforcement.
[405] K Kaesling, ‘Evolution statt Revolution der Plattformregulierung’ (2021) 65(3) ZUM (Zeitschrift für Urheber- und Medienrecht) 177, 180; cf also B Raue and H Heesen, ‘Der Digital Services Act’ (2022) 75(49) NJW (Neue Juristische Wochenschrift) 3537 para 32. On the recognition requirement: B Raue in F Hofmann and B Raue (ed), Digital Services Act (Nomos 2023) Art 22 para 38.
[406] Art 22(3) and (4) DSA Regulation.
[407] To this end, as to the desideratum of feeding reference files into the platforms’ AI moderation system by trusted whistleblowers (similar to the already existing PhotoDNA and Content ID reference databases): Appelman and Leerssen (n 399), (2022) 24 Yale Journal of Law & Technology 452, 473.
[408] However, this is the requirement of the Meta Oversight Board in Case 2022-007-IG-MR – Drill Music, 21 f (‘It is therefore critical that Meta evaluate these requests itself and reach an independent conclusion. [...] Independence is crucial, and the evaluation should require specific evidence of how the content cause harm’) and 23 (‘While there may be good reasons to adopt a prioritization framework that ensures reports from law enforcement are assessed swiftly, that process needs to be designed to ensure that such reports include sufficient information to make independent assessment possible, including seeking further input from the requesting entity or other parties where necessary’).
[409] In German: ‘Grundsatz der Gesetzmäßigkeit der Verwaltung’.
[410] In German: ‘Vorbehalt des Gesetzes’, see the German Constitutional Court (Bundesverfassungsgericht: BVerfG), 3 February 1959, 2 BvL 10/56, BVerfGE 9, 137 = (1959) 12(21) NJW (Neue Juristische Wochenschrift) 931: ‘The principle of the rule of law requires that the administration may only intervene in the legal sphere of the individual if it is authorized to do so by law, and that this authorization is sufficiently determined and limited in terms of content, subject matter, purpose and extent, so that the interventions are measurable and to a certain extent predictable and calculable for the citizen [...]’. Further: BVerfG (Germany), 8 August 1978, 2 BvL 8/77, BVerfGE 49, 89 = (1979) 32(8) NJW 359, 360: ‘The same standards are used to assess whether the legislature, as the constitutional reservation of the right to legislate further requires [...], has itself determined the essential normative foundations of the area of law to be regulated with the norm submitted for review and has not left this to the actions of, for example, the administration’.
[411] Cf the German Constitutional Court (Bundesverfassungsgericht: BVerfG, 20 April 1982, 2 BvL 26/81, BVerfGE 60, 253 = (1982) 35(43) NJW (Neue Juristische Wochenschrift) 2425: ‘Art 19(4), 20(2) s 2 and Section IX of the Grundgesetz [Constitution] prove the rule-of-law idea of binding state power to the law with the establishment of legal protection by independent courts This commitment to the law is indispensable for an order that has placed itself under the claim of the ideas of human dignity, freedom and equality as well as social justice. Freedom requires, above all, the reliability of the legal order. For freedom means above all the possibility of shaping one’s own life according to one’s own life plans. An essential condition for this is that the circumstances and factors which can have a lasting influence on the possibilities of shaping such life plans and their execution, in particular the state's influence on them, can be assessed as reliably as possible’. Cf also BVerfG (Germany), 15 January 1958, 1 BvR 400/51, BVerfGE 7, 198 = (1958) 11(7) NJW (Neue Juristische Wochenschrift) 257 (on judicial observance of a third-party effect of fundamental rights in private law).
[412] Cf recital 61 s 4 DSA Regulation.
[413] F Saurwein, ‘Regulierung von Internet-Inhalten: Ombudsstellen als Governance-Option an der Schnittstelle von Recht und Ethik’, in G Marci-Boehncke, M Rath, M Delere and H Höfer (ed), Medien – Demokratie – Bildung (Springer VS 2022) 47-63 https://doi.org/10.1007/978-3-658-36446-5_5 accessed 31 December 2023.
[414] See only Art 16(3) in connection with Art 6(1) lit b) DSA Regulation.
[415] Thus, the case structure in Meta Oversight Board in Case 2022-007-IG-MR – Drill Music.
[416] Cf also P Schneiders, ‘Hate Speech auf Online-Plattformen: Problematization, Regulation and Evaluation against the Background of the Proposal for a Digital Services Act’ (2021) 85(2) UFITA (Archiv für Medienrecht und Medienwissenschaft) 269, 303 ff https://doi.org/10.5771/2568-9185-2021-2-269 accessed 31 December 2023.
[417] Similarly, Art 11 subpara 2 s 2 P2B Regulation for the internal complaint management system for business users. On the whole: M Berberich, ‘§ 5 Sorgfaltspflichten, Moderationsverfahren und prozedurale Fairness’ in B Steinrötter (ed), Europäische Plattformreguliereng (Nomos 2023) para 56; D Holznagel, ‘Zu starke Nutzerrechte in Art. 17 und 18 DSA’ (2022) 38(9) CR (Computer und Recht) 594, 598.
[418] B Raue in F Hofmann and B Raue (ed), Digital Services Act (Nomos 2023) Art 20 para 47.
[419] See recital 52 s 3 DSA Regulation, but only in the case of a platform-based enforcement of state-granted rights; an analogous application of this principle to the enforcement of platform-owned standards suggests itself when the protection of third-party interests is at issue.
[420] Cf recital 87 s 7 DSA Regulation for moderation decisions of very large online platforms; also, B Raue in F Hofmann and B Raue (ed), Digital Services Act (Nomos 2023) Art 20 para 47.
[421] So also the Facebook Oversight Board, ‘Art. 2 section 1 of the FOB Charter’ https://about.fb.com/wp-content/uploads/2019/09/oversight_board_charter.pdf accessed 31 December 2023.
[422] For more details, see below para 153-154.
[423] Citron and Pasquale (n 389), (2014) 89(1) WASH L REV 1, 20.
[424] Other view: Van Loo (n 183), (2016) 33(2) Yale Journal on Regulation 547, 565 f; Van Loo (n 100), (2021) 88(4) University of Chicago LRev 829, 876 f. On the similar issue of equal access rights to the legal services market, D Simshaw, ‘Access to A.I. Justice: Avoiding an Inequitable Two-Tiered System of Legal Services’ (2022) 24 Yale Journal of Law & Technology 150, 183 ff.
[425] So the understanding under German law, see the Federal Court of Justice (Bundesgerichtshof), 29 July 2021, III ZR 192/20, (2021) 25(11) ZUM-RD (Zeitschrift für Urheber- und Medienrecht – Rechtsprechungsdienst) 612 para 66, 77, 80; B Raue in F Hofmann and B Raue (ed), Digital Services Act (Nomos 2023) Art 14 para 90.
[426] This is also explicitly pointed out by recitals 45, 52 s 2 DSA Regulation.
[427] Thus Art 14(1) s 2 as well as recitals 45, 58 s 2 DSA Regulation. – Cf also: Van Loo (n 100), (2021) 88(4) University of Chicago LRev 829, 876; Lester and Pachamanova (n 41), (2017) 24(1) UCLA Entertainment Law Review 51, 68.
[428] Art 16(6) DSA Regulation permits the use of automated means for processing reports and decision-making in the notification and redress procedure. However, information must be provided about their use (Art 16(6) s 1 DSA Regulation). – In the case of the internal complaints procedure, on the other hand, Art 20(6) DSA Regulation obliges all providers of online platforms to ensure that relevant decisions are made under the supervision of appropriately qualified personnel and not solely by automated means – Cf Mamaar (n 366) § 4 para 71; Spindler (n 339), (2021) 123(4) GRUR (Gewerblicher Rechtsschutz und Urheberrecht) 545, 552.
[429] Nevertheless, the capacity problem of platforms to maintain sufficient staff to review decisions remains.
[430] Of course, there are limits to the context sensitivity of human decision-makers At least with appropriately qualified personnel, however, they are likely to remain superior to algorithmic decision-making in this respect, given the current state of the art.
[431] One reason for this may be that the use of human decision-makers can prevent the presumed feeling of not being ‘at the mercy of a machine’, see also R Koulu, ‘Proceduralizing control and discretion: Human oversight in artificial intelligence policy’ (2020) 27(6) Maastricht Journal of European and Comparative Law 720, 722, 729; about the importance of explained and socially accepted decisions see also F von Ameln, ‘Führen und Entscheiden unter Unsicherheit’ (2021) 52(4) GIO (Gruppe. Interaktion. Organisation. Zeitschrift für Angewandte Organisationspsychologie) 567, 570.
[432] For corresponding transparency obligations of providers of very large online search engines, see Art 42(2) lit a) and b) DSA Regulation.
[433] The Oversight Board of the Meta Group is a paradigm for this Here, decisions are made by panels consisting of five Board members, see Art 3, in particular Sections 2, 4; Art 7.1 Charter; Art 1, Section 1.1.4.4, 3.1.3 (‘Standard Cases’), Art 2, Section 2.1.2 (‘Expedited Review’) and 2.1.3 (‘Summary Decisions’) Oversight Board Bylaws.
[434] This is the case, for example, in the Network Enforcement Act counter-proceedings (§ 3b NetzDG), but only for illegal content.
[435] To this end: S Cooper, C Rule and L Del Duca, ‘From Lex Mercatoria to Online Dispute Resolution’ (2011) Penn State Legal Studies Research Paper No 09/2011, 13; also, C Busch, ‘Mehr Fairness und Transparenz in der Plattformökonomie?’ (2019) 121(8) GRUR (Gewerblicher Rechtsschutz und Urheberrecht) 788, 796; Mast (n 337), (2023) 78(7) JZ (Juristenzeitung) 287, 291 f; A Ohly in A Ohly and O Sosnitza (ed), UWG Gesetz gegen den unlauteren Wettbewerb: UWG (8th edn, Beck 2023) § 8a para 1, 2.
[436] R Van Loo argues from a U.S. perspective that enforcement decisions of platforms should at the same time meet the standard of ‘reputational accuracy and completeness’, in Van Loo (n 100), (2021) 88(4) University of Chicago Law Review 829, 883 f.
[437] See Art 14(4) DSA Regulation: ‘Providers of intermediary services shall act in a diligent, objective and proportionate manner in applying and enforcing the restrictions referred to in paragraph 1, with due regard to the rights and legitimate interests of all parties involved, including the fundamental rights of the recipients of the service, such as the freedom of expression, freedom and pluralism of the media, and other fundamental rights and freedoms as enshrined in the Charter’. – Art 16(6) s 1 DSA Regulation: ‘Providers of hosting services shall process any notices that they receive under the mechanisms referred to in paragraph 1 and take their decisions in respect of the information to which the notices relate, in a timely, diligent, non-arbitrary and objective manner’.
[438] On this point, in more detail: Van Loo (n 100), (2021) 88(4) University of Chicago LRev 829, 871.
[439] Cf also recital 52 s 1 DSA Regulation: ‘[...] on the basis of rules that are uniform, transparent and clear [...]’.
[440] Cf also T R Tyler, ‘What is procedural justice? Criteria used by citizens to assess the fairness of legal procedures’ (1988) 22(1) Law & Society Review 103, 105: ‘Consistency refers to similarity of treatment and outcomes across people or time or both’.
[441] These are, for example, the facts of the so-called spirit of policy allowance and the newsworthiness allowance; on the latter doctrine, taken from U.S. constitutional law and applied to Facebook’s content moderation, see T Kadri and K Klonick, ‘Facebook v. Sullivan: Public Figures and Newsworthiness in Online Speech Online Speech’ (2019) 93(1) Southern California Law Review 37-99.
[442] Oversight Board, 9 March 2023, 2022-014-FB-MR – Sri Lanka Pharmaceuticals, Part 8.1.I, 14 (‘Secret discretionary exemptions to Meta’s policies are incompatible with the legality standard.’); further: Oversight Board, 27 September 2021, 2021-010-FB-UA – Colombia Protests, Part 6 (‘The Board notes that Facebook does not make its criteria for escalation publicly available.’); affirmed in: Oversight Board, 17 June 2022, 2022-001-FB-UA – Knin Cartoon, 4, 18; 4: ‘The fact that the content was not sent to Meta’s specialized teams for assessment before it reached the Board shows that the company’s processes for escalating content are not sufficiently clear and effective’; Oversight Board, 14 December 2022, 2022-012-IG-MR – India Sexual Harassment Video, under Part 8.1 and 8.3: The Board criticizes the process and, in particular, access to the escalation process as not sufficiently clear and effective.
[443] Oversight Board, 17 June 2022, 2022-001-FB-UA – Knin Cartoon, Oversight Board, 17 June 2022, 2022-001-FB-UA, Part 8.1 (Meta’s review process), 17.
[444] Cf also the ‘Media Matching Service Bank’ (‘escalations bank’) used by the Meta Group. Content classified as ‘violating’ is fed into this database as part of the escalation procedure, and all subsequent identical or ‘core-similar’ content (‘matching content’) is automatically sanctioned on this basis Affected users can file a complaint against this, see for example Oversight Board, 14 December 2022, 2022-011-IG-UA – Video after Nigeria Church Attack, About the Case, 2: The FOB reversed Meta’s decision to remove a video from Instagram showing the situation immediately following a 5 June 2022 terrorist attack in Nigeria. The panel concluded that by restoring the post with a warning message (‘disturbing content’), the privacy of the victims is protected while allowing a discussion of the events that some states may wish to prevent.
[445] Cf Oversight Board, 9 January 2023, 2022-013-FB-UA – Iran Protest Slogan, Parts 6 and 8.1 I. b), 14.
[446] The entire content moderation system structurally exhibits elements of scaling. In regular moderation, this is evident in the technically simple mechanism of hash matching, for example. Here, too, existing content classified as sanction-worthy – such as words or images – is fed into a database, the corresponding content is detected, and this decision is then replicated (for the technical mode of operation, see above para 30-38.
[447] Scaled exceptions apply to entire categories of content, not just individual posts – Sri Lanka Pharmaceuticals, the Oversight Board (9 March 2023, 2022-014-FB-MR, Part 8.2, 14) upheld Meta’s decision to allow a Facebook post soliciting drug donations for Sri Lanka to be published when there was a financial crisis there. Meta concluded that under a strict interpretation of the rule, the post violated the Community Standard on Restricted Goods and Services This prohibits content asking for medications However, the FOB applied a scaled ‘within the meaning of the policy’ exception, but in doing so found that undisclosed, arbitrary policy exceptions were inconsistent with Meta’s human rights responsibilities To make ‘in the spirit of the policy’ exceptions more transparent and consistent, the Board made recommendations including that, if applied consistently, they be standardized in the relevant community standards themselves, including the criteria Meta uses to decide whether to scale the exception.
[448] Art 4 FOB Charter; Art 3, Section 2.3.1 FOB Bylaws For the detailed implementation process of the Board’s decisions by dividing them into the different levels of comparability such as ‘Case Content’, ‘Identical Content with Parallel Context’, and ‘Similar Content’, see Meta, ‘Sharing More Details on How We Will Implement the Oversight Board's Decisions, Responding to the Oversight Board’s First Decisions’, 28 January 2021 https://about.fb.com/news/2021/01/responding-to-the-oversight-boards-first-decisions/ accessed 31 December 2023.
[449] Moreover, there is a corporate reporting requirement on this issue, E Douek, ‘How Much Power Did Facebook Give Its Oversight Board?’, in Lawfare, 25 September 2019 https://www.lawfaremedia.org/article/how-much-power-did-facebook-give-its-oversight-board accessed 31 December 2023.
[450] According to E Douek, even representatives of the Meta Group estimated the number of such identically stored contents to be very small due to the context-specificity. This assessment was made in the context of a discussion between representatives of the Meta Group and ‘stakeholders’ which Douek also attended: ‘The Oversight Board Moment You Should've Been Waiting For’, in Lawfare, 26 February 2021 https://www.lawfaremedia.org/article/oversight-board-moment-you-shouldve-been-waiting-facebook-responds-first-set-decisions accessed 31 December 2023.
[451] CJEU, 3 October 2019, C-18/18 – Glawischnig-Piesczek, ECLI:EU:C:2019:821, para 33 ff.
[452] On the instrument of geo-blocking see: R Achleitner, ‘The Fight against Geo-Blocking – A Never Ending Story? Policy Paper on Geo-Blocking’, 2 February 2021 https://ssrn.com/abstract=4246896> or <http://dx.doi.org/10.2139/ssrn.4246896 both accessed 31 December 2023.
[453] See below para 167-178.
[454] The guiding consideration should also be how likely scaling is to cause serious, irreversible damage.
[455] Regulation (EU) 2024/1689 of the European Parliament and of the Council of 13 June 2024 laying down harmonised rules on artificial intelligence and amending Regulations (EC) No 300/2008, (EU) No 167/2013, (EU) No 168/2013, (EU) 2018/858, (EU) 2018/1139 and (EU) 2019/2144 and Directives 2014/90/EU, (EU) 2016/797 and (EU) 2020/1828 (Artificial Intelligence Act, hereinafter: AI Act). Pursuant to Art 113(2) AI Act, the Regulation applies mainly from 2 August 2026. The AI Act was published on 12 July 2024 and enters into force on the twentieth day following its publication in accordance with Art 113(1).
[456] Cf Art 2(1) AI Act. According to Art 3(3) AI Act, ‘provider’ means ‘a natural or legal person, public authority, agency or other body that develops an AI system or a general-purpose AI model or that has developed an AI system or a general-purpose AI model and places it on the market or puts the AI system into service under its own name or trademark, whether for payment or free of charge’.
‘Deployer’ is defined in accordance with Art 3(4) as ‘a natural or legal person, public authority, agency or other body using an AI system under its authority except where the AI system is used in the course of a personal non-professional activity’.
[457] Annex III no 8 lit a) AI Act.
[458] Proposal for a Regulation of the European Parliament and of the Counc of 21 April 2021, COM(2021) 206 final, Annex III no 8 lit a).
[459] In general on the autonomous interpretation of the law of the Union: K Riesenhuber, ‘§ 10 Die Auslegung‘ in K Riesenhuber (ed) Europäische Methodenlehre (De Gruyter 2021) 285 para 4 ff; R Stotz, ‘Die Rechtsprechung des EuGH’ in K Riesenhuber (ed) Europäische Methodenlehre (De Gruyter 2021) 653 para 19.
[460] Directive 2013/11/EU of the European Parliament and of the Council of 21 May 2013 on alternative dispute resolution for consumer disputes and amending Regulation (EC) No 2006/2004 and Directive 2009/22/EC (Directive on consumer ADR), OJ L 165, 18 June 2013, 63.
[461] G Rühl, ‘The Alternative Dispute Resolution Directive: Handlungsperspektiven und Handlungsoptionen’ (2014) 127(1) ZZP (Zeitschrift für Zivilprozess) 61, 67. So also the implementation of the Directive in the German Verbraucherstreitbeilegungsgesetz (Consumer Dispute Resolution Act), there § 1(1).
[462] Recital 5 s 1 Directive 2013/11/EU, albeit limited to disputes arising from sales contracts or service contracts between consumers and traders. See also Art 2(1) Directive 2013/11/EU. – Even more precise: Commission Recommendations 98/257/EC of 30 March 1998 on the principles applicable to the bodies responsible for out-of-court settlement of consumer disputes, OJ L 115, 17 April 1998, 31, 32: ‘Whereas this recommendation must be limited to procedures which, no matter what they are called, lead to the settling of a dispute through the active intervention of a third party, who proposes or imposes a solution’. This definition was adopted from recitals 3 s 2 and 9 s 1 Commission Recommendation 2001/310/EC of 4 April 2001 on the principles for out-of-court bodies involved in the consensual resolution of consumer disputes (Text with EEA relevance) (notified under document number C(2001) 1016), OJ L 109, 19 April 2001, 56.
[463] See recital 32, Art 1 s 1, Art 2(2)(a) and Art 6 Directive 2013/11/EU. Cf also Commission Recommendations 98/257/EC of 30 March 1998 on the principles applicable to the bodies responsible for out-of-court settlement of consumer disputes, OJ L 115, 17 April 1998, 31, 32: English: ‘essential’; German: ‘unerlässliche Voraussetzung’; French: ‘qualités nécessaires’. Also: Commission Recommendation 2001/310/EC (n 463), II.A.
[464] Recital 61 s 1 (German version): ‘Bestimmte KI-Systeme, die für die Rechtspflege und demokratische Prozesse bestimmt sind, sollten angesichts ihrer möglichen erheblichen Auswirkungen auf die Demokratie, die Rechtsstaatlichkeit, die individuellen Freiheiten sowie das Recht auf einen wirksamen Rechtsbehelf und ein unparteiisches Gericht als hochriskant eingestuft werden’. French version: ‘tribunal impartial’.
Recital 4: ‘The use of AI tools can support the decision-making power of judges or judicial independence, but should not replace it’.
[465] Art 5(1), Art 6 (1), (2) Directive 2013/11/EU.
[466] The fact that the AI Act restricts the use of high-risk AI systems to functionally judicial proceedings also follows indirectly from its recital 48 s 2: ‘Those rights include [...] the right to an effective remedy and to a fair trial [...]’.
[467] Art 2(2)(b) Directive 2013/11/EU. See also recital 9 s 4 Commission Recommendation 2001/310/EC (n 463), II.A.
[468] Cf Art 2(1) Directive 2013/11/EU, German version: ‘Beilegung von [...] Streitigkeiten’; English version: ‘resolution of [...] disputes’; French version: ‘une solution, ou réunit les parties en vue de faciliter la recherche d'une solution amiable’.
[469] See above para 144-147 with further references.
[470] German version: ‘die Ergebnisse [...] Rechtswirkung für die Parteien entfalten’; French version: ‘les résultats [...] produisent des effets juridiques pour les parties’.
[471] Arg Art 86(1) AI Act: ‘[...] and which has legal effects or significantly affects them in a similar manner [...]’.
[472] Thus, expressly recital 26 AI Act.
[473] M von Welser, ‘Die KI-Verordnung – ein Überblick über das weltweit erste Regelwerk für künstliche Intelligenz’ (2024) 16(15) GRUR-Prax (Gewerblicher Rechtsschutz und Urheberrecht in der Praxis) 485; regarding the Proposal AI Act: I Orssich, ‘Das europäische Konzept für vertrauenswürdige Künstliche Intelligenz’ (2022) 33(6) EuZW (Europäische Zeitschrift für Wirtschaftsrecht) 254, 255.
[474] Recital 61 S 1 AI Act, German verson: ‘erhebliche Auswirkungen’; French version: ‘incidence [...] significative’.
[475] See recital 48 s 2 AI Act.
[476] See above 1.2 (Dangers and drawbacks of AI-based law and standards enforcement), para 67 ff.
[477] See above 1.3.2.3 (Intra-company legal protection proceedings: Basic structures and procedural guarantees), para 127 f.
[478] Cf the German Federal Court of Justice (Bundesgerichtshof), 25 October 2011, VI ZR 93/10 – Blogeintrag (2012) 114(3) GRUR (Gewerblicher Rechtsschutz und Urheberrecht) 311 para 25-27 (right of personality). Van Loo (n 183), (2016) 33(2) Yale Journal on Regulation 547, 566 f (‘third-party adjudicator’, ‘network trial’, ‘various court-like roles’), 576 (2016); Van Loo (n 100), (2021) 88(4) University of Chicago LRev 829, 832 (‘most important private judicial system’) 846 (‘quasi-judicial role’ with respect to review and enforcement of the ‘right to be forgotten’ by search engines such as Google), 849, 850 (‘The expanded privatization of U.S. justice through platforms' internal dispute systems deserves scrutiny’), 865; Haber (n 217), (2016) 40(1) Seattle University Law Review 115, 129 ff; D Holznagel, ‘Melde- und Abhilfeverfahren zur Beanstandung rechtswidrig gehosteter Inhalte nach europäischem und deutschem Recht im Vergleich zu gesetzlich geregelten notice and take-down-Verfahren’ (2014) 63(2) GRUR Int (Gewerblicher Rechtsschutz und Urheberrecht Internationaler Teil) 105, 108 (‘Judge’s Role’); see also F Hofmann and T Sprenger, ‘Privatization of Enforcement’ (2021) 85(2) UFITA (Archiv für Medienrecht und Medienwirtschaft) 249, 254 (‘to settle disputes’).
[479] Cf Art 16(4), (6); Art 14(4), Art 23(3) and recitals 24 s 3, 26 s 2 DSA Regulation.
[480] See already Laukemann (n 272) 276 f. On the structural bias of the U.S. tourism website Trip-Advisor due to the dependence of its business model on advertising revenue from user-rated companies, see Van Loo (n 100), (2021) 88(4) University of Chicago LRev 829, 869.
[481] Cf Radu (n 262) 179: ‘Local values representation is the second point of contention towards Facebook community. The unilateral definition of what is and what is not acceptable online by a company headquartered in the United States is harder to sustain as more than 2 billion people use the platform. Facebook's largest user base at the moment is India, but little of the social and cultural norms there appear to transpire in the global policy of the company’.
[482] On the legal classification of terms of use and community standards as GTCs, see only German Federal Court of Justice (Bundesgerichtshof), 29 July 2021, III ZR 192/20, (2021) 25(11) ZUM-RD (Zeitschrift für Urheber- und Medienrecht – Rechtsprechungsdienst) 614, para 44 (§§ 305 BGB ff are applicable); further: M Mayer, Soziale Netzwerke im Internet im Lichte des Vertragsrechts (Boorberg 2018) 120, 359. – Cf also the broad definitions in Art 3 lit u) DSA Regulation; Art 2 no 10 P2B Regulation and Art 2 no 8 TCO Regulation.
[483] 33.700 of all 34.806 pieces of content deleted or blocked by Facebook in the relevant period, ie just under 97%, already (possibly only) violate community standards and are therefore deleted worldwide. Only the remaining share of 3.2% (= 1.106) falls through the ‘international grid’ and is thus only blocked in Germany, see Meta, ‘Facebook Transparency Report of January 2023’ 19 327151920_907084790305794_6193992151844220602_n.pdf (fbcdn.net) accessed 28 September 2023.
[484] On this aspect, see Nahmias and Perel (n 78), (2021) 58(1) Harvard Journal on Legislation 145, 178, referring to Radu (n 262) 179.
[485] § 3(6) no 3 of the German Network Enforcement Act (NetzDG) links the recognition of an institution as a so-called institution of regulated self-regulation, among other things, to the existence of rules of procedure that regulate submission obligations of the affiliated social networks.
[486] For Facebook, the remediation rate for the period January to March 2022 is about 8.3% (587.000 appeals to 48.700 remediations); but in some cases also proactive correction due to previous, similar violations.
[487] These complaints were forwarded to the FSM (Freiwillige Selbstkontrolle Multimedia-Diensteanbieter e.V.) in the period between 1 July 2022 and 31 December 2022: Meta, ‘Facebook Transparency Report of January 2023 for the 2nd HY 2022’ 14 327541832_1414754302684176_3061551644115119140_n.pdf (fbcdn.net) accessed 31 December 2023.
[488] In the first half of 2023, YouTube submitted 12 of 193.131 reported content: Google, ‘Transparency Report for YouTube Platform for January to June 2023’, 4, 7 (https://transparencyreport.google.com/netzdg/youtube?hl=de). X, formerly Twitter, submitted 66 complaints to German law firms during the same period: Twitter, ‘Network Enforcement Report: January-June 2023’, 25 https://transparency.twitter.com/content/dam/transparency-twitter/country-reports/germany/NetzDG-Jan-Jun-2023.pdf accessed 31 December 2023.
[489] This is the case if platforms can only avoid a fine if the decision is expected to take longer by submitting the complaint to an out-of-court dispute resolution body before the expiry of a decision deadline (of 7 days). Similarly, the solution of the German NetzDG in § 3(2) s 1 no 3 lit b): ‘The procedure must ensure that the social network provider removes or blocks access to any unlawful content without undue delay, usually within seven days of receipt of the complaint; the seven-day period may be exceeded if […] b) within seven days of receipt of the complaint, the social network provider transfers the decision on the unlawfulness to a body of regulated self-regulation recognized in accordance with paragraphs 6 to 8 and submits to its decision [...]’. However, a violation of the procedural requirements in a specific individual case is not sufficient to warrant a fine (Bundestags-Drucksache 18/12356, 24; M Liesching, Netzwerkdurchsetzungsgesetz (Nomos 2018) § 4 para 13. The exact requirements for going beyond the individual case remain unclear. – The Austrian KoPlG does not recognize such an outsourcing of the decision to an independent body.
[490] See most recently Art 21(1) DSA Regulation, with regard to certified out-of-court dispute resolution bodies, which are to be accredited by Member State coordinators for digital services (Art 21(3) DSA Regulation; revocation of accreditation under paragraph 7).
[491] For both the enforcement addressee and the rightsholder, state legal protection should thus not only be possible downstream, as was recently also the case in Art 21(1) subpara 3 DSA Regulation.
[492] According to the DSA, the relevant criteria for authorization include, in particular, the independence and impartiality of the body, as well as expertise and clear and fair rules that enable an easily accessible procedure aimed at efficient, ie, rapid and cost-effective, dispute resolution. According to the cost regulation of Art 21(5) DSA Regulation, which is advantageous for both users and notifiers, in the event of an unsuccessful complaint, the latter shall in principle only bear their own fees and other reasonable costs, but not those of the online platform. In the opposite case, the online platform bears the full cost burden, including the costs of the prevailing opposing party. – In addition, the DSA stipulates a duty to inform the platforms about the possibility of appealing to such institutions.
[493] The dispute resolution body could be called upon in two ways: on the one hand as a kind of ‘complaints authority’ by users or whistleblowers, and on the other, on mandatory submission by the platform in the event of an unclear legal or infringement situation.
[494] According to Art 21(4) subpara 3 DSA Regulation, a decision should in principle not take longer than 90 calendar days, and in the case of ‘highly complex’ disputes a maximum of 180 days.
[495] See Hofmann and Specht-Riemenschneider (n 348), (2021) 13(1) ZGE (Zeitschrift für geistiges Eigentum) 48, 103 f, who in this context refer to a study by Fiala and Husivec, according to which the risk of overblocking can be significantly reduced by using out-of-court dispute resolution mechanisms: F Fiala and M Husivecm, ‘Using Experimental Evidence to Improve Delegated Enforcement’ (2018) International Review of Law and Economics, Forthcoming TILEC Discussion Paper no 2018-028, 25. – It is therefore to be criticized if Art 21(2) subpara 3 DSA Regulation explicitly denies a corresponding binding effect.
[496] Art 21(4) and Art 24(5) DSA Regulation refrain from this and only require the bodies to report annually to the Digital Services Coordinator on the number, duration and outcome of disputes the lack of storage of these often-important decisions under the DSA is to be criticized. – As a flanking measure, Art 24(1) lit a) DSA Regulation obliges online platform providers to report on the number of disputes submitted to the out-of-court dispute resolution bodies referred to in Art 21, the results of dispute resolution and mediation duration until the conclusion of dispute resolution proceedings, as well as the proportion of disputes in which the online platform providers implemented the body’s decisions.
[497] Art 21(2) subpara 3 DSA Regulation. Critical: D Holznagel in R Müller-Terpitz and M Köhler (ed), Digital Services Act (Beck 2024) Art 21 para 39-41.
[498] Art 6(2), Annex III no 8 lit a) AI Act. See in more detail above para 162 ff.
[499] See in more detail above para 161 ff.
[500] Cf P McColgan, ‘Das wird man wohl noch löschen dürfen? – Control Standards for Opinion Rules on the Internet’ (2021) 1(12) RDi (Recht Digital) 605, 610 f, 615 f; Mast (n 337), (2023) 78(7) JZ (Juristenzeitung) 287, 292; further Klonick (n 94), (2018) 131(6) Harvard Law Review 1598, 1639.
[501] Mast (n 337), (2023) 78(7) JZ (Juristenzeitung) 287, 292.
[502] For more information on this and on the linkage of § 307(1) BGB (German Civil Code) to the standards of European Union law, see Mast (n 337), (2023) 78(7) JZ (Juristenzeitung) 287, 290. – In the case of the review of unlawful platform GTCs by state courts – for example in German law on the basis of the Unterlassungsklagegesetz (UKlaG) or supplementary pursuant to Art 14(1) P2B Regulation, court decisions have a broad effect, for example under § 11 UKlaG or by means of a nullity order against GTCs that violate Art 3(1) P2B Regulation (thus Art 3(3) in conjunction with recital 20 of the P2B Regulation).
[503] Thus aptly: Mast (n 337), (2023) 78(7) JZ (Juristenzeitung) 287, 289, 292, 295, who in this respect speaks of the ‘normative force of platform GTCs’.
[504] See initially K-H Ladeur, ‘Neue Institutionen für den Daten- und Persönlichkeitsschutz im Internet: „Cyber-Courts“ für die Blogosphere‘ (2012) 36(10) DuD (Datenschutz und Datensicherheit) 711 ff; further: Van Loo (n 100), (2021) 88(4) University of Chicago Law Review 829 ff.
[505] Cf also Ladeur (n 385).
[506] On identity and personality formation as a communicative process, see D Wielsch, ‘Medienregulierung durch Persönlichkeits- und Datenschutzrechte’ (2020) 75(3) JZ (Juristenzeitung) 105, 107 ff.
[507] T Wu argues that in order to cope with the mass problem, also for reasons of optimized resource allocation, simple cases (‘easy cases’: simple facts; clear, little context-dependent legal situation) should be decided by means of automated moderation, while complex case constellations (‘hard cases’: context-, consideration-intensive and complex legal situations) the decision should be reserved for humans – especially committees, in: T Wu, ‘Will Artificial Intelligence Eat the Law? The Rise of Hybrid Social-Ordering Systems’ (2019) 119(7) Columbia Law Review 2001 ff; see also M Denga, ‘Platform Regulation by European Values: On the Binding of Opinion Platforms to EU Fundamental Rights’ (2021) 56(5) EuR (Europarecht) 569, 572. – In contrast, the approach is to reduce the scrutiny of significant communication processes by raising the threshold of infringing conduct within social media. At the same time, the concretization of platform-typical communication customs contributes to the formation of area-specific rules, see on the whole: Ladeur (n 385).
[508] Van Loo (n 100), (2021) 88(4) University of Chicago Law Review 829, 867 ff. However, the link to precedents should not be strict so as not to block innovative adaptations to a rapidly changing online environment.
[509] Ladeur (n 385), (2012) 36(10) DuD (Datenschutz und Datensicherheit) 711, 714.
[510] Comparisons to a cross-provider ‘cyber court of second instance’, with greater consideration of regional characteristics: Ladeur (n 385). – In contrast, R Van Loo does not want to take the word of the idea of a sector-wide uniform private regulatory regime – for example, for social media platforms. Consequently, he rejects the idea of (strict) cross-platform precedence. Instead, he considers taking his cue from the common law model of persuasive authority of foreign court decisions. Similar to the idea of a ‘market for rules’, platforms should compete with each other: Van Loo (n 100), (2021) 88(4) University of Chicago Law Review 829, 867 f; also McColgan (n 501), (2021) 1(12) RDi (Recht Digital) 605, 613. Critical of the idea of competition in view of a monopolization of the platform market, in turn, the Union legislature on the occasion of the Digital Services Act, on this L Kumkar, ‘Plattform-Recht revisited: Umgang mit den Marktordnungen digitaler Plattformen de lege lata et ferenda’ (2022) 30(3) ZEuP (Zeitschrift für Europäisches Privatrecht) 530, 551; B Raue, ‘Plattformnutzungsverträge im Lichte der gesteigerten Grundrechtsbindung marktstarker sozialer Netze’ (2022) 75(4) NJW (Neue Juristische Wochenschrift) 209 para 4, referring to the lock-in and network effect for users. – M Land also argues in favor of different rules adapted to the practices of different platform types, in: M Land, ‘The Problem of Platform Law: Pluralistic Legal Ordering on Social Media’, in P SA Berman (ed), The Oxford Handbook of Global Legal Pluralism (2020) 974.
[511] On this parallel: Ladeur (n 385).
[512] Directive 2008/52/EC of the European Parliament and of the Council of 21 May 2008 on certain aspects of mediation in civil and commercial matters, Official Journal of the European Union from 24 May 2008, L 136/3 ff.
[513] See also D Rodi, in Staudinger-BGB, Buch 2 (19th rev edn, De Gruyter 2022) Anh. zu §§ 305-610 BGB M 33; M Fehrenbach, in BeckOGK-BGB (Beck 1 November 2023) § 307 Schlichtungsklausel para 4; P Röthemeyer, ‘Die Schlichtung‘ (2013) 16(2) ZKM (Zeitschrift für Konfliktmanagement) 47, 49.
[514] Röthemeyer (n 514), (2013) 16(2) ZKM (Zeitschrift für Konfliktmanagement) 47, 48; M Fehrenbach, in BeckOGK-BGB (Beck 1 November 2023) § 307 Schlichtungsklausel para 4.
[515] M Fehrenbach, in BeckOGK-BGB (Beck 1 November 2023) § 307 Schlichtungsklausel para 4; R Greger und C Stubbe, Schiedsgutachten (1st edn, Beck 2007) para 27.
[516] See R Greger, ‘D Recht der alternativen Konfliktlösung’ in R Greger, H Unberath and F Steffek (ed), Recht der alternativen Konfliktlösung (2nd edn, Beck 2016) para 245.
[517]However, according to § 19 of the German Consumer Dispute Resolution Act (Verbraucherstreitbeilegungsgesetzes: VSBG), the accepted conciliation proposal is legally binding. – In contrast, the result of mediation can also end in the conclusion of a settlement agreement, see F Kreis, ‘KI und ADR-Verfahren’ in M Kaulartz and T Braegelmann (ed), Rechtshandbuch Artificial Intelligence und Machine Learning (2020) 633, 638 para 17 f.
[518] S J Heetkamp and C Piroutek, ‘ChatGPT in Mediation und Schlichtung‘ (2023) 26(3) ZKM (Zeitschrift für Konfliktmanagement) 80.
[519] T Deichsel, Digitalisierung der Streitbeilegung (1st edn, Nomos 2022) 201 f; C Leeb, Digitalization, Legal Technology and Innovation (1st edn, Dr. Otto Schmidt 2019) 238.
[520] When using legal advice chatbots in Germany, the limits of the Legal Services Act (Rechtsdienstleistungsgesetz: RDG) must be observed: For example, in the legal services market, non-lawyer legal advice is prohibited as soon as the threshold of independent provision of legal services is exceeded, § 2(1) RDG. – In the case of debt collection services, § 10(1) s 1 no 1 RDG and § 6(1) RDG must be observed, which define the limits of the permissible scope of activities Accordingly, legal advice (free of charge) provided by private companies such as legal techs is not permitted. A debt collection service provider’s authority to provide advice is limited to the activity of asserting claims – in particular, it is therefore not comprehensive, see in more detail M Hartung, ‘Sonstige Akteure und Rahmenbedingungen’ in M Hartung, M-M Bues and G Halbleib (ed), Legal Tech: Die Digitalisierung des Rechtsmarkts (1st edn, Beck 2018) 215 para 1044. The question of whether and under what conditions chatbots can even provide legal services within the meaning of § 2 RDG is still highly controversial. In particular, it is unclear to what extent the use of artificial intelligence constitutes a ‘legal examination of the individual case’ (pursuant to § 2(1) RDG) in view of its technical functionality. This, in turn, is partly denied due to the lack of subsumption. Programming abstract legal decision trees is only an abstract activity, not an examination of the concrete legal situation, cf C Deckenbrock and M Henssler, in C Deckenbrock and M Henssler (ed), Rechtsdienstleistungsgesetz: RDG (5th edn, Beck 2021) § 2 para 54g. By contrast, it will be necessary to differentiate: If only abstract information is provided or a chatbot is only used to establish the facts of the case, there is no requirement for a legal review. However, as soon as legal information is provided that is adapted to the data previously entered by the person seeking legal advice, the threshold for a specific examination of the individual case is likely to be exceeded, see also B Brechmann, Legal Tech und das Anwaltsmonopol (Mohr Siebeck 2021) 61; F Remmertz and M Krenzler, in M Krenzler amd F Remmertz (ed), Rechtsdienstleistungsgesetz(3rd edn, Nomos2023) § 2 para 71a; see also the final report of the State Working Group: Abschlussbericht der Länderarbeitsgruppe, ‘Legal Tech: Herausforderungen für die Justiz’ (2019) 40 f, https://www.schleswig-holstein.de/DE/landesregierung/ministerien-behoerden/II/Minister/Justizministerkonferenz/Downloads/190605_beschluesse/TOPI_11_Abschlussbericht.pdf?__blob=publicationFile&v=1. – If legal chatbots are used in court-connected dispute resolution, the scope of application of the RDG – which is limited to out-of-court legal services – is already excluded, § 1(1) s 1 RDG. The activities of conciliation boards, arbitrators as well as mediation and any comparable form of alternative dispute resolution – insofar as the activity does not intervene in the discussions of the parties involved by proposing legal regulations – do not constitute legal services within the meaning of § 2(2) no 2, 4 RDG. For the whole, see Leeb (n 520) 75, 280 ff.
[521] Large language models such as ChatGPT are based on machine learning technology and form neural networks that enable the AI system to answer questions or work assignments (so-called ‘prompts’) posed by the user on the basis of the underlying training data. These are statistical models that do not retrieve ‘knowledge’, but calculate probable word sequences based on an analysis of recognized text patterns and contexts and output them as answers In addition, large language models are able to contextualize a user’s input information and create new content based on their training data, which can then be restructured or linguistically adapted according to a user’s requirements, see Heetkamp and Piroutek (n 519), (2023) 26(3) ZKM (Zeitschrift für Konfliktmanagement) 80.
[522] S Meder, ‘Die Zukunft der juristischen Methode: Rehabilitierung durch Chat-GPT?’ (2023) 78(23) JZ (Juristenzeitung) 1041, 1051 at fn 109; Heetkamp and Piroutek (n 519), (2023) 26(3) ZKM (Zeitschrift für Konfliktmanagement) 80.
[523] T Deichsel, ‘Verbraucherschlichtungsstellen – Ein Anwendungsfeld für Legal Tech?’ (2020) 35(8) VuR (Verbraucher und Recht) 283, 287.
[524] Thus, the proposal for an automated conflict system by H M Anzinger, ‘10 Jahre Modria – KMS und Online-Mediation auf dem Weg zur Digitalisierung der Justiz – Teil 1’ (2021) 24(2) ZKM (Zeitschrift für Konfliktmanagement) 53, 56.
[525] Kreis (n 518) 633, 640 para 27 f.
[526] In this respect, the problem is likely to be similar to that in arbitration proceedings due to a lack of suitable databases and other information bases, see on automated selections of arbitrators below para 206-207.
[527] The Singapore Small Claims Tribunal is responsible for disputes between buyers and sellers in the range of 20.000 to 30.000 Singapore Dollars, see https://www.judiciary.gov.sg/civil/file-small-claim accessed 31 December 2023.
[528] H M Anzinger, ‘10 Jahre Modria – KMS und Online-Mediation auf dem Weg zur Digitalisierung der Justiz – Teil 2’ (2021) 24(3) ZKM (Zeitschrift für Konfliktmanagement) 84, 87.
[529] Deichsel (n 524), (2020) 35(8) VuR (Verbraucher und Recht) 283, 287.
[530] Kreis (n 518) 633, 640 para 28.
[531] Anzinger (n 525), (2021) 24(2) ZKM (Zeitschrift für Konfliktmanagement) 53, 56 f.
[532] Anzinger (n 525), (2021) 24(2) ZKM (Zeitschrift für Konfliktmanagement) 53, 57.
[533] Deichsel (n 524), (2020) 35(8) VuR (Verbraucher und Recht) 283, 288.
[534] Fundamental in this regard: Van Loo (n 183), (2016) 33(2) Yale Journal on Regulation 547, 566 ff.
[535] Van Loo (n 183), (2016) 33(2) Yale Journal on Regulation 547, 551 f.
[537] W Voß, ‘Gerichtsverbundene Online-Streitbeilegung’ (2020) 84(1) RabelsZ (Rabels Zeitschrift für ausländisches und internationales Privatrecht) 62, 65; also G Rühl, ‘Digitale Justiz’ (2020) 75(17) JZ (Juristenzeitung) 809, 811 f; A Sela, ‘The Effect of Online Technologies on Dispute Resolution System Design’ (2017) 21(3) Lewis & Clark Law Review 633, 662; H Barton, ‘Rebooting Justice’ (2018) 44(4) Law Practice 32, 35 f; C Rule, ‘Making Peace on eBay’ (2008) ACR Resolution 8, 10.
[538] A Sela (n 538), (2017) 21(3) Lewis & Clark Law Review 633, 662.
[539] For example, certified out-of-court dispute resolution bodies such as ‘Der Online-Schlichter’; also: complaint and ombudsman procedures such as the ‘Internet Ombudsman’ in Austria, see Braegelmann (n 521) 215, 218 para 930. An instructive overview can be found in R Greger (n 517) Part D.
[540] ODR providers include Modria (https://www.tylertech.com/products/online-dispute-resolution accessed 31 December 2023), SquareTrade, Cybersettle (https://www.cybersettle.com/ accessed 31 December 2023) and Smartsettle (https://www.smartsettle.com/ accessed 31 December 2023). The NCTDR (The National Center for Technology & Dispute Resolution) provides an overview of private providers at: https://odr.info/provider-list / accessed 31 December 2023.
[541] An instructive overview can be found in Anzinger (n 529), (2021) 24(3) ZKM (Zeitschrift für Konfliktmanagement) 84, 85; W Brazil, ‘Informalism and Formalism in the History of ADR in the United States’ in J Zekoll, M Bälz and I Amelung (ed), Formalisation and Flexibilisation in Dispute Resolution (Brill 2014) 250, 280 ff; D Hensler, ‘The Private in Public, the Public in Private’ in J Zekoll, M Bälz and I Amelung (ed), Formalisation and Flexibilisation in Dispute Resolution (Brill 2014) 45, 48 ff, 53-55.
[542] The dispute resolution of the platform ‘uitelkaar.nl’ (as successor to the provider ‘Rechtwijzer’) only takes place online, see Anzinger (n 529), (2021) 24(3) ZKM (Zeitschrift für Konfliktmanagement) 84, 87.
[543] See: https://civilresolutionbc.ca/solution-explorer/ accessed 31 December 2023. The Civil Resolution Tribunal is responsible for car accidents, small claims up to 5.000 Canadian dollars, special tenancy cases (strata property) and proceedings against companies based in British Columbia; see in detail V Tan, ‘Online Dispute Resolution for Small Civil Claims in Victoria’ (2019) 24 Deakin Law Review 101, 116-118; Anzinger (n 529), (2021) 24(3) ZKM (Zeitschrift für Konfliktmanagement) 84, 86.
[544] S Salter and D Thompson, ‘Public-Centred Civil Justice Redesign’ (2016-2017) 3 McGill Journal of Dispute Resolution 113, 129; Tan (n 544), (2019) 24 Deakin Law Review 101, 121.
[545] Deichsel (n 524), (2020) 35(8) VuR (Verbraucher und Recht) 283, 286 f, 288; S and H Kumar, ‘Mediation and Artificial Intelligence’ (2021) 4(4) International Journal of Law Management & Humanities 1472, 1477.
[546] S and H Kumar (n 546), (2021) 4(4) International Journal of Law Management & Humanities 1472, 1476 f; F Specht, ‘Chancen und Risiken einer digitalen Justiz für den Zivilprozess’ (2019) 22(3) MMR (Multimedia und Recht) 153, 156; Deichsel (n 524), (2020) 35(8) VuR (Verbraucher und Recht) 283, 286 f.
[547] On the so-called ‘digital justice gap’, see already: Braegelmann (n 521) 215 para 922; Anzinger (n 529), (2021) 24(3) ZKM (Zeitschrift für Konfliktmanagement) 84, 87 f.
[548] Critical of the assumption of an extended access: Voß (n 538), (2020) 84(1) RabelsZ (Rabels Zeitschrift für ausländisches und internationales Privatrecht) 62, 64-66. – On the assessment of online dispute resolution mechanisms as a catalyst for effective access to justice, see C Menkel-Meadow, ‘Is ODR ADR? Reflections of an ADR Founder from 15th ODR Conference, The Hague’ (2016) 3(1) IJODR (International Journal on Online Dispute Resolution) 4; O Rabinovich-Einy and E Katsh, ‘The New New Courts’ (2017) 67(1) American University Law Review 165, 169. – On the historical roots of the ADR concept of overcoming structural weaknesses of state court systems and milieu-specific access barriers, see G Wagner, ‘Private Law Enforcement and ADR’ in J Zekoll, M Bälz and I Amelung (ed), Formalisation and Flexibilisation in Dispute Resolution (Brill 2014) 369 f; M Wendland, Mediation und Zivilprozess (2017), 199 ff.
[549] See Deichsel (n 524), (2020) 35(8) VuR (Verbraucher und Recht) 283, 286 f.
[550] Meder (n 523), (2023) 78(23) JZ (Juristenzeitung) 1041, 1047.
[551] For arbitration proceedings: C Sim, ‘Will Artificial Intelligence Take Over Arbitration?’ (2018) 14(1) Asian International Arbitration Journal 1, 8 f; G Vannieuwenhuyse, ‘Arbitration and New Technologies: Mutual Benefits’ (2018) 35(1) Journal of International Arbitration 119, 124.
[552] S and H Kumar (n 546), (2021) 4(4) International Journal of Law Management & Humanities 1472, 1478 f. On the judicial process: G Rühl, ‘KI in der gerichtlichen Streitbeilegung’ in M Kaulartz and T Braegelmann (ed), Rechtshandbuch Artificial Intelligence und Machine Learning (2020) 617, 627 para 20.
[553] In addition, Heetkamp and Piroutek (n 519), (2023) 26(3) ZKM (Zeitschrift für Konfliktmanagement) 80.
[554] Anzinger (n 525), (2021) 24(2) ZKM (Zeitschrift für Konfliktmanagement) 53, 57: ‘Models can be found in systems such as A2JAuthor [https://www.a2jauthor.org/] or Law Lift [https://de.lawlift.com/>] and Smart Law [https://www.smartlaw.de/].’ – On consumer arbitration: Deichsel (n 524), (2020) 35(8) VuR (Verbraucher und Recht) 283, 288.
[555] See below para 212-214.
[556] See commentary to Guideline 1 of the Silicon Valley Arbitration & Mediation Center Guidelines, Draft of August 31, 2023 https://thearbitration.org/wp-content/uploads/2023/08/SVAMC-AI-Guidelines-CONSULTATION-DRAFT-31-August-2023-1.pdf accessed 31 December 2023.
[557] Rühl (n 553) 617, 624 f para 15 f; M Scherer, ‘Artificial Intelligence and Legal Decision-Making’ (2019) 36(5) Journal of International Arbitration 539, 557.
[558] In other words, simultaneous updating of training data is impossible, cf Meder (n 523), (2023) 78(23) JZ (Juristenzeitung) 1041, 1042, 1050.
[559] Meder (n 523), (2023) 78(23) JZ (Juristenzeitung) 1041, 1047.
[560] For the arbitration proceedings: D Lindquist and Y Dautaj, ‘AI in International Arbitration’ (2021) (1) Journal of Dispute Resolution 39, 54 f; M Fries, ‘Legal Tech im Schiedsverfahren’ in R Wilhelmi and M Stürner (ed), Mehrparteienschiedsverfahren (Springer 2021) 85, 93; see also D Nink, Justice and Algorithms (Duncker & Humboldt 2021), 230.
[561] This problem arises both in continental European law and in case law, see Scherer (n 558), (2019) 36(5) Journal of International Arbitration 539, 557.
[562] Scherer (n 558), (2019) 36(5) Journal of International Arbitration 539, 557. An overview of the various models of artificial intelligence in the context of legal decision-making processes can be found in: ibid, 546 ff.
[563] For the parallel issue of arbitration, see below para 215 and para 225.
[564] S and H Kumar (n 546), (2021) 4(4) International Journal of Law Management & Humanities 1472, 1478.
[565] Heetkamp and Piroutek (n 519), (2023) 26(3) ZKM (Zeitschrift für Konfliktmanagement) 80, 81.
[566] B Gsell, ‘Die Umsetzung der Richtlinie über alternative Streitbeilegung Juristisches Fachwissen der streitbeilegenden Personen und Rechtstreue des Verfahrensergebnisses’ (2015) 128(2) ZZP (Zeitschrift für Zivilprozess) 189, 199 f; M Fries, Verbraucherrechtsdurchsetzung (Mohr Siebeck 2016), 245.
[567] However, cf § 19 of the German Consumer Dispute Resolution Act (Verbraucherstreitbeilegungsgesetzes: VSBG): ‘The conciliation proposal shall be based on the applicable law and shall in particular comply with the mandatory consumer protection laws’. See §§ 16, 17 of the German
[568] W Voß, for example, attests to a considerable simplification of state law on service disruption and consumer protection, in Voß (n 538), (2020) 84(1) RabelsZ (Rabels Zeitschrift für ausländisches und internationales Privatrecht) 62, 66 f; see also Specht (n 547), (2019) 22(3) MMR (Multimedia und Recht) 153, 155; Anzinger (n 525), (2021) 24(2) ZKM (Zeitschrift für Konfliktmanagement) 53, 56.
[569] If a conflict arises between merchant and customer, the employees of the PayPal group essentially ‘decide’ on the basis that money and goods must not be with the same person, see M Fries, ‘PayPal Law und Legal Tech – Was macht die Digitalisierung mit dem Privatrecht?’ (2016) 69(39) NJW (Neue Juristische Wochenschrift) 2860, 2861 f; also: J Adolphsen, ‘Der Zivilprozess im Wettbewerb der Methoden’ (2017) 48(4) BRAK Mitteilungen 147, 149; Rühl (n 538), (2020) 75(17) JZ (Juristenzeitung) 809, 812. See, however, the decision of the German Federal Court of Justice (Bundesgerichtshof) on the PayPal Buyer Protection Directive: BGH (Germany) 22 November 2017, VIII ZR 83/16, (2018) 71(8) NJW (Neue Juristische Wochenschrift) 537 and VIII ZR 213/16, (2018) 21(3) MMR (Multimedia und Recht) 156.
[570] See Wendland (n 549) 192-213.
[571] J Adolphsen even speaks of a ‘primitive legal system’: Adolphsen (n 570), (2017) 48(4) BRAK Mitteilungen 147, 149; of a ‘banalization of private law’: C Althammer, ‘Alternative Streitbeilegung im Internet’ in F Faust and H-B Schäfer (ed), Zivilrechtliche und rechtsökonomische Probleme des Internet und der künstlichen Intelligenz (Mohr Siebeck 2019), 249, 266.
[572] M Fries chooses the term ‘de facto privatization of civil law’, in Fries (n 570), (2016) 69(39) NJW (Neue Juristische Wochenschrift) 2860, 2860 f.
[573] To this end: Althammer (n 572) 249, 260.
[574] Thus, according to M Wendland, the result of mediation (as conventionally understood) is a singular product of the individual case and not a ‘party contract law’ or an autonomous private order, see Wendland (n 549) 196.
[575] See H Prütting, ‘Das neue Verbraucherstreitbeilegungsgesetz: Was sich ändert – und was bleiben wird’ (2016) (3) AnwBl (Anwaltsblatt) 190, 192 f; Althammer (n 572) 249, 260 f. – The phenomenon of the ‘conservatism’ of machine-based decision-making cannot be addressed by publishing paradigmatic conflict cases. On the benefits of corresponding publications outside of AI-based decision-making, see Hess (n 330), (2015) 70(11) JZ (Juristenzeitung) 548, 553.
[576] U Gläßer, ‘Mediation und Digitalisierung’ in T Riehm and S Dörr (ed), Digitalisierung und Zivilverfahren (De Gruyter 2023) 529, 531. On the power of platforms to exclude companies as well: Althammer (n 572), 249, 262; M Fries, ‘Erfüllung von Geldschulden über eigenwillige Zahlungsdienstleister’ (2018) 33(4) VuR (Verbraucher und Recht) 123, 124.
[577] Heetkamp and Piroutek (n 519), (2023) 26(3) ZKM (Zeitschrift für Konfliktmanagement) 80, 81. For example, the Digital Services Act requires disclosure of (AI-based) decision parameters and statistical bases of dispute resolution, cf. on the one hand Art 17(3) lit d) DSA Regulation (reference to the legal basis and explanations as to why the information is considered unlawful content on this basis); according to lit c), the hosting service provider’s decision to impose usage restrictions must also indicate whether automated means were used to make the decision, including whether the decision was made in relation to content that was identified or detected by automated means. – Secondly, Art 24(1) DSA Regulation formulates periodic reporting obligations for online platform providers, including on the number, outcome and duration of disputes and dispute resolution.
[578] See Regulation (EU) 2024/1689 of 13 June 2024 laying down harmonised rules on artificial intelligence and amending Regulations (EC) No 300/2008, (EU) No 167/2013, (EU) No 168/2013, (EU) 2018/858, (EU) 2018/1139 and (EU) 2019/2144 and Directives 2014/90/EU, (EU) 2016/797 and (EU) 2020/1828 (Artificial Intelligence Act), OJ L, 12 July 2024. According to this, the AI Act is to apply alongside the Digital Services Act (DSA), see Art 2(5) AI Act.
[579] Art 50 AI Act.
[580] Also covered – although less relevant for the types of procedure examined here – are systems for biometric categorization, Art 50(3) AI Act and AI systems in connection with deep fakes, Art 50(2) AI Act.
[581] P Richter and J Mendelsohn, ‘§ 21 Plattformspezifische Vorgaben des Data Acts’ in B Steinrötter (ed), Europäische Plattformreguliereng (Nomos 2023) para 17 ff (on the Proposal AI Act).
[582] Richter and Mendelsohn (n 582) para 31.
[583] Art 16 AI Act.
[584] Art 52(1), subpara 2 of the proposal amended by the EU Parliament: ‘Where appropriate and relevant, this information shall also include which functions are AI enabled, if there is human oversight, and who is responsible for the decision-making process, as well as the existing rights and processes that, according to Union and national law, allow natural persons or their representatives to object against the application of such systems to them and to seek judicial redress against decisions taken by or harm caused by AI systems, including their right to seek an explanation’. See Synopsis AI Act, Commission-Parliament.pdf (P9_TA(2023)0236), https://www.europarl.europa.eu/doceo/document/TA-9-2023-0236_EN.pdf accessed 31 December 2023.
[585] Art 21 DSA Regulation.
[586] See §§ 16, 17 of the German Copyright Service Provider Act (Urheberrechts-Diensteanbieter-Gesetz: UrhDaG) for out-of-court dispute resolution by private and official conciliation boards.
[587] Heetkamp and Piroutek (n 519), (2023) 26(3) ZKM (Zeitschrift für Konfliktmanagement) 80, 81, who consider forms of hybrid decision-making to be particularly suitable in the area of e-commerce. The reason for this is the high rate of automated dispute resolution that can be found there anyway.
[588] See above para 108-175.
[589] Art 6(2), Annex III no 8 lit a) AI Act. See in more detail above para 162 ff.
[590] See in more detail above para 161 ff.
[591] In 2021, for example, 15% of respondents to the International Arbitration Survey stated that they regularly or frequently use artificial intelligence; Queen Mary University and White & Case, ‘International Arbitration Survey: Adapting Arbitration to a Changing World’ (2021) 21, https://www.qmul.ac.uk/arbitration/media/arbitration/docs/LON0320037-QMUL-International-Arbitration-Survey-2021_19_WEB.pdf accessed 31 December 2023.
[592] J Rajendra and A Thuraisingam, ‘The deployment of artificial intelligence in alternative dispute resolution, the AI augmented arbitrator’ (2022) 31(2) Information & Communications Technology Law 176-193.
[593] Queen Mary University and White & Case (n 592) 22.
[594] M Scherer and O Jensen, ‘Die Digitalisierung der Schiedsgerichtsbarkeit’ in T Riehm and S Dörr (ed), Digitalisierung und Zivilverfahren (De Gruyter 2023) 591, 604 para 34. On such software solutions, see Rühl (n 553) 617, 618.
[595] E Zorrilla, ‘Towards a Credible Future’ (2018) 16(2) SchiedsVZ (Zeitschrift für Schiedsverfahren) 106, 113; G Zekos, Advanced Artificial Intelligence and Robo-Justice (Springer 2022), 328. One provider of such software is DISCO https://csdisco.com/offerings/review.
[596] Kreis (n 518) 633, 639 para 22.
[597] Zorrilla (n 596), (2018) 16(2) SchiedsVZ (Zeitschrift für Schiedsverfahren) 106, 113; Zekos (n 596) 328; Eidenmüller and Wagner (n 195) 192.
[598] L Bizikova, P Hancock, D Jewell and I Sherr, ‘IA Meets AI’ (2 October 2023) https://dailyjus.com/legal-tech/2023/10/ia-meets-ai-rise-of-the-machines accessed 31 December 2023; Eidenmüller and Wagner (n 195) 192 f; Zekos (n 596) 326. Existing software is, for instance, the e-discovery tool eBravia (https://www.dfinsolutions.com/products/ebrevia).
[599] Zorrilla (n 596), (2018) 16(2) SchiedsVZ (Zeitschrift für Schiedsverfahren) 106, 113.
[600] Existing tools include Trint (https://trint.com/), Fireflies (https://firefliesai/) and Otter (https://otter.ai).
[601] Kreis (n 518) 633, 639 para 22; Zekos (n 596) 326; Bizikova, Hancock, Jewell and Sherr (n 599), ‘IA Meets AI’ (2 October 2023).
[602] The research field of Natural Language Processing deals with the algorithmic processing of spoken and written language. This can be implemented using artificial intelligence, among other things, see F Deusch and T Eggendorfer, ‘IT-Sicherheit’ in J Taeger and J Pohle (ed), Computerrechts-Handbuch (38th edn, Beck 8/2023) para 232n.
[603] This technology is used to recognize and convert text – such as a scan or photo of a typed or handwritten text – into a machine-readable format, see P von Bünnau, ‘Künstliche Intelligenz im Recht’ in S Breidenbach and F Glatz (ed), Rechtshandbuch Legal Tech (2nd edn, Beck/Manz 2021) 71 para 18; regarding arbitration: Eidenmüller and Wagner (n 195) 192.
[604] Predictive coding – also known as technology-assisted review – involves (technically simplified) complex algorithms that are able to search and analyze large volumes of documents. This technology is primarily used in e-discovery, see in detail C Yablon and N Landsman-Ross, ‘Predictive Coding’ (2013) 64(3) South Carolina Law Review 633, 643, 638; regarding arbitration: Zorrilla (n 596), (2018) 16(2) SchiedsVZ (Zeitschrift für Schiedsverfahren) 106, 111; corresponding technology is used, for example, by the service provider DISCO (https://www.csdisco.com/offerings/ediscovery/features-ai).
[605] Scherer and Jensen (n 595) 591, 604 para 34; Rühl (n 553) 617, 620 para 6.
[606] Examples include Westlaw Edge https://legal.thomsonreuters.com/en/products/westlaw-edge and Lexis+ https://www.lexisnexiscom/en-us/products/lexis-pluspage.
[607] S Barona Vilar, ‘Effizienzsteigerung und Suche nach Beschleunigung von Schiedsverfahren im Spannungsfeld von Mythos, Sublimierung und Vierter Industrieller Revolution (4.0)’ (2019) 23 ZZPInt (Zeitschrift für Zivilprozess International) 295, 313; Bizikova, Hancock, Jewell and Sherr (n 599), ‘IA Meets AI’ (2 October 2023).
[608] Eg, the ‘Jus-AI’ of the provider JusMundi https://jusmundi.com/en; Daily Jus, ‘Jus Mundi Introduces Jus-AI’ (29 June 2023) https://dailyjus.com/news/2023/06/jus-mundi-introduces-jus-ai-a-game-changing-gpt-powered-ai-solution-for-the-arbitration-community accessed 31 December 2023; overview in Bizikova, Hancock, Jewell and Sherr (n 599), ‘IA Meets AI’ (2 October 2023).
[609] Zekos (n 596) 330. – On the pre-drafting of court decisions, see Fries (n 570), (2016) 69(39) NJW (Neue Juristische Wochenschrift) 2860, 2864.
[610] Zorrilla (n 596), (2018) 16(2) SchiedsVZ (Zeitschrift für Schiedsverfahren) 106, 113. Specific providers for the review of documents in arbitration proceedings cannot be found. Technically comparable software for contract drafting is, for example, Luminance (https://www.luminance.com/overview.html).
[611] Kreis (n 518) 633, 639 para 23.
[612] Rajendra and Thuraisingam (n 593), (2022) 31(2) Information & Communications Technology Law 176, 183 f; Zekos (n 596) 325 f.
[613] Zorrilla (n 596), (2018) 16(2) SchiedsVZ (Zeitschrift für Schiedsverfahren) 106, 116; Kreis (n 518) 633, 639 para 23.
[614] Eidenmüller and Wagner (n 195) 191 f.
[615] C Aschauer, ‘Automated Decision-Making and Artificial Intelligence (AI) in Arbitration’ in C Leyens, I Eisenberger and R Niemann (ed), Smart Regulation (Mohr Siebeck 2021) 130, 133.
[616] See H Prütting, ‘Die rechtliche Stellung des Schiedsrichters’ (2011) 9(5) SchiedsVZ (Zeitschrift für Schiedsverfahren) 233, 235.
[617] Specifically on the use of AI: Kreis (n 518) 633, 644 para 46-50 in relation to ‘producing’, not merely ‘examining tasks’ of the arbitrator. – On the supreme personal nature of the mandate: Prütting (n 617), (2011) 9(5) SchiedsVZ (Zeitschrift für Schiedsverfahren) 233, 235 – On the admissibility of delegation to auxiliary persons: O Jensen, Tribunal Secretaries in International Arbitration (Oxford University Press 2019) para 805 ff.
[618] Zorrilla (n 596), (2018) 16(2) SchiedsVZ (Zeitschrift für Schiedsverfahren) 106, 113; Kreis (n 518) 633, 644 para 46-50. – On the use of tribunal secretaries, see Sim (n 552), (2018) 14(1) Asian International Arbitration Journal 1, 6; Jensen (n 618) para 805 ff; M Polkinghorne, ‘Different Strokes for Different Folks?’ Kluwer Arbitration Blog (16 May 2014), https://arbitrationblog.kluwerarbitration.com/2014/05/17/different-strokes-for-different-folks-the-role-of-the-tribunal-secretary-2/ accessed 31 December 2023.
[619] Kreis (n 518) 633, 644 para 50.
[620] One example is Kluwer Arbitration https://www.kluwerarbitration.com/ accessed 31 December 2023.
[621] The subject of research is, for example, data that arbitrators voluntarily or involuntarily leave behind in social media, for example on family circumstances, political inclinations or general sensitivities, see in more detail Aschauer (n 616) 130, 135 f.
[622] Y Rhim and K Park, ‘The Artificial Intelligence in International Law’ in E Y J Lee (ed), Revolutionary Approach to international Law: The Role of international Lawyer in Asia (Springer 2023) 215, 224 f.
[623] The lack of publication of arbitration awards makes it difficult to draw conclusions about the working methods and ‘habitus’ of the arbitrators.
[624] The Arbitrator Intelligence project, for example, has created an evaluation database for arbitrators (based on questionnaires and crowd-sourced arbitration awards). However, the initiator of the project, Professor Catherine Rogers, herself pointed out that the project does not yet have a sufficient data basis for processing by means of machine learning; see in detail Zorrilla (n 596), (2018) 16(2) SchiedsVZ (Zeitschrift für Schiedsverfahren) 106, 111. – Other databases are: Global Arbitration Review Arbitrator GAR ART, https://globalarbitrationreview.com/tools/arbitrator-research-tool accessed 31 December 2023 or Jus mundi, https://jusmundi.com/en accessed 31 December 2023, see Aschauer (n 616) 130.
[625] Scherer and Jensen (n 595) 591, 615 para 61; Rühl (n 553) 617, 619, para 6-11; Zekos (n 596) 329 f; L Bull and F Steffek, ‘The Decoding of Legal Conflicts’ (2018) 21(5) ZKM (Zeitschrift für Konfliktmanagement) 165, 166.
[626] Bizikova, Hancock, Jewell and Sherr (n 599), ‘IA Meets AI’ (2 October 2023); Aschauer (n 616) 130, 135; Barona Vilar (n 608), (2019) 23 ZZPInt (Zeitschrift für Zivilprozess International) 295, 314.
[627] Scherer and Jensen (n 595) 591, 615 para 61.
[628] On the increasing importance of litigation funding in arbitration proceedings, see only S Wilske, L Markert and B Ebert, ‘Entwicklungen in der internationalen Schiedsgerichtsbarkeit im Jahr 2022 und Ausblick auf 2023’ (2023) 21(3) SchiedsVZ (Zeitschrift für Schiedsverfahren) 121, 125.
[629] Scherer and Jensen (n 595) 591, 615 para 61.
[630] Predictive tools that work on the basis of metadata analysis include: LexMachina of patent disputes https://lexmachina.com/ and Predictice https://predictice.com, all accessed 31 December 2023. Another well-known example is the study on the prediction of decisions of the US Supreme Court, see T Ruger, P Kim, A Martin and K Quinn, ‘The Supreme Court Forecasting Project’ (2004) 104(4) Colum. L. Rev. 1150, 1163 ff.
[631] The factual data analysis is based on a comparison of the facts of the case with the facts of relevant preliminary decisions, see Rühl (n 553) 617, 621 f para 9 f.
[632] Rühl (n 553) 617, 620 para 6. On predictive tools that work on the basis of factual data analysis, see also the study on the prediction of decisions of the ECtHR: M Medvedeva, M Vols and M Wieling, ‘Using machine learning to predict decisions of the European Court of Human Rights’ (2020) 28(2) Artificial Intelligence and Law 237, 266; see also the study on predicting decisions of the Financial Ombudsman in the UK (Case Cruncher Alpha): R Cellan-Jones, ‘The robot lawyers are here’ (1 November 2017) BBC News, https://www.bbc.com/news/technology-41829534 accessed 31 December 2023.
[633] The random forest method is used in the context of machine learning. This is a combination of decision trees (random forest): Each tree depends on the values of a random vector that is determined independently and with the same distribution for all decision trees in the forest, see L Breiman, ‘Random Forests’ (2001) 45(1) Machine Learning 5 ff.
[634] Deichsel (n 520), 98 f.
[635] This applies both to the use by parties and litigation funders as well as by the arbitrator himself, see S Marmont, ‘Keeping Up with Legal Technology’ (2019) 1(2) ITA in Review 37, 41, 48.
[636] Aschauer (n 616) 130, 137.
[637] See J Lew, L Mistelis and S Kröll, Comparative International Commercial Arbitration (Kluwer Law International 2003) 265.
[638] On the existence and scope of this obligation to investigate and the consequences of its violation, see Lew, Mistelis and Kröll (n 638) 269.
[639] The scope of this obligation to investigate is limited by the principle of reasonableness. For the whole, see S Marmont (n 636), (2019) 1(2) ITA in Review 37, 41, 47 f.
[640] Art 33 Law of March 23, 2019 (Loi de programmation 2018-2022 et réforme pour la justice) no 2019/222, available at https://www.legifrance.gouv.fr/ accessed 31 December 2023. In that regard, the French legislator has generally prohibited the use of data on the identity of judges for the evaluation, analysis, comparison or prediction of their decisions.
[641] Aschauer (n 616) 130, 137.
[642] This is also the case with Kreis (n 518) 633, 647 para 62.
[643] J Schwartz, ‘Artificial Arbitration?’ in R Wilhelmi and M Stürner (ed), Mehrparteien-Schiedsverfahren: Unter besonderer Berücksichtigung gesellschaftsrechtlicher Streitigkeiten (Springer 2021) 95, 120 f.
[644] Bizikova, Hancock, Jewell and Sherr (n 599), ‘IA Meets AI’ (2 October 2023); different, albeit without further justification: Eidenmüller and Wagner (n 195), 203.
[645] M Kaulartz, ‘Smart Contract Dispute Resolution’ in M Fries and B-P Paal (ed), Smart Contracts (Mohr Siebeck 2019) 73, 80 f.
[646] Interesting here is ChatGPT’s answer to the question ‘Can you act as an arbitrator in an arbitration?’: ‘No, I cannot act as an arbitrator in an arbitration. The role of an arbitrator requires a specific set of skills, experience and qualifications that include human characteristics and judgment. As an AI model, I lack these qualifications and the ability to make human decisions’ (as of 8 December 2023).
[647] See above para 189-190. On arbitration proceedings: Lindquist and Dautaj (n 561), (2021) (1) Journal of Dispute Resolution 39, 48 f, 51 ff; Vannieuwenhuyse (n 552), (2018) 35(1) Journal of International Arbitration 119, 124; G Halis Kasap, ‘Can Artificial Intelligence (“AI”) Replace Human Arbitrators?’ (2021) (2) Journal of Dispute Resolution 209, 232 ff; Nink (n 561), 231 ff; Schwartz (n 644), 95, 122 f; Bizikova, Hancock, Jewell and Sherr (n 599), ‘IA Meets AI’ (2 October 2023); Aschauer (n 616) 130, 134; Kreis (n 518) 633, 648 para 67 ff.
[648] Schwartz (n 644) 95, 122 f; Aschauer (n 616) 130, 134; Lindquist and Dautaj (n 561), (2021) (1) Journal of Dispute Resolution 39, 48 f, 51 ff.
[649] Lindquist and Dautaj (n 561), (2021) (1) Journal of Dispute Resolution 39, 49.
[650] Rhim and Park (n 623) 215, 225; Zekos (n 596) 329; Schwartz (n 644) 95, 123; Halis Kasap (n 648), (2021) (2) Journal of Dispute Resolution 209, 221 ff.
[651] J Münch in Münchener Kommentar zur ZPO (6th edn, Beck 2022) vor § 1025, para 5, 127; Rhim and Park (n 623) 215, 225; K Paisley and E Sussman, ‘Artificial Intelligence Challenges and Opportunities for International Arbitration’ (2018) 11(1) New York Dispute Resolution Lawyer 35, 37.
[652] Cf for example: Arbitrator Intelligence https://arbitratorintelligence.vercel.app/ accessed 31 December 2023, see P Shaughnessy and C Rogers, ‘Arbitrator Intelligence – An Interview with its Founder and Director, Professor Catherine Rogers’ (2015) 87 Journal on Technology in International Arbitration 87, 96; Dispute Resolution Data https://www.disputeresolutiondata.com/ accessed 31 December 2023; Global Arbitration Review Arbitrator Research Tool (GAR ART, https://globalarbitrationreview.com/tools/arbitrator-research-tool accessed 31 December 2023). On the whole: Paisley and Sussman (n 652), (2018) 11(1) New York Dispute Resolution Lawyer 35, 38; Rhim and Park (n 623) 215, 225 f.
[653] Eidenmüller and Wagner (n 195) 202 f.
[654] Rhim and Park (n 623) 215, 226. – Regarding data protection concerns, cf Paisley and Sussman (n 652), (2018) 11(1) New York Dispute Resolution Lawyer 35, 38.
[655] Kreis (n 518) 633, 646-648 para 61 ff, 76; Kaulartz (n 646) 73, 80; Halis Kasap (n 648), (2021) 2 Journal of Dispute Resolution 209, 237 ff.
[656] Cf Art. 11, 12(1) UNCITRAL Model Law. Agreeing: Rhim and Park (n 623) 215, 225; Eidenmüller and Wagner (n 195) 215.
[657] Zekos (n 596) 340; H Snijders, Arbitration and AI, Arbitration (1st edn, Wolters Kluwer 2023) 224, 234 ff.
[658] Eidenmüller and Wagner (n 195) 209 f; Rhim and Park (n 623) 215, 225; Zekos (n 596) 381 f; Halis Kasap (n 648), (2021) (2) Journal of Dispute Resolution 209, 237.
[659] Scherer and Jensen (n 595) 591, 617 f.
[660] Art 1450(1) Code de procédure civile: ‘La mission d’arbitre ne peut être exercée que par une personne physique jouissant du plein exercice de ses droits’. See M Scherer, ‘International Arbitration 3.0. How Artificial Intelligence Will Change Dispute Resolution’ in C Klausegger et al (ed), Austrian Yearbook on International Arbitration (1st edn, Beck 2019) 503, 512 fn 22; Bizikova, Hancock, Jewell and Sherr (n 599), ‘IA Meets AI’ (2 October 2023).
[661] Art 11.1 s 1 NAI-SchO (Schiedsordnung Nederlands Arbitrage Instituut: Arbitration Rules of the Netherlands Arbitration Institute): ‘Any natural person [‘natuurlijke persoon’] of legal capacity may be appointed as arbitrator’.
[662] Art 13 Spanish Arbitration Act: ‘All natural persons in full possession of their civil rights may act as arbitrators, provided that they are not restricted by the legislation applicable to them in the exercise of their profession’.
[663] Turkish International Arbitration Law, Article 7(B)(l): ‘Only natural persons can be selected as arbitrators’.
[664] G Maxwell and G Vannieuwenhuyse, ‘Robots Replacing Arbitrators: Smart Contract Arbitration’ (2018) (1) ICC Dispute Resolution Bulletin 24, 31.
[665] J Münch in Münchener Kommentar zur ZPO (6th edn, Beck 2022) vor § 1034 para 18-21; vor § 1025 para 5; § 1025 para 10. – The opposing view, which allows the parties to appoint AI as arbitrator, also places party autonomy under the proviso that basic procedural guarantees are preserved. It is based on state protection obligations and barriers that claim to be valid both in the interests of the parties and the general public, see Kaulartz (n 646) 73, 80 f; J Münch in Münchener Kommentar zur ZPO (6th edn, Beck 2022) vor § 1025 para 6.
[666] Art 13(1) ICC Arbitration Rules 2021, referring to the nationality of the arbitrator; furthermore, Art. 16(1) of the Vienna Rules 2018, which refers to legal capacity, see Aschauer (n 616) 130, 133.
[667] With regard to a possible legal capacity of the arbitrator, it is proposed, in line with the discussion held in the EU Parliament in 2017, to provide automated systems with legal capacity (‘e-personality’) or to allow the fully automated management of legal entities (‘self-driving corporation’), so: Eidenmüller and Wagner (n 195) 201 f, 157 ff.
[668] Section 7 LoS: ‘Var och en som råder över sig själv och sin egendom’ [Anyone who has legal capacity]), cf J Münch in Münchener Kommentar zur ZPO (6th edn, München 2022) vor § 1025 para 5 fn 13.
[669] Art 812(1) Codice di procedura civile: ‘La norma in analisi indica il requisito fondamentale di capacità degli arbitri, ovvero il pieno possesso della piena capacità legale di agire’.
[670] Section 26(1) of the English Arbitration Act 1996: ‘The authority of the arbitrator is personal and ceases on his death’.
[671] Kreis (n 518) 633, 694 para 72.
[672] Scherer and Jensen (n 595) 591, 618.
[673] § 1054(1) ZPO (German Code of Civil Procedure).
[674] Such as a clarification, evidence, authentication and concluding function. – Affirmative: Kreis (n 518) 633, 648 f para 70 f.
[675] Maxwell and Vannieuwenhuyse (n 665), (2018) (1) ICC Dispute Resolution Bulletin 24, 30; generally on data protection in arbitration, see A Cervenka and P Schwarz, ‘Data Protection in Arbitration Proceedings’ (2020) 18(2) SchiedsVZ (Zeitschrift für Schiedsverfahren) 78, 79 f; G Fritz, D Prantl, N Leinwather and M Hofer, ‘Data Protection in International Arbitration Proceedings’ (2019) 17(6) SchiedsVZ (Zeitschrift für Schiedsverfahren) 301, 302 f; C Boll-Kempelmann, ‘Data protection and the evidence procedure in arbitration proceedings’ (2022) 20(5) SchiedsVZ (Zeitschrift für Schiedsverfahren) 241.
[676] Art 22(2) lit c) GDPR. On the judicial procedure, see Nink (n 561) 251 ff.
[677] Explicitly Zekos (n 596) 331 ff.
[678] In this context, enforcement is self-executing, meaning that legal conformity with the New York Convention and other arbitration law is irrelevant for the enforcement of the arbitral award. Generally with regard to blockchain arbitration: T Kindt, ‘Blockchainbasierte dezentrale Streitbeilegungsverfahren und ihr Verhältnis zur Schiedsgerichtsbarkeit’ (2023) 21(5) SchiedsVZ (Zeitschrift für Schiedsverfahren) 241 ff.; G Wagner, Legal Tech und Legal Robots (2nd edn, Springer 2020) 34 f.
[679] One provider of such procedures is the company Kleros, offering peer-to-peer arbitration proceedings for small claims, as well as in e-commerce, IP law and insurance law https://kleros.io/en/ accessed 31 December 2023; another one is Aragon (‘Aragon Court’).
[680] On this and in the following: Kindt (n 679), (2023) 21(5) SchiedsVZ (Zeitschrift für Schiedsverfahren) 241, 245 f.
[681] First of all, judicial independence and impartiality are questionable, as there is in fact no possibility of reviewing the anonymous jurors. Furthermore, their remuneration is directly linked to the outcome of the proceedings. Also, the blockchain-based procedure probably does not meet the right to a fair hearing due to lacking opportunities for the parties to express themselves after the proceedings have been initiated. Finally, doubts about the legal form of the decision arise in view of a lack of uniform decision-making standards. To all these aspects, see Kindt (n 679), (2023) 21(5) SchiedsVZ (Zeitschrift für Schiedsverfahren) 241, 246, 248-251.
[682] Such blockchain procedures may complement arbitration proceedings by either preceding them (eg within the framework of an escalation clause) or being integrated into them, see Kindt (n 679), (2023) 21(5) SchiedsVZ (Zeitschrift für Schiedsverfahren) 241, 252 f.
[683] Generally: D Kahneman, Thinking, Fast and Slow (1st edn, Penguin 2013) 119; Scherer (n 558), (2019) 36(5) Journal of International Arbitration 539, 557-562. In the context of arbitration: Zorrilla (n 596), (2018) 16(2) SchiedsVZ (Zeitschrift für Schiedsverfahren) 106, 113; on this point.
[684] Rhim and Park (n 623) 215, 225. – On the technical foundations of the discrimination problem, see Ebers (n 192) 75 para 101-138; Scherer (n 661) 503, 510; Sim (n 552), (2018) 14(1) Asian International Arbitration Journal 1, 7 ff; Halis Kasap (n 648), (2021) (2) Journal of Dispute Resolution 209, 225 ff.
[685] Scherer (n 558), (2019) 36(5) Journal of International Arbitration 539, 561.
[686] The empirical situation is not clear. On the whole: Scherer (n 558), (2019) 36(1) Journal of International Arbitration 539, 559-561.
[687] On the neutrality requirement: Kreis (n 518) 633, 645 f para 54-57.
[688] Zorrilla (n 596), (2018) 16(2) SchiedsVZ (Zeitschrift für Schiedsverfahren) 106, 112 ff; Scherer (n 661) 503, 511; Halis Kasap (n 648), (2021) (2) Journal of Dispute Resolution 209, 229 ff.
[689] Sim (n 552), (2018) 14(1) Asian International Arbitration Journal 1, 8 f; Vannieuwenhuyse (n 552), (2018) 35(1) Journal of International Arbitration 119, 124.
[690] On judicial proceedings: Rühl (n 553) 617, 627 para 20; Scherer (n 661) 503, 511; Scherer (n 558), (2019) 36(5) Journal of International Arbitration 539, 562.
[691] It is well known that trust in the expertise, reputation and personality of an arbitrator is particularly relevant. On the whole: Scherer (n 558), (2019) 36(5) Journal of International Arbitration 539, 565; Maxwell and Vannieuwenhuyse (n 665), (2018) (1) ICC Dispute Resolution Bulletin 24, 32; Halis Kasap (n 648), (2021) (2) Journal of Dispute Resolution 209, 230.
[692] Scherer (n 661) 503, 512; Scherer (n 558), (2019) 36(5) Journal of International Arbitration 539, 562.
[693] Sim (n 552), (2018) 14(1) Asian International Arbitration Journal 1, 8 f.
[694] Scherer (n 661) 503, 512; Scherer (n 558), (2019) 36(5) Journal of International Arbitration 539, 562.
[695] See above para 189-190.
[696] In some cases, disputes relating to entire areas of law are mainly settled in arbitration proceedings.
[697] Kreis (n 518) 633, 645 para 51 f.
[698] Bizikova, Hancock, Jewell and Sherr (n 599), ‘IA Meets AI’ (2 October 2023).
[699] Silicon Valley Arbitration & Mediation Center, SVAMC Guidelines on the Use of Artificial Intelligence in Arbitration, Draft of 31 August 2023 https://thearbitration.org/wp-content/uploads/2023/08/SVAMC-AI-Guidelines-CONSULTATION-DRAFT-31-August-2023-1.pdf accessed 31 December 2023.
[700] The purpose of the guidelines (reference framework) is described as follows (SVAMC Guidelines, 3): ‘The Guidelines seek to establish a set of general principles for the use of AI in arbitration. Intended to guide rather than dictate, they are meant to accommodate case-specific circumstances and technological developments, promoting fairness, efficiency, and transparency in arbitral proceedings’ – The term AI is defined as follows, SVAMC Guidelines, 3: ‘[...] the term ‘AI’ refers to computer systems that perform tasks commonly associated with human cognition, such as understanding natural language, recognizing complex semantic patterns, and generating human-like outputs’.
[701] See also SVAMC Guidelines (n 700) 17 (Commentary to Guideline 6).
[702] See also SVAMC Guidelines (n 700) Guideline 7. – The degree of depth of review must, of course, be weighed in each individual case against the cost and time savings hoped for (and achieved) through the use of AI.
[703] Schwartz (n 644) 95, 124 f.
[704] Similarly, Cohen, who would like to use AI to correct a bias in human arbitrators: P Cohen, ‘Bytes and Prejudice’ (2015) 1(1) Journal of Technology in International Arbitration 57, 66. This, in turn, could avoid anchor effects due to an upstream machine decision.
[705] In individual cases, this may also relate to the question of the extent to which the use of AI or a specific AI tool is preferable to the use of primary source material.
[706] See also SVAMC Guidelines (n 700) 9 ff, 13 (Commentary to Guideline 3), with the additional concern that innocuous and uncontroversial uses of AI should not be prevented by overly strict procedural requirements.
[707] See also SVAMC Guidelines (n 700) 6 f (Commentary to Guideline 1).
[708] SVAMC Guidelines (n 700) 7 f (Commentary to Guideline 1).
[709] SVAMC Guidelines (n 700) 18 f (Commentary to Guideline 7).
[710] SVAMC Guidelines (n 700) 14 (Commentary to Guideline 4).
[711] SVAMC Guidelines (n 700) 16 (Commentary to Guideline 5).
[712] SVAMC Guidelines (n 700) 9 (Commentary to Guideline 2).
[713] See in more detail above para 162 ff.
[714] See in more detail above para 161 ff.