Some background papers about how policies of various sorts might be enforced in the context of LLMs...
The Big Picture on Regulating LLMs
- "Frontier AI regulation: Managing emerging risks to public safety" (11 Jul 2023)
https://arxiv.org/abs/2307.03718
Note: Read this paper if you have no time to read anything else! It's probably the most substantial paper on regulation to-date, even more concrete than the Philipp Hacker paper (below). - "Regulating ChatGPT and other Large Generative AI Models" (updated 12 May 2023)
https://arxiv.org/abs/2302.02337
Note: Philipp Hacker is regarded as one of the leading thinkers on this topic worldwide. This paper, which has already been cited by many other writers, will be presented at the ACM Conference on Fairness, Accountability, and Transparency (ACM FAccT). June 12-15, 2023 in Chicago https://facctconference.org/ - "Two models of AI oversight - and how things could go deeply wrong" (08 Jun 2023)
https://garymarcus.substack.com/p/two-models-of-ai-oversight-and-how
Note: Gary Marcus is well-known for his challenges to contemporary AI, anticipating many of the current limitations decades in advance, and for his research in human language development and cognitive neuroscience. - "Dual Governance: The intersection of centralized regulation and crowdsourced safety mechanisms for Generative AI" (02 Aug 2023)
https://arxiv.org/abs/2308.04448
Note: The authors propose a framework designed to synergize centralized government regulations in a U.S. specific context and safety mechanisms developed by the community to protect stakeholders from the harms of generative AI. - "This is how GPT-4 will be regulated" (17 Mar 2023)
https://techmonitor.ai/technology/ai-and-automation/this-is-how-gpt-4-will-be-regulated - "Policy paper: A pro-innovation approach to AI regulation" (UK) (29 Mar 2023)
https://www.gov.uk/government/publications/ai-regulation-a-pro-innovation-approach/white-paper
Note: Presented to Parliament by the Secretary of State for Science, Innovation and Technology by Command of His Majesty on 29 March 2023 - "EU moves closer to passing one of world’s first laws governing AI" (Jun 2023)
https://bit.ly/3JAHKKr - "Congress is racing to regulate AI. Silicon Valley is eager to teach them how" (17 Jun 2023)
https://www.washingtonpost.com/technology/2023/06/17/congress-regulating-ai-schumer/
Congressional salary caps that pale in comparison to Silicon Valley’s sky-high paychecks make it difficult to retain staff technologists, putting US lawmakers at a disadvantage in getting up to speed — a goal that has become increasingly urgent as the European Union has leaped ahead of Washington... - "Schumer launches ‘all hands on deck’ push to regulate AI in keynote address" (21 Jun 2023)
https://www.washingtonpost.com/technology/2023/06/21/ai-regulation-us-senate-chuck-schumer/
Note: Sen. Schumer laid out five principles on Wednesday to guide lawmakers on AI legislation — including security, accountability and explainability — that expand on a “high-level” framework he first teased in April 2023.
Identifying and Tracking LLM output
- "A Watermark for Large Language Models" (6 Jun 2023)
https://arxiv.org/abs/2301.10226 - "On the Reliability of Watermarks for Large Language Models" (9 Jun 2023)
https://arxiv.org/abs/2306.04634 - "The Science of Detecting LLM-Generated Texts" (2 Jun 2023)
https://arxiv.org/abs/2303.07205 - "Can AI-Generated Text be Reliably Detected?" (17 Mar 2023 )
https://arxiv.org/abs/2303.11156 - "DetectGPT: Zero-Shot Machine-Generated Text Detection using Probability Curvature" (26 Jan 2023)
https://arxiv.org/abs/2301.11305
Note: DetectGPT, which only works for GPT-2, has been frequently cited in 2023. Googling around indicates there are many possible clones/fakes... - "New AI classifier for indicating AI-written text" (31 Jan 2023)
https://openai.com/blog/new-ai-classifier-for-indicating-ai-written-text
OpenAI's own classifier, trained to distinguish between AI-written and human-written text. Use the service itself at https://platform.openai.com/ai-text-classifier
Identifying and Tracking Diffusion Output
- "Tree-Ring Watermarks: Fingerprints for Diffusion Images that are Invisible and Robust" (01 Jun 2023)
https://arxiv.org/abs/2305.20030
Note: This steganographic technique, specific to diffusion models, has been cited as a possible way to track the sources of offensive output, including virtual child pornography.
Generative AI and Virtual Child Pornography
- "AI-generated child sex images spawn new nightmare for the web" (19 Jun 2023)
https://www.washingtonpost.com/technology/2023/06/19/artificial-intelligence-child-sex-abuse-images/
Note: Child-safety experts said many of the AI-generated images being shared appeared to have relied on open-source tools, such as Stable Diffusion, which is available as open source and can be run in an unrestricted and unpoliced way. - "Generative ML and CSAM: Implications and Mitigations" (24 Jun 2023)
https://purl.stanford.edu/jv206yg3793
Note: In this new paper Stanford Internet Observatory and Thorn (a nonprofit that fights the spread of child sexual abuse online) researchers found that, since August 2022, there has been a small but meaningful uptick in the amount of photorealistic A.I.-generated child sexual abuse material circulating on the dark web. The researchers found that the the majority of these images were generated not by Dall-E but by open-source tools that were developed and released with few protections in place.
Evaluating Safety, Trustworthiness and Fairness in Generative AI
- "DecodingTrust: A Comprehensive Assessment of Trustworthiness in GPT Models" (20 Jun 2023)
https://arxiv.org/abs/2306.11698
Detecting Hate Speech
- "Towards Legally Enforceable Hate Speech Detection for Public Forums" (23 May 2023 )
https://arxiv.org/abs/2305.13677
Note: This paper is potentially of high interest w.r.t. the application of KGs as authoritative references for policy enforcement; in their abstract the authors claim, "Our work introduces a new task for enforceable hate speech detection centred around legal definitions, and a dataset annotated on violations of eleven possible definitions by legal experts" - "ToxiGen: A Large-Scale Machine-Generated Dataset for Adversarial and Implicit Hate Speech Detection" (May 2022)
https://bit.ly/3NfPXWk Discussion: https://bit.ly/42pcHaY
Detecting & Mitigating Bias
- "Linguistic Models for Analyzing and Detecting Biased Language" (Aug 2013)
https://aclanthology.org/P13-1162.pdf
Note: Not specific to LLMs - "Challenging the appearance of machine intelligence: Cognitive bias in LLMs" (3 Apr 2023 )
https://arxiv.org/abs/2304.01358
Note: The authors demonstrate the presence of of over 180 known cognitive biases in LLMs and discuss the implications of using biased reasoning under the guise of expertise. - "Debiasing ChatGPT: Creating an LLM that isn’t racist or sexist" (13 May 2023)
https://glassboxmedicine.com/2023/05/13/from-chatgpt-to-puregpt-creating-an-llm-that-isnt-racist-or-sexist/ - "Bias and Toxicity in Large Language Models" (31 Oct 2022)
https://www.cs.princeton.edu/courses/archive/fall22/cos597G/lectures/lec14.pdf
Building Biases into AI
- "Why we need biased AI: How including cognitive and ethical machine biases can enhance AI systems" (18 Mar 2022)
https://arxiv.org/abs/2203.09911
Note: This was the focus of a recent TWC Lunchtime Reading Group meeting. This paper is provocative, as its abstract implies: "...this paper is the first tentative step to explicitly pursue the idea of a re-evaluation of the ethical significance of machine biases, as well as putting the idea forth to implement cognitive biases into machines."
LLM-assisted Copyright Infringement Detection
- "Playing with machines: Using machine learning to understand automated copyright enforcement at scale" (22 Feb 2023)
https://journals.sagepub.com/doi/pdf/10.1177/2053951720919963
Note: The authors train a ML classifier to identify videos in categories that reflect ongoing controversies in copyright takedowns, and use it to explore how copyright is enforced on YouTube.
Generative AI and Copyright Law
- "Generative AI Meets Copyright - Pamela Samuelson" (June 2023)
https://www.youtube.com/watch?v=6sDGIrVO6mo
Note: Pam Samuelson is the Richard M. Sherman Distinguished Professor of Law at UC Berkeley. In this piece she talks quite a bit about fair use and generative AI, and also suggests that copyright law may be the "Achille's Heel" of generative AI. - "Generative Artificial Intelligence and Copyright Law" (Updated May 11, 2023)
https://crsreports.congress.gov/product/pdf/LSB/LSB10922
Note: Prepared by the Congressional Research Service, a group of nonpartisan shared staffers
to congressional committees and Members of Congress - "Copyright Registration Guidance: Works Containing Material Generated by AI" (16 March 2023)
https://www.copyright.gov/ai/ai_policy_guidance.pdf
Note: This Guidance from the US Copyright Office advises that the only parts of a work that are copyrightable are the human-contributed ones, and the work is not copyrightable if an AI technology determines the expressive elements of the work and the creativity is not the product of human authorship. - "Authors file a lawsuit against OpenAI for unlawfully ‘ingesting’ their books" (05 Jul 2023)
https://www.theguardian.com/books/2023/jul/05/authors-file-a-lawsuit-against-openai-for-unlawfully-ingesting-their-books
Note: The plaintiffs claim their books were unlawfully “ingested” and “used to train” ChatGPT because the chatbot generated “very accurate summaries” of the novels. - Google & DeepMind are sued for their LLMs, including Bard, Lamda, Palm, and the to-be-released Gemini. (File 11 Jul 2023)
https://storage.courtlistener.com/recap/gov.uscourts.cand.415223/gov.uscourts.cand.415223.1.0.pdf
Note: This lawsuit cites a number of issues including copyright. It points to C4 as a problematic training dataset among others. More on C4 here...
Generative AI and Creative Commons Licensing
- "Why the Great AI Backlash Came for a Tiny Startup You’ve Probably Never Heard Of" (WIRED 14 Aug 2023)
https://www.wired.com/story/prosecraft-backlash-writers-ai/
Note: Raises the point that having terms of use in a CC license that exclude automated scrapers (such as LLM training code) might not be valid, since non-humans cannot be licensees.
LLMs and Scientific Communication
- "The role of ChatGPT in scientific communication: writing better scientific review articles" (15 Apr 2023)
https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10164801/ - "As scientists explore AI-written text, journals hammer out policies" (22 Feb 2023)
https://www.science.org/content/article/scientists-explore-ai-written-text-journals-hammer-policies
LLMs and Education
- "Educator considerations for ChatGPT" (2023)
https://platform.openai.com/docs/chatgpt-education
Note: OpenAI's overview prepared for educators seeking to learn more about the capabilities, limitations, and considerations for using ChatGPT for teaching and learning
LLMs and Legal Practice
- "The ChatGPT Lawyer Explains Himself" (08 Jun 223)
https://www.nytimes.com/2023/06/08/nyregion/lawyer-chatgpt-sanctions.html - "Generative AI and the Law" (Spring 2023)
https://www.lexisnexis.com/html/lexisnexis-generative-ai-story/
Note: A LexisNexis summary piece on the role of GenAI in practicing law. - "Who’s Afraid of ChatGPT? An Examination of ChatGPT’s Implications for Legal Writing" (26 Jan 2023)
https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4336929
Note: A very extensive article, which first explores ChatGBT's legal capabilities, then considers the practical applications and ethics. Among other uses, the author found that ChatGPT was able to identify logical flaws in contract clauses and could be a promising learning tool when teaching contract analysis.
Jailbreaking Generative AI-based Systems
- "The Amateurs Jailbreaking GPT Say They're Preventing a Closed-Source AI Dystopia" (Mar 2023)
https://www.vice.com/en/article/5d9z55/jailbreak-gpt-openai-closed-source
Note: Systems such as ChatGPT implement moderation filters designed to prevent users from retrieving harmful content; this kind of middleware may be essential to regulating generative AI. Communities of users share prompts designed to overrise these measures, with varying degrees of success.
Please contact John Erickson if you have suggestions for this list!