Constructing AI Speech
abstract. Artificial Intelligence (AI) systems such as ChatGPT can now produce convincingly human speech, at scale. It is tempting to ask whether such AI-generated content “disrupts” the law. That, we claim, is the wrong question. It characterizes the law as inherently reactive, rather than proactive, and fails to reveal how what may look like “disruption” in one area of the law is business as usual in another. We challenge the prevailing notion that technology inherently disrupts law, proposing instead that law and technology co-construct each other in a dynamic interplay reflective of societal priorities and political power. This Essay instead deploys and expounds upon the method of “legal construction of technology.” By removing the blinders of technological determinism and instead performing legal construction of technology, legal scholars and policymakers can more effectively ensure that the integration of AI systems into society aligns with key values and legal principles.
Legal construction of technology, as we perform it, consists of examining the ways in which the law’s objects, values, and institutions constitute legal sensemaking of new uses of technology. For example, the First Amendment governs “speech” and “speakers” toward a number of theoretical goals, largely through the court system. This leads to a particular set of puzzles, such as the fact that AI systems are not human speakers with human intent. But other areas of the law construct AI systems very differently. Content-moderation law regulates communications platforms and networks toward the goals of balancing harms against free speech and innovation; risk regulation, increasingly being deployed to regulate AI systems, regulates risky complex systems toward the ends of mitigating both physical and dignitary harms; and consumer-protection law regulates businesses and consumers toward the goals of maintaining fair and efficient markets. In none of these other legal constructions of AI is AI’s lack of human intent a problem.
By going through each example in turn, this Essay aims to demonstrate the benefits of looking at AI-generated content through the lens of legal construction of technology, instead of asking whether the technology disrupts the law. We aim, too, to convince policymakers and scholars of the benefits of the method: it is descriptively accurate, yields concrete policy revelations, and can in practice be deeply empowering for policymakers and scholars alike. AI systems do not in some abstract sense disrupt the law. Under a values-driven rather than technology-driven approach to technology policy, the law can do far more than just react.
Introduction
Artificial Intelligence (AI) systems such as ChatGPT can now produce convincingly human speech, at scale.1 That speech can take many forms, typically in response to user prompts and queries. AI systems can produce political manifestos, poetry, and fascinating facts. If it were produced by a human, such “high-value speech” would sit at the heart of the First Amendment’s protections.2 AI systems can also produce, unprompted or with very little nudging, wildly untrue hallucinations, recipes for explosives, and harmful lies about individual people. These are examples of “low-value speech,” some of which are unprotected by the Constitution and thus readily regulated, at least in theory.3
It is tempting to ask whether AI speech “disrupts” the law. That is, is there something special about AI-generated speech? Will the law be able to keep pace with it? Will the harms caused by AI speech at scale drive the creation of new and sui generis areas of law and regulatory systems? Or will regulation of AI speech be an example of Frank H. Easterbrook’s infamous “law of the horse”: old law applied, with minimal alteration, to newish things and behavior?4
These are the questions frequently asked in law-and-technology scholarship. And these are the wrong questions to ask. They lead to an approach that examines how the law responds to technology, which then leads to several major mistakes, including underexamination of the role existing law plays and underscrutinization of the often surprising conditions under which technology has developed. Instead of asking whether the law can keep up with technology, we should be asking how the law shapes and intervenes in technology and how it can do so—or refrain from doing so—in ways that further important values. A values-first approach to technology law would start not by asking whether the law is capable of keeping up with the development of technological or scientific expertise (it almost always is), but by asking what values technology, and the social practices of technology, further. It would ask how the law and its systems of sensemaking can and should be shaped to further values.
This Essay therefore examines what we have elsewhere termed the legal construction of technology, and more specifically the legal construction of content-generating AI systems.5 Legal construction of technology looks to the ways in which the law itself contributes to its encounter with, or really its incorporation of, the recently evolving uses of AI systems as “speech engines.”6 By shifting the main object of inquiry to the law and its features, the method reveals tropes and policy baggage, along with possible new points of intervention. To follow the well-trod path of tech-exceptionalist reasoning and ask, “How does AI speech disrupt the law?” is to blind oneself to more effective ways of discussing how the law encounters, and indeed actively constructs, AI speech.
We proceed as follows: In Part I, we define the method of legal construction of technology. The method examines how the law makes sense of technology, querying the objects, values, and institutions involved in such sensemaking. This is not a purely descriptive move; legal construction makes room for awareness of the pathologies of a particular regulatory path, and the possibility of other interventions.
We then apply the method in Part II. We begin by defining “AI”—not only in the sense of abstract technical specifications, but by placing the development of these systems in social, historical, and legal contexts. We then explore four ways of legally constructing AI speech: (1) as speech by speakers (within First Amendment law), (2) as speech at scale (within content moderation law), (3) as the output of risky complex systems (within risk regulation or tort law), and (4) as market behavior implicating consumer protection (within the FTC’s Section 5 authority). For each analysis, we query the values, objects, and institutions of regulation.
Together, these analyses show that no one feature of speech-generating AI systems is driving “disruption” of the legal system. The law, in encountering AI systems, already demonstrates a substantial array of tools, institutions, and orienting values. From analogies to impact assessments, the law is already deeply involved in making legal sense of AI.
Part III outlines the policy implications of this method. We each have clear normative leanings, spelled out in other works, as to which of these and other tools work best for governing AI systems.7 But our goal here is less to argue for particular tools than to show the things legal construction of technology can empower us to do. A values-first approach to regulating technology would leave each lawyer, policymaker, or scholar to determine—and to own—what values they want to advance and what harms they want to prevent, and to then choose their legal tools or regulatory abstentions accordingly.
We summarize and conclude with some parting thoughts. Legal constructions of technology can intertwine, bend, evolve, and mutate. What they resolutely do not do, in the face of purportedly new technologies like AI, is passively get acted upon and break. To see law, its institutions, and individual actors in this way is to afford new technologies some special force they do not have. Law and technology are each aspects of, and deeply embedded in, society. As such, they co-construct each other, in the context of deeply political systems that channel or constrain political power.8
I. legal construction of technology
We start with an introduction to our approach to answering the question, “Does AI speech disrupt the law?” We approach the question by rearranging its premise. Technology does not disrupt the law. The law does not follow technology around trying to “keep up”—that is not the nature of the relationship, or so argues what we have called the legal construction of technology. The legal construction of technology method takes the view that the law conducts its own kind of sensemaking of new technological developments and their social uses.9 The law makes sense of technology in very particular ways, using language, enforcement, abstention, institutional arrangements, and means for garnering nonlegal expertise, that change as legal cultures change. Law makes sense of, and indeed shapes, technology and its trajectories through legal tools and institutions. These tools and institutions are designed to advance particular values, and are aimed at particular people, behavior, or things.10
A. Defining Legal Construction of Technology
Legal construction of technology rests on the theory that technology should be understood through its cultural interpretation—related but not identical to a science-and-technology-studies (STS) approach.11 It is an analysis of how the long stories of law intersect with the long stories of technology, asking not whether some technology will require legal overhaul, but what choices were made across both social contexts such that a technology developed that would present legal questions worth considering. The law plays as much of a role as technology does: it embeds values, crafts institutions, and even sets up its own fault lines that later may crack. This method emphasizes that the law is worthy of deeper scrutiny; its values and theories, objects of regulation, and institutions all contribute to how technology is brought into contact with, and made meaningful within, the law.
What legal construction of technology does not do is ask as an endpoint, “What about this technology disrupts the law?” This framing presumes that technology does the acting, and law the responding. It treats technology as some outside force, perhaps even external to society. The deterministic framing typically also presumes that technological development inherently moves at a pace law is not equipped to follow (the so-called “pacing problem”).12 Echoes of this framing sound across early cyberlaw debates about whether the Internet was an “exceptional” technology,13 and whether it presented radically new features that would require radical changes in law.14
There are immediate and practical costs to the other way, of technological essentialism. Starting with the technology and asking how the law responds to it locks lawyers and policymakers into preexisting paths, as a matter of both law and rhetoric. As a consequence of politics and history, this often leads to a crippled legal system and strong technological system, in which innovation wins because innovation is best. Path dependencies occur.15 Legal scholars have detailed path dependencies in judicial use of precedent,16 administrative expertise and resources,17 and legislative policy.18 Part of the path, too, includes institutional expertise, resources, and culture. The path-dependent ways the law makes meaning of technology can hide significant policy choices in design specs,19 the quantification of risk or harms,20 and procurement policies,21 shielding substantive policymaking from democratic processes.22
In a more recent conversation, two prominent voices in the field of law and technology again debated technological essentialism.23 In an article comparing robotics to the Internet, Ryan Calo argued that the “essential qualities” of robots would necessitate “distinct conversation[s]” about the law.24 Robots’ essential technical qualities, Calo posited, would require changes in the law.25 Jack M. Balkin responded that drawing out a technology’s essential qualities was not particularly useful when thinking about legal application.26 Nor is boiling a technology down to essential qualities really even possible, given the way technology in society evolves over time.27
Calo certainly recognizes the social aspects of technology.28 But his initial approach—distilling technology to its essential features that act upon the passive law—reflects technological essentialism. We reject this. Technologies develop in societies with laws that set expectations about how they will work and why. Then, law does the daily work of maintaining those expectations. Law is always producing, maintaining, and “keeping up” with technology. Sometimes this is a matter of structuring how the law incorporates expertise from other fields.29 Other times, law “keeps up” with technology the same way it “keeps up” with other evolving social facts: through a combination of fuzzier standards, common law, and new legislation and regulation.30
These tasks do not always achieve evolving social goals and values, but that is a matter of politics, not the nature of the relationship between law and technology. 31 So, while Calo could claim that the fact of robots’ embodiment “disrupted” the way U.S. law handles software and liability, the same is not necessarily true of European law.32 Likewise, the U.S. legislation passed to change the legal category for connected computers (Communication Decency Act Section 230, discussed below), was a political effort made in response to lower-level court cases and only part of the legislative effort to shape content on the Internet. Today, Section 230’s special platform protections are not under threat from new embodied robots, but rather from political actors in Florida and Texas where state legislation attempts to restrict and govern how platforms moderate content.33
We find that legal construction benefits from a comparative perspective—whether comparisons within domestic legal systems between different subject-matter areas, or comparisons across legal cultures.34 A comparative approach, while not required, affords researchers greater perspective on what they are taking as given, versus where there is possibility to imagine things very differently, from analogies to values and institutions. In other words, if a technology appears to be “exceptional” to torts but not to privacy laws, or in the United States but not in Europe, it says more about the political histories of those areas of law and places than about the particular attributes of that technology.35
Legal construction of technology requires uncovering more than “what the law says” and “how the technology works”; it demands a deep understanding of the relationship between law and technology in context over time, and of the cultural evolution of governance. Often law-and-technology scholarship is interesting precisely because technological development makes salient something about existing law and governance structures that might not have been visible before.36 But again, that is not some special function of a technical artifact; it is the result of a process of continual legal meaning-making that sometimes, in constructing technology, highlights a feature or a bug that we failed to notice about the law.
The ways we make legal sense of technologies changes the way we design, integrate, and regulate them. For example, designers, marketers, and lawyers all use analogies to advocate for certain understandings of technologies. A designer might build a portable drive with cameras and microphones to be seen as a “family member,” while a marketer might give it a name like “home assistant,” and a lawyer might analogize it to a “tiny constable.”37 Is the cloud a fluffy ether, remote storage, or someone else’s computer? Are we thinking of email as letter-like, the way a user experiences it, or as a physical network?38 These sensemaking choices have value-laden impacts: they drive policy choices, like whether courts will protect an expectation of privacy and require a warrant for a search. There are normative and deeply pragmatic ramifications of understanding a driverless car as a chauffeur, versus group transportation, versus a proxy driver.
As outlined in Part II below, we already see wranglings over analogies for AI-generated speech, both in existing law, and in how commentators have suggested changing the law. A speech-generating AI system could be a speaker,39 or a platform or tool;40 a data processor,41 or a risky complex system.42 AI-generated speech can be analogized to encyclopedias; or speech-generating AI systems can be analogized to advice-offering physicians.43 Each of these analogies illustrates the law’s attempts at meaning-making and carry policy consequences. The fact that AI systems could, and in fact probably do, fall into multiple legal buckets or no existing legal buckets in a particular area of law, does not tell us that the law is failing to keep up with technology. It tells us that legal systems, or really legal actors, are already using the law’s tools of sensemaking to enact underlying values in decisions about what harms and benefits should be afforded, and to whom.
Legal construction of technology, unlike technological-essentialist and pacing-problem-oriented framings, puts legal actors where they accurately sit: not chasing technology, but as potentially powerful interveners in conception and design through to integration and enforcement. The method insists that law not needlessly cede its role in shaping society to technology and its benefactors. As one of us has explained, “[b]y chasing new technologies with legal solutions, law-and-technology scholars, as well as policymakers, unnecessarily accept a degree of irrelevance.”44
Legal construction of technology is also unavoidable. We all do it, searching for analogies or trying to match policy outcomes to existing institutional goals. Yet it is very easy to do it badly, by tacitly accepting what is, and turning to technological essentialism as the way to frame a core set of questions. Every law-and-technology scholar, wittingly or not, asks how the law makes meaning of technology. But not every scholar steps back to ask not, “What is special about this artifact?” but, “What are the costs of constructing technology from within existing governance systems, and is there a better way?” Legal constructionists see moments of technological sensemaking as opportunities for intervention, and interventions as part of the sensemaking process. While seemingly descriptive, legal construction as a method provides new possibilities for normative action, some of which we discuss in Part III below.45
B. Three Core Elements
Here, we identify three core elements of legal construction: (1) the values (theories) behind different legal subfields, (2) the institutions involved and the regulatory tools they deploy, and (3) the objects or targets of regulation. These elements make up a great deal, but not all, of how law might construct technology.46 Rather than just performing legal interpretation, effective legal construction of technology seeks to explicitly identify these elements and what they contribute to existing law’s sensemaking. It aims to reveal what we are not seeing: policy baggage, path dependencies, and blind spots. This, in turn, lets us step back to think about other possible interventions: how we might legally construct a particular technology differently.
Values. Each area of law has its driving theories, which typically prioritize particular values.47 These theories and their values aim, and indeed typically constrain, the law’s construction. For example, U.S. copyright law is widely understood to have utilitarian underpinnings, prioritizing the production of creative works toward “net social welfare,” often with an economic bent.48 Scholars have debated the extent to which this framing drives stronger rights for authors versus broader exceptions for remixes and reuse.49 But the values underlying U.S. copyright law are typically economic, not dignitary. This both explains the Supreme Court’s willingness to carve out special treatment for copyright law under the First Amendment (property rights being given special treatment to promote capitalism) and limits the ways in which copyright exceptionalism might be extended to analysis of regulation that protects dignitary rights, such as privacy law. Foregrounding the values of different parts of the legal system allows legal constructionists to better identify consequent blind spots. It allows us to move back from analogies and legal boxes to query whether we agree with underlying regulatory goals.
Institutions. Each area of the law, too, involves differing institutions and institutional combinations, which use different regulatory tools and perform different methods of legal construction toward different ends. For instance, without federal data-protection legislation, Americans have come to understand computer technology as consumer services regulated by the Federal Trade Commission (FTC). The FTC is culturally an enforcement agency, enabled and restricted by its charge to protect consumers from unfair and deceptive trade practices. As a consequence of history, it struggles to promulgate regulation; instead, it issues consent decrees.50 The FTC’s oversight has historically been limited to certain commercial entities and does not extend, for example, to nonprofits or to telecommunications companies acting as communications networks regulable by the Federal Communications Commission (FCC). While the FCC, as an independent expert agency, attracts the talents and expertise of economists and telecommunications engineers, the FTC largely attracts justice-oriented attorneys and investigators.51
These stories of institutions, their tools, their capacities, and their cultures enable legal constructionists to examine not just the content of substantive law, such as FTC consent decrees, but also the consequences. The consequences of framing affected individuals as “consumers” who have been “deceived,” rather than “data subjects” whose “fundamental rights” have been violated are both rhetorical and immensely practical.52 Further, these stories allow us to look at specific legal tools such institutions use and identify what one of us has termed the “policy baggage” of said tools: the pathologies inherent in different practices within the law and its particular ways of practical sensemaking.53
Objects. The third element we highlight in the examples in Section II.B below is the object or target of regulation. Law-and-technology scholars have long discussed whether laws are best crafted as technology-specific or technology-neutral.54 However, this is just one aspect of what we mean by examining the chosen object of regulation. The object of regulation tells us a lot about how an area of the law constructs technological practices. A law can target facial-recognition systems. It can target drones. It can target personal data, or content-moderation systems, or monopolies, or money. It can target an entire meta-class of “anything that harms consumers.” As we show in the below analyses, not only does identifying the object of regulation often provide a practical answer as to whether speech-generating AI systems are subject to that regulation, it also lets us understand what aspects of said systems are not being regulated or even noticed through a particular branch of the law. Identifying the object of regulation could help identify how such systems might be better regulated, or are already regulated, through broad, existing, technology-neutral laws that construct such systems not as an instantiation of the pacing problem, but as the product of readily regulable corporate actors. The recent U.S. (re)turn to antitrust enforcement against Big Tech illustrates this potential move.55
In short, legal construction asks: What system of legal meaning-making does technology encounter, toward what ends, using what institutions and tools? The rest of this Essay seeks to illustrate this by way of a closer examination through legal construction of the question: “Do AI speakers disrupt the law?”
II. constructing ai speech
A. Definitions
What is AI? The answer is more complex than the abstract technical specifications of an artifact. We begin by “defining AI,” for purposes of legal construction of AI-generated speech, as a technical, socio-historical, and ultimately legal matter. We then progress into how different parts of the legal system already construct AI systems.
When we talk about AI systems in the context of AI-generated speech, we are largely referring to Large Language Models (LLMs). Nina Brown offers a helpful working technical explanation for legal scholars: “Large language models (LLMs) are a type of deep learning algorithm used to model statistical relationships between words and phrases in large bodies of text data in order to generate human-like language.”56 LLMs are prediction machines: they predict what words or phrases come next, as a matter of (complex) statistics. As technical artifacts, such systems are trained on large databases, often containing personal data and copyrighted data, with substantial and ongoing human involvement.
LLMs, like other AI systems, are always human-machine systems; there is always a human in the loop, in both the narrower and broader senses of the term.57 Humans and organizations of humans pick goals and craft design parameters for these systems.58 Humans participate in reinforcement learning to refine and “optimize” system output.59 Humans and their institutions matter for legal construction of these systems in at least three ways: (1) Depending on how much one broadens the lens on what constitutes an AI system, there is always a human who could be regulated.60 (2) However, human-machine systems and interfaces are notoriously tricky to get right.61 (3) As we discuss below, the complexity of human-machine systems means that where law or policy attempts to treat AI as a fungible substitute for individual human actors, the fit is often off.
Like Google’s relevance search algorithm, the AI at the center of policy debates today should not be understood as objective math, though of course there is plenty of math involved. Trying to pull understanding “how LLMs work” apart from understanding LLMs as a product or a cultural artifact is a pointless and impossible endeavor. So, we provide some background not just on technical specifications but on LLMs as products or artifacts here.
The most relevant generative AI system is ChatGPT, which operates both as a standalone app and integrated into other products. As its own service, ChatGPT made headlines. As of September 2023, it touts 52.2 million downloads and $4.6 million in gross revenue from its subscription service.62 ChatGPT is also integrated into other systems like Loona the robot dog63 and Canva’s Magic Studio.64 It also offers enterprise licensing.65
ChatGPT is the product of OpenAI, which originally began as a nonprofit in 2015 but formed a “capped profit” entity with a specified and strange investment structure that early backers and founders like Elon Musk have criticized.66 Although the details have not been made public, according to reports, Microsoft invested $13 billion in OpenAI, starting with $1 billion in 2019. After taking a seventy-five percent share of OpenAI’s profit until the company makes its money back, Microsoft will have a forty-nine percent stake in OpenAI.67 OpenAI was also a product of a crypto-stung venture capital scene that dove into AI, investing more than one in every four dollars on AI startups in 202368 and driving up salaries for AI labor to close to seven figures.69 Other generative-AI systems have histories too, each different from the specific history of OpenAI.
None of these histories are predetermined, and all of them take place against an existing regulatory backdrop, whether the law has actively intervened or politics have driven a laissez-faire approach. Imagine how different these prediction machines might have been if U.S. states had amended their constitutions to include human rights to data protection, or if a market for personal data had been developed in the 1990s. Who would have built AI systems and why? What type of data would they have been trained on and what kind of audiences would be consuming their outputs? In an alternate universe somewhere, two law scholars have been asked how AI-generated speech disrupts communications law, wherein the FCC has been strictly regulating inputs to communications produced on or by networked systems since the telegraph—instead of largely taking a back seat to regulating the Internet and subsequent information technologies. What do AI systems look like in that universe?
Legal analyses often include a legal definition of a technological subject. This, too, is legal sensemaking. Typically, the policy debate over legal definitions focuses on how technologically specific a law should be, versus how much regulation should be oriented toward addressing non-technologically specific harms.70 For example, U.S. copyright law has over time become increasingly technologically neutral, moving from regulating, for example, books and maps, to attempting to encompass any creative work fixed in a tangible medium of expression, including online.71 Legal definitions can be the site of fervent political battles. They are often the targets of attempts to get out of regulatory coverage through characterizations of a technology as too new and thus challenging to regulate (e.g., we don’t understand General Purpose AI yet!72) or as not worthy of extra attention relative to existing technology (e.g., surveillance drones are no different from cell phone cameras!73).
The varied and changing legal definitions of AI systems already evidence these and other tensions.74 Mark A. Lemley and Bryan Casey claim that while it is tempting to start the regulation of AI or robots with a definition of the technology, “it can’t be done, at least not well.”75 Instead, they argue for a number of strategies for offsetting pacing issues, including regulating behavior rather than technology, and housing definitional decisions with courts or regulators rather than in legislatures.76
Claire Boine and David Rolnick track the evolving and competing definitions of AI systems in the EU AI Act.77 Boine and Rolnick’s work demonstrates the ways in which legal definitions participate in legal construction of technology, and how definitions interact with other aspects of legal construction. They note that the definition of AI systems in the initial 2018 draft of the Act focused on intelligence and agency,78 while the 2021 text from the EU Commission offered a far more general definition that encompassed nonautonomous statistical software.79 Later definitions, adopted by the other legislative bodies of the EU, returned to focusing on autonomy,80 distinguishing autonomous systems from statistical software.81 Boine and Rolnick also point to the European Parliament’s attempts to add a definition of General Purpose AI systems to the Act.82 They note, however, that the whole structure of the EU AI Act, with its focus on risks in the context of specific uses, and fit to intended purposes, is poorly designed for regulating General Purpose AI systems that are adaptable to many unpredictable uses.83 Their work demonstrates how definitions interact with other aspects of legal sensemaking—for example, decisions to focus on the developers of systems and on regulating uses in specific contexts.
B. Four Analyses of the Law(s) of AI-Generated Speech
In the following four analyses of the law (or laws) of AI-generated speech, we identify and query the values, objects of regulation, and institutions that different parts of the legal system bring to bear on the “problem” of AI-generated speech. In doing so, we hopefully illustrate what legal construction of technology brings to the methodological table.
Each analysis takes AI speech in a different legal context and reveals how the prioritized values of each space, which may also be contested and evolving, shape how we understand what the technology is and how it works. Depending on these values, legal actors ask different questions about AI’s structure, arrangements, and functions. Institutions with their own cultures, instruments, and political statuses address sometimes very different conceptions of the objects of regulation. The law, we hope to show, can come to construct AI-generated speech in these four different ways—and many more—with very different policy outcomes.
1. AI as a Speaker
Is AI-generated speech covered by the First Amendment? Is it protected? In this Section, we identify where First Amendment sensemaking of AI-generated content is still very much contested and ongoing—and at what levels (doctrinal versus theoretical versus institutional). We largely reason within existing First Amendment doctrine and try to be direct about the scope and feasibility of different possible interventions.84
The prototypical regulatory objects of First Amendment law are speech and speakers. That is, the First Amendment regulates speech by speakers (who are only sometimes human). A significant part of the First Amendment puzzle thus constructs the question of AI-generated speech as a “substitution problem”: a question of what happens to the law when you substitute an individual human actor with AI.85 This is, as we will see, only one of a number of ways one could legally construct speech generated by AI systems—even within current First Amendment law.
First Amendment values entail multiple theories of the First Amendment, some of which valorize and even prioritize quantity of speech, while others distinguish between higher- and lower-value speech. Nearly all of these theories, one of us has argued, are likely to drive First Amendment coverage of much AI-generated content, though they diverge in how much of that content might ultimately be constitutionally protected.86
The typical institution of the First Amendment is courts, particularly the politicized Supreme Court, interpreting the Constitution through two-party legal challenges. That is, First Amendment sensemaking happens through a particularly American form of legal analysis. Legislators have an attenuated role, but a role nonetheless; they craft the laws that courts then analyze.
To the extent that AI-generated content can be understood to “disrupt” First Amendment law, it does so largely because of an oddity about the doctrine’s regulatory objects. Different aspects of First Amendment doctrine differ in whether the target object of legal sensemaking is speech versus a human speaker. This potentially creates a scenario in which a great deal of AI-generated content is likely covered by the First Amendment as abstract and disembodied speech, but it is not clear how regulators can intervene to govern even historically unprotected speech, where the interpretative focus currently is on the intent of human speakers.87
By way of basic background, First Amendment analysis generally asks questions in two steps. The Court typically asks: (1) whether a genre of communication is covered by the First Amendment; and (2) if it is covered, whether it is constitutionally protected such that a law must be struck down as unconstitutional.88 Numerous Supreme Court cases ask whether something is “covered” as First Amendment speech or speech-entwined activity such as movies,89 parades,90 video games,91 the Internet,92 graphic design for websites,93 nude dancing,94 and transfers of doctors’ prescription data for purposes of advertising.95 It is only once something is found to be covered by the First Amendment—that is, constitutionally salient—that the Court will apply some form of scrutiny to the applicable law, including asking whether such covered activity nonetheless falls into a historically unprotected category of speech.96
Without getting into the edge cases, most actual speech—what the Court refers to as “pure speech,” whether spoken or written—is covered by the First Amendment. The Court recently referred to “pure speech,” distinct from embodied speech or speech-entwined conduct, as “[a]ll manner of speech—from ‘pictures, films, paintings, drawings, and engravings,’ to ‘oral utterance and the printed word.’”97 There are historic exceptions of “pure speech” that are nonetheless invisible to and arguably not even covered by the First Amendment.98 But it is likely most AI speech will be found by courts to be covered by the First Amendment.99 That is, if AI-generated content produces an antitrust violation or facilitates a crime, like a human-generated equivalent, that content probably will not be covered speech, but if AI-generated content constitutes “pictures, films, paintings, drawings . . . oral utterance and the printed word,” its regulation will likely trigger First Amendment analysis of some kind.100
But how can this be so? An AI system is not a human speaker with human rights. It doesn’t have emotions; it doesn’t participate in the political process; it doesn’t experience a chilling effect in which it stops speaking for fear of retaliation or shunning. The fact that AI-generated pure speech will probably be covered by the First Amendment tells us as much about First Amendment law as it does about AI systems.101 As Toni M. Massaro and Helen Norton observe, “[V]ery little in current free speech theory or doctrine makes First Amendment coverage contingent upon a human speaker.”102
Why is this the case? The values that drive First Amendment law, as interpreted by courts and the Supreme Court as First Amendment institutions, allow constitutional coverage and possibly protection of speakerless speech.
First Amendment coverage is explained by multiple theoretical justifications.103 The three most influential positive arguments for First Amendment coverage and protection are that protecting speech promotes democratic self-governance; protecting speech contributes to the marketplace of ideas and thus to the search for enlightenment; and protecting speech preserves individual autonomy and freedom.104 Under all of these positive theories, First Amendment rights can attach to speech because of the value of that speech to human audiences.105 That is, AI-generated political diatribes that influence a human reader’s vote would be covered by the First Amendment as speech that promotes democratic self-governance, contributes to the reader’s search for truth, and contributes to the reader’s autonomy. What matters is the speech, not the speaker; the marketplace and its consumers, not the humanity of the vendor.106
In recent years, the Court has additionally shown a growing trend toward invoking a negative justification for First Amendment protection, under which speaker identity matters possibly even less. These arguments focus on “the need to constrain the government’s potentially dangerous exercise of control over expression, and are rooted in distrust of the government as regulator.”107 If the Court’s focus is on what kind of actions the government may or may not take, the humanity of the speaker matters little if at all. If Congress bans AI speech that criticizes members of Congress, courts will and probably should find that law unconstitutional because of censorial government intent.
Thus for purposes of gatekeeping whether there is a First Amendment issue in the first place, the law typically asks, “Is there speech?” not, “Is there a human speaker?”108 Absent a radical reconfiguration of the Court’s current reasoning and doctrine, if AI-generated content is found to be speech, courts will typically next ask whether said speech falls into any historic categorical exceptions to First Amendment protection.109 Otherwise, if there is “pure speech,” a court will typically apply strict scrutiny and often overturn the law.110 (There are other exceptions—for example, for generally applicable laws that nonetheless affect expression.111)
The Court has carved out a list of increasingly ossified categorical exceptions to First Amendment protection of “pure speech”: defamation, fighting words, obscenity, and true threats, among others.112 For many of these historic exceptions, the doctrine asks, “How do we define unprotected speech so as to protect a (human) speaker?” not, “How do we carve out objectively unprotected speech?” This sets up a potentially thorny puzzle, wherein the internal logic of current First Amendment doctrine is likely to result in coverage of AI-generated speech but lacks a clear answer as to when such speech might not be protected and thus be regulable.
For example, First Amendment doctrine constrains defamation law such that to show defamation of a public figure, one must show not just that speech is false but that a speaker has actual malice—that is, knowledge of or reckless disregard for the falsity of its assertions.113 What does it mean to show that an AI speaker—or more probably its developer or user—has knowledge of, or has shown reckless disregard for falsity?114 This analysis is further complicated by the fact that AI systems are not individual speakers at all but are complex human-machine systems developed using complex supply chains.115
The Court’s selective focus on human speakers and human intent in First Amendment law is not going away anytime soon. In three recent cases, the Court doubled down on the centrality of (human) intent to historic exceptions to First Amendment protection.116 In Counterman v. Colorado, the Court found that the state of Colorado could not use an objective reasonable-person standard for determining whether speech constituted unprotected true threats.117 The majority held that the First Amendment requires some form of scienter for true threats, as it does for defamation and other categories of unprotected speech, in order to build a buffer around unprotected speech and prevent a chilling effect on (human) speakers.118 A standard of recklessness, according to the Court, struck the right balance between competing concerns of chilling protected speech and making it harder to prosecute unprotected harmful threats.119
In another 2023 case, United States v. Hansen, the Court again addressed intent, analyzing the category of “speech integral to criminal conduct.”120 The Court noted that both criminal solicitation and criminal facilitation contain a fairly strict intent requirement: “an intent to bring about a particular unlawful act.”121 It is unlikely that an AI developer creating a general-purpose content-generating system would ever exhibit such intent.
While Twitter v. Taamneh is not, strictly speaking, a First Amendment case, the Court’s discussion of intermediary liability for online terrorist content again indicated its current emphasis on intent as a backstop for expressive values.122 In Taamneh, the Court interpreted a federal statute that imposed liability for facilitating an act of terrorism and found that it required a level of scienter—”truly culpable conduct”—that did not encompass general-purpose online platforms and probably would not encompass most AI developers or platforms hosting general-purpose AI systems.123
If courts hew to these kinds of strict intent requirements, then AI “speakers” would get off the hook where human speakers would not.124 This would perversely incentivize more otherwise unlawful speech by AI systems—protecting more speech generated by AI than speech by actual humans. But as is often the case, this problem of First Amendment sensemaking of speech generated by AI systems is not an unsolvable one. It is not “disruption,” in the sense that the use of such systems has somehow broken the law.
There are multiple ways in which scholars, and courts, could intervene. They could intervene—and indeed have already been intervening—at the level of doctrine, arguing that there are more appropriate ways to think about duties of care and liability with respect to content-generating AI systems. For example, Nina Brown invokes products-liability law. Brown suggests that courts analyzing defamation claims arising from AI-generated content should look not to intent but to whether AI developers are making negligent design and development choices, such as training their systems on databases filled with inaccurate content or failing to adequately test their systems.125 Eugene Volokh, too, invokes a products-liability approach to AI defamation claims and argues that developers should be subject to a “notice-and-blocking” framework, which we discuss further in the next Section.126 Jane Bambauer explores negligence and duties of care, suggesting that depending on the specific facts of each case, the producers of AI systems might be analogized to doctors, lawyers, encyclopedia publishers, or even parents who negligently supervise their precocious AI children.127
Each of these scholarly moves shifts the legal analysis of AI systems from understanding such systems as speakers to trying to construe them as something else: products, encyclopedias, advisors, or precocious children. Within some of these more familiar analogies, we see hints, too, of something more interesting: the construction of the problem of AI systems as a problem of speech at scale,128 or as risky complex systems,129 or as a problem for consumer protection,130 each of which we discuss more below. That is, shifting the construction of AI systems away from the constricting box of speaker with human intent opens room for other ways of making legal sense of AI systems.
That’s all nice in the rhetorical abstract. But does First Amendment law as it currently exists leave room for these and other shifts in legal sensemaking? The answer is a very strong maybe.
First, while the Counterman majority made some strong claims that all pure-speech exceptions require some showing of intent, Justice Barrett’s dissent pointed out that several existing categorical exceptions appear to have objective standards, if they have any standards at all.131 Second, the Counterman majority outlined a recklessness standard that at times appears to leave room for liability for AI developers when they knowingly ignore substantial risks, and could maybe serve as a template for a sui generis scienter standard for AI-generated content in other categories of unprotected speech.132
Third, both Counterman and Taamneh described balancing tests or sliding scales of scienter, potentially leaving room for striking a different balance for AI-generated speech than human speech.133 The Court in Counterman balanced the harm of not prosecuting unprotected speech against the harm of a chilling effect on human speakers, declaring that a recklessness standard strikes the right balance.134 However, AI systems do not themselves experience a chilling effect. AI-generated speech, like commercial speech, could be characterized as inherently hardier than ordinary speech.135 There are plenty of economic incentives for AI-speech generation, just as there are with commercial speech, and thus less of a fear of a “chilling effect” on the companies that develop and distribute AI systems. If courts show a willingness to rebalance harms and chilling effects, this might result either in a version of recklessness that could be applied to AI developers across the board, or in the sort of objective product-liability-like standards Brown or Volokh discuss.
There is also room for adaptation at the level of First Amendment values. As one of us has suggested with co-authors elsewhere, courts could justify a different doctrinal standard for AI-generated speech by focusing on harms to the human listeners on whose backs coverage of AI-generated content is theoretically justified in the first place.136 When lawmakers restrict AI speech in the name of protecting the interests of a human audience, courts might mimic areas of law similarly justified by listeners’ interests such as commercial-speech doctrine or professional-speech doctrine by applying less exacting scrutiny and upholding regulations that aim to protect human listeners by preventing, for example, deception.137 How successful this argument might be with the current Court is another matter.
In these ways, law makes room for alternate or changing sensemaking. The law uses language, and language often leaves room for interpretation. The law is justified by values, and values might dictate different results, or might themselves change over time. The law is also a social system set up to handle evolving social facts. Nonetheless, the law is also made up of humans, and humans and their institutions can be deeply political.
This brings us back to institutions. The institutions enforcing the First Amendment are primarily the courts, especially the Supreme Court.138 The Court has recently been described as pursuing First Amendment Lochnerism, using the blunt force of the Constitution to overturn a host of regulations.139 The Court has been especially solicitous of religious and politically conservative speakers, potentially constraining regulation in all spaces, including those occupied by AI.140 The issue of platform liability has itself become highly politicized, with the NetChoice cases addressing Florida and Texas laws that were enacted to require purportedly liberal internet platforms to carry conservative speech making their way from the Fifth and Eleventh Circuits up to the Court.141 All this is to say: much of what is in motion on the legal side is spurred more by a shifting political landscape than by the introduction of purportedly fast-paced new technology.
In summary, the First Amendment constructs the puzzle of AI-generated content as largely a problem of intent and individual liability, or lack thereof. This may limit regulatory interventions and, at first glance, back lawmakers into the particular legal struggles of the “substitution” approach: AI does not itself have intent, and responsibility may be extremely hard to determine or apply, given the black-box nature of AI systems, the complexity of human-machine systems, and the complexity of AI supply chains.142 But as we will subsequently explain, other approaches to regulating AI-generated speech touch little if at all upon intent, opening up the possibility of broad regulatory interventions. Scholars of First Amendment law might do well to think about if or how that doctrine could make space for these other approaches, including regulation.143
2. AI Speech as a Problem of Speech at Scale
If the First Amendment focuses on speech by individual rights-bearing speakers, the closely adjacent field of content-moderation law instead legally constructs AI-generated content as a problem of speech at speed and at scale.144 (To be clear: content-moderation law may itself be backstopped by the First Amendment; that is not our focus here.145) For our purposes, we refer to this as “speech-at-scale” construction. It differs from the First Amendment construction not just in the substantive law—statutory schemes rather than constitutional interpretation—but also in its objects, values, and institutions.
Here is the argument for speech-at-scale construction: AI systems will quickly and at ostensibly low cost produce an enormous quantity of content. Some of that content will be high-value content (e.g., political speech), some of it will be low-value but legal content (e.g., hate speech and porn),146 and some of it will be illegal content (e.g., defamation). That content will, like human-generated content, be distributed both to individual users as the result of individual queries and to the general public online. The puzzle of AI speech at scale both involves and resembles existing discussions of online-platform liability and content moderation.147 Online platforms will have to moderate AI-generated speech, just as they do human speech. And AI systems may trigger questions about whether they should themselves be treated as speakers or information intermediaries.
Unlike First Amendment construction, which frames AI as an individual speaker and backs lawyers into a substitution problem focused on speaker intent, speech-at-scale construction focuses less on individual speakers and more on the problems of fairly moderating quickly distributed networked content at scale. Thus, the objects of speech-at-scale construction are typically communications platforms; content at scale, rather than speech from a rights-bearing individual; and individual platform “users,” rather than speakers in the Constitutional sense.
Speech-at-scale construction asks what kind of law establishes the best incentives and processes for ensuring that content at scale is on balance not too harmful, while also avoiding an undue chilling effect on both speech and innovation (whatever that means).148 The value central to AI as speech at scale is something like efficient fairness. And its sensemaking institutions are markedly different as well. Legislators serve as the primary, albeit not exclusive, policymakers—in many instances deferring to private platforms’ decision-making. Courts serve as the primary institution of interpretation of legislation.
While First Amendment rights and values exist in the backdrop of speech-at-scale construction, the central policy tools look little to nothing like First Amendment law. Under speech-at-scale construction, the law of AI-generated speech online is not much different from the law of human-generated speech online. Both are the law of intermediary liability and content moderation.
There have been, until very recently, two approaches to intermediary liability and content moderation in the United States. The first approach, exemplified by Section 230 of the Communications and Decency Act (CDA 230), immunizes online intermediaries from liability for either hosting or removing most user-generated content.149 The much-debated rationale behind CDA 230 is that if online platforms are on the hook for user-generated content, they will be incentivized to overcensor, and innovation and free speech will suffer.150 This rationale is paired with recognizing platform hosts’ need to moderate content to create domains in which people want to participate, without being held liable as publishers of any legally troublesome user-posted content. In the legal gap created by CDA 230, private governance of online speech has flourished, with companies “voluntarily” creating extensive private content-moderation regimes largely to ward off government scrutiny or additional law, or in response to collateral governmental pressures.151
The second approach to online liability and content moderation in the United States is the “notice-and-takedown” regime established in the Digital Millennium Copyright Act (DMCA), which creates a different system for copyright law: one that directly and in great detail dictates how companies should moderate online content.152 The DMCA regime is granular, highly procedural, and filled with minutiae aimed at balancing the legitimate interests of copyright holders and the potential chilling effects on those who post online content. The DMCA regime divides intermediaries into several different classes, requiring platforms that cache, host, or link to content to respond to a formal notification of copyright infringement by taking down the identified content.153 The user who posted said content has the opportunity to state that it is not infringing and request that it be reinstated (a “put back”).154 In recent years, companies such as Google have created a “para DMCA” or “DMCA-plus” regime that bypasses the DMCA’s requirements and safeguards and attempts to automate and monetize copyright content moderation.155 The takedown system is designed to achieve the same speech-at-scale value of efficient fairness, to avoid overburdening platforms with monitoring efforts, but also to address harms to copyright holders.
Until only very recently, these two approaches represented the two main ways of legally constructing speech at scale in the United States: the hands-off privatized CDA 230 approach and the highly proceduralized DMCA approach. However, in the past few years, several states (e.g., Florida and Texas) have imposed detailed transparency and must-carry requirements on online intermediaries that would interfere with privatized content-moderation policies—what might be characterized as a developing “third way” of platform liability.156 Aspects of these laws are soon to be challenged at the Supreme Court in NetChoice, LLC v. Paxton.157
There is no perfect content-moderation system. Under any approach, some legal content will be blocked, and some illegal content will slip through the guardrails. The policy value is thus to foster an overall system that works well enough and appears to be fair and effective—hence, the focus on procedure and transparency.158 In an ideal world, current content-moderation systems are calibrated to target undesired and sometimes illegal content efficiently and accurately.
There is also no inevitable content-moderation system. Content-moderation policy is the result of politics, and politics play out through institutions and people. As one of us has written:
Other countries have extended liability to platforms once the operator has knowledge of legally actionable content. Platforms are not considered neutral, automated systems and accountability is effectuated through human involvement and design choices.159
While CDA 230 may feel inevitable in the United States, countries in the European Union have long done content moderation differently. In Europe, until very recently, the E-Commerce Directive established a DMCA-like notice-and-takedown regime that applied to all potential online liability, not just to alleged copyright infringement.160 The system is calibrated differently. Platforms have higher monitoring burdens, but they publicly host less harmful content as a consequence. As we discuss in greater detail in the next Section, the recent Digital Services Act (DSA) supplants the old regime while also retaining its basic notice-and-takedown framework.161
For the most part, then, speech-at-scale construction will treat AI-generated content just as it treats human-generated content, as a problem of content at scale, solved by systems aiming toward efficient fairness. Perhaps policymakers will intervene in any number of ways: to require that AI-generated content be identified as such,162 or to recalibrate notice-and-takedown policies to adapt to faster content generation by AI systems. But these changes are neither inevitable nor required by the development of AI technology. The law may adapt, but once again, it does not break.
Then there is what we call the legal construction of AI systems as “splatforms” (speaker-platform hybrids). This analysis asks whether companies that produce AI-generated content themselves might be directly protected from liability by CDA 230. Scholars are split on this question.163 AI systems arguably loosely resemble search engines in that they both simultaneously create content (as speakers or platform users) and point users to information they did not generate themselves (as platforms or conduits).164 The majority of scholars have concluded that CDA 230 does not protect companies that offer algorithms that directly produce content, characterizing those algorithms as more like users/speakers than platforms/search engines. But Derek E. Bambauer and Mihai Surdeanu argue, in an interesting moment of legal construction, that the answer will be highly fact dependent.165 They argue that LLMs can be understood under CDA 230 as both a platform (for existing content within training databases) and a user (for newly generated content).166 The user of LLMs, they argue, is even more actively involved in the creation of content than the user of, say, a search engine.167 This characterization of LLMs recalls James Grimmelmann’s work on search engines as “speech engines,” wherein he highlighted the ways in which search engines operate as both conduits and speakers.168 We take no position here, but note the ways in which the splatform debate hinges not just on attributes of the technology but on existing categories in the law.
As Volokh’s recent work on generative AI systems illustrates, speech-at-scale framing can also be deployed in discussions of the direct, rather than intermediary, liability of the companies that develop and distribute such systems.169 Volokh reasons that OpenAI and Google are not covered by CDA 230, as the output of their AI systems is content their software has produced, not hosted content generated by others.170 He nonetheless incorporates speech-at-scale reasoning into what is ostensibly First Amendment-driven analysis.171 Volokh concludes that companies that create and distribute AI systems should be required under libel law to establish a “notice-and-blocking” regime similar to the DMCA—and that a failure to do so would constitute negligence and even “actual malice.”172
Volokh thus shifts from the substitution construction of First Amendment law (what, exactly, does it mean to ask if AI systems show intent of “actual malice”?) to speech-at-scale construction of content moderation (how does one balance the benefits of AI system output against their harms, through an efficient but procedurally fair system?).
To summarize, legally constructing AI-generated speech as a problem of speech at scale invokes particular legal practices, tools, and policy conversations. In the context of content moderation, it treats AI speech more or less as it does human-generated speech, with the object of regulation being networks and networked content: the large corporate entities that serve as online intermediaries or platforms, and their users. Speech-at-scale construction attempts to handle content moderation either by handing it off to private governance or by treating it as a problem of properly calibrating ex post individual processes and systemic transparency. The institutions relevant to the private management of these network systems include courts occasionally maintaining otherwise private management, though state legislators have started to try to restructure the network or recalibrate the balance in service of shifting values. Constructing AI speech as speech at scale is less concerned with respecting individual speakers’ rights or policing the substantive boundaries of protected and unprotected speech. Its primary value is calibrating efficient fairness between platforms and users, with some free-speech and innovation values sprinkled in.
3. AI as a Risky Complex System
A third way to construct speech by AI systems is through risk regulation.173 This approach constructs AI not as a speaker, nor as the generator of speech at scale, but as a risky complex system. The value this approach seeks to achieve is the ex ante mitigation of widespread, significant, and typically quantifiable harms: some quite concrete, and others harder to measure and more contestable.
The risky-complex-system approach appears in at least two places. One construction, currently largely taking place in Europe, is highly regulatory in nature. It uses regulatory tools borrowed from other areas of the law, such as data-protection law, environmental law, and product-safety regulation. That approach focuses primarily on ex ante risk assessment and risk mitigation, either in soft law or with centralized agencies as institutions exerting light-touch to moderate top-down control.174 Another construction, which may be more familiar to U.S. readers, constructs AI as risky complex systems through American tort law.
The object of a risky-complex-system construction is not just the AI system itself but also the humans and organizations around it. This construction is more like the AI-human systems of content moderation than the individual AI speaker constructed through First Amendment doctrine, but unique in important ways that present opportunities for powerful forms of intervention.175 To construct AI as a risky complex system, one shifts the conception of AI away from viral social media posts and toward nuclear plants, cars, and medical devices, where regulation aims to mitigate the risk of failure and bolster resilience through a number of regulatory tools. Whether through mandated design specs or more general recording and reporting requirements, legal construction of risky systems typically accepts that such complex systems will inevitably crash, but, with varying degrees of firmness, tries to task the developers and users of such systems with lowering risks to the population at large.
The risky-complex-system construction is increasingly popular, including outside of the context of AI-generated speech. Light-touch risk regulation has rapidly proven to be the go-to tool for governing AI systems in general.176 For example, both the EU’s General Data Protection Regulation (GDPR) and the EU AI Act use ex ante risk assessment and risk mitigation to address perceived problems with algorithms and AI.177 Each law requires the developers of “risky” AI systems to conduct an impact assessment (GDPR) or conformity assessment (AI Act) and to mitigate discovered risks. While neither the GDPR nor AI Act are, strictly speaking, targeted at AI systems as speakers, they both target AI as risky systems: the GDPR applies to systems developed using personal data, and the AI Act applies to systems it categorizes as “high risk” or “unacceptably risky.”178 The value this approach seeks to achieve is the prevention of widespread, significant, and quantifiable damage. Many other regulations and guidelines center around risk regulation, including: the National Telecommunications and Information Administration AI Risk Management Framework, the Singapore Model AI Governance Framework, and several proposed U.S. laws both at the state and federal level.179
A similar move to risk regulation is now occurring in the context of speech and content moderation. There has been a recent shift in legal scholarship from talking about individual First Amendment rights, or intermediary liability and efficient process, to talking about speech systems and population-level risks.180 This kind of sensemaking funnels lawmakers to a different kind of regulatory instrument: constructing the old online speech-at-scale problem as a problem of risk regulation. It moves the conversation from being about process or substantive individual rights to one about designing a regulatory system that involves ex ante mitigation and ex post resilience, and often centralized regulators.181 This approach shifts the object of study to system design and the institutions to administrative agencies and potentially state courts.
This move to risk regulation of online content systems is echoed and implemented in the EU’s new DSA. The DSA in part preserves the EU’s historic 2000s framework for intermediary liability discussed above; that is, it constructs online content as a problem of speech at scale, focusing on due-process protections and systemic transparency. But the DSA also shifts toward legally constructing content moderation as a problem of large-scale risky systems. The DSA establishes for the first time risk-assessment and mitigation requirements for Very Large Online Platforms and Very Large Online Search Engines.182 The DSA’s legal toolkit includes: formal risk assessments, risk-mitigation plans, third-party audits,183 the appointment of compliance officers,184 oversight by accredited researchers who get access to internal data,185 and a slew of public transparency measures, including publicly available repositories of information about advertisements, and public transparency reports that include information about risk assessments, mitigation, and audits.186 The DSA’s risk regulation involves ex ante intervention, including in product design.187
Once again, this is an example of how the law does not just react to technology. The EU could have stuck to speech-at-scale construction for online content moderation, pursuing the value of efficient fairness. Instead, it returned to the problem and supplemented speech-at-scale construction with a risky-complex-systems approach.
Stateside, there is no centralized risk regulation—yet. But scholars have pushed to establish risk mitigation for AI systems by making sense of them through tort law. One could imagine courts assessing tort claims eventually establishing a requirement of ex ante risk regulation for libelous AI systems, just as tort litigation eventually led to risk-mitigation requirements for car manufacturers.188 For example, Bryan H. Choi details the long history of courts dismissing claims of buggy software, because they rarely involve a physical injury and it is difficult to determine causality in complex computational systems.189 Choi explains that even if a physical injury is presented, software errors “cannot be reasonably prevented via ex ante design or ex post testing.”190 Like software before it, AI systems of today certainly cannot be made error proof. Choi argues that we should shift from a prevention to mitigation pathology, arguing for “crashworthy code” that would create a duty to minimize injuries when the inevitable error occurs.191
Nina Brown, too, discusses risk regulation through tort law. As discussed, Brown analyzes the specific issue of chatbots generating defamatory content.192 After recounting the challenges AI chatbots present to assigning liability under traditional defamation law, namely the tenuous link between developers and the actual defamatory statement and the difficulty of proving the requisite mental state, she suggests products-liability law as an alternative.193 For defective design, Brown would construct AI as a consumer technology with expected uses and require that developers mitigate foreseeable harms.194
Calls to treat software as a product for products-liability purposes stretch back at least to the 1980s.195 The reasons those calls have been unsuccessful have little to do with the difference between software from decades ago and AI today. Nonetheless, a strong sense of novelty, societal pressure, and institutional drive by the right actors, may yet bring this construction to the fore.
The risky-complex-systems construction thus brings different regulatory tools to the table. It expands regulatory imagination—or, at least, moves to transplant tools from other legal areas such as environmental law—demonstrating that there are more possible points of legal intervention than notice-and-takedown regimes.196 But as one of us has noted elsewhere, risk regulation, like any form of regulation, is not a neutral tool. It brings with it “policy baggage” that includes a tendency to frame what should be policy discussions as obscure mathematical or scientific questions for technical experts, and a tendency to undervalue harms that are harder to quantify. Moreover, it reflects a willingness to accept risks, rather than prevent harms, and to deploy a technology regardless of the risks it brings or the other policy interventions its use eschews.197
In the speech context, aspects of the risky-complex-system construction raise important questions about freedom of speech and the extent to which regulators can pursue ex ante, wholesale interventions that downplay individual rights.198 Understanding AI through the value of risk mitigation, the institutions of expert agencies and state courts, and the object of system design, directs us to investigate, define, and demand different things about the technology than constructions of it as speaker, or as the producer of networked content or a network itself.
4. AI and Consumer Protection
In our fourth and final example, AI can be understood as speech that messes with the marketplace, either by deceiving consumers or by providing inadequate notice of how such systems work, under consumer-protection law. The regulatory objects of consumer protection law are consumers, businesses, products, and more broadly speaking, the marketplace. The consumer-protection construction understands AI not as a speaker, but as a product or a manipulative salesperson. AI may fail to adequately disclose information pertinent to consumer choice, or may generate information that manipulates consumer choice in a marketplace. The consumer-protection approach is situated within consumer-protection agencies (institutions), contextualized by experts in investigation and enforcement. Its values are market fairness and efficiency—in contrast to content-moderation law’s narrower goal of procedural efficiency.
It would be hard to overstate the power the consumer-protection construction has had over the way computers are understood in American law and beyond. The United States created an area of laws, rights, and institutions for computers through consumer-protection regimes the way Europeans did over the course of the latter half of the 20th century through data protection. Enthusiastic and important political actors carved out a path through the Department of Commerce and the FTC (the relevant institutions in the consumer-protection framing), paying less attention to alternatives like the FCC, courts, or changes to state constitutions to recognize data-privacy rights as Europe does.199 As such, the FTC was relatively well-positioned to move into the AI arena, providing a legal understanding of AI as a factor in markets and defining its object of regulation as the consumer experience in a fair marketplace.200
The FTC frames AI as data and a model to scrutinize,201 though its most recent actions have also prioritized scrutiny of talent and computational resources.202 This framing capitalizes on the FTC’s mandate in Section 5 of the Federal Trade Commission Act to prevent unfair or deceptive practices. As such, AI garners the attention of the FTC when used to “influence people’s beliefs, emotions, and behavior.”203 Broadly, the FTC will investigate models that cause more harm than good, “that is, in Section 5 parlance, if it causes or is likely to cause substantial injury to consumers that is not reasonably avoidable by consumers and not outweighed by countervailing benefits to consumers or to competition.”204 According to the agency, this includes discriminatory outcomes like the case of digital redlining against Facebook brought by the Department of Housing and Urban Development in 2019.205
Maintaining an efficient and fair market involves correcting information asymmetries. Thus, transparency and duties to warn are essential policy tools for the consumer-protection objective of keeping consumers informed. There are a range of potential notice requirements that could become the dominant legal approach to AI, similar to the way an overemphasis on notices swallowed the legal approach to consumer-data privacy in the United States. There is the potential to label AI simply and appropriately as potentially dangerous products, leaving it to consumers to accept risks, abstain, or go shopping.206
The consumer-protection construction of AI systems may be gaining momentum in the United States. In July 2023, the FTC announced its investigation into OpenAI, the high-profile startup that created ChatGPT. In particular, the agency is investigating unfair practices related to “reputational harm.”207 It wants details of all complaints OpenAI received about “false, misleading, or disparaging” statements generated by its products.208 The agency is also looking into the way that ChatGPT stores and uses interactions with users after an incident in March 2023 that exposed some users’ conversations with the system—and whether proper notice about recording and training are effectively presented to consumers. 209
The FTC has also focused on deceptive notices. This could lead to an agency emphasis on requiring companies to effectively articulate the way AI systems collect and use data as well as checking bold statements about what the systems can do. For example, the FTC brought enforcement actions against Facebook in 2019210 and Everalbum (a photo app) in 2017211 for misrepresenting their uses of facial-recognition technology. Regulatory actors might also take issue with statements about AI systems’ capacity. “AI snake oil” is a term popularized by Arvind Narayanan, who shared presentation slides with the title on Twitter.212 Commentators use the term to emphasize the hype around AI and focus the public’s attention on the limitations of AI systems. In consumer-protection construction, rather than constructing AI systems as producing dangerous public discourse, critics and regulators construct them as the objects of false commercial speech intended to direct consumers to products and services.213
Preventing consumer manipulation is a central mandate of the FTC. Thus, another aspect of the OpenAI investigation is how difficult it is to opt-out or change defaults, in particular how difficult it is to disable chat history and whether those steps constitute a “dark pattern.” “Dark pattern” is a term that refers to computer-interface designs and strategies that trick, nudge, or mislead consumers to act contrary to their intended goals. While tricks, nudges, and misdirection are not new to marketplaces, “dark patterns” was coined by a researcher in 2010 specifically for digital interaction.214 The FTC’s 2022 report, Bringing Dark Patterns to Light, provides extensive examples of design changes, like font sizes and colors, and nudges and tricks, like confusing language and hidden settings.215 Although the report makes no mention of AI, the FTC has since warned companies not to overstate claims about AI products216 and not to skimp on notifying consumers about how data may be used to train AI systems.217 The agency is on the lookout for dark patterns in chatbots, targeting, and embedded ads.218 This narrow set of relatively aggressive activities represent the institutional tools and culture of law enforcers promoting consumer empowerment that enables purportedly free choices in an open market.
The values of the FTC all relate to consumer protection that serves efficient markets. As a consequence, the FTC has been limited to only a tangential authority over commercial computers and data as objects of regulation, with broad gaps in its authority (for example, over nonprofits, air carriers, banks, and telecommunication companies).219 Yet the FTC’s political prominence, resources, and tools have made it the de facto governance force for AI developers to consider in the United States.
One positive feature of the consumer-protection construction is that the FTC, with its legislative mandate to go after the broad and evolving standards of deception and “unfairness,” has a long history of being unphased by technological newness. FTC Chair Lina Khan has made it clear that she is not interested in AI exceptionalism.220 However the FTC’s efforts shape AI in the future, it is important to consider that the FTC has already been fundamental to the existence of the AI we have now. The information economy and the data, algorithms, and computing power that make AI possible were fostered by an FTC understanding of technology and governance. So, as the FTC makes sense of AI, it does so as a coproducer of AI in the U.S. context.
II. legal construction as a policy choice
This brings us to the “so what?” section, in which we address the normative valence of legal construction of technology. We have walked through four different legal constructions of AI-generated speech: as speech or speaker, as speech at scale, as the product of risky complex systems, and as a faulty digital product subject to snake-oil sales tactics under consumer-protection law. We have examined the meaning-making modes, tools, and institutions of the law. Assuming all of this has seemed descriptively accurate enough, what is a lawyer, scholar, or policymaker to do with it?
We each have thoughts, spelled out at length elsewhere, as to which of these legal constructions might be more descriptively accurate as to the technology. Additionally, we each have thoughts as to which institutions and policy tools, or combinations thereof, might best achieve regulatory goals.221 Our aim here is not to win the reader over to what we’ve each already argued about regulating AI systems. Nor does this Essay’s length afford us the space to once and for all solve the puzzles and problems raised by AI-generated speech. Rather, we want to win the reader over to the legal-construction-of-technology method and its concrete normative policy benefits, by showing its significant implications for conversations about AI-generated speech.
As ongoing practitioners of the method, we are still uncovering what it affords us. We hope to convince you that while one key benefit of legal construction of technology is its descriptive accuracy, the method isn’t merely descriptive. Sometimes, it is revelatory. And often, it is deeply empowering.
Legal construction of technology is descriptively accurate. We don’t want to shy away from the importance of the fact that legal construction of technology accurately describes what we think the law is doing. The law conducts sensemaking of technology, through language and institutions, toward particular values or goals. Each of the above constructions involves different institutions with different histories, performing meaning-making toward sometimes wildly different normative ends. The First Amendment attempts to legally construct through litigated doctrine, AI-generated content as speech or AI systems as a speaker, toward values of democratic participation, pluralism, and the protection of individual speaker and listener autonomy. The FTC attempts to legally construct AI-generated content as consumer deception or a problem of inadequate disclosure or bad data governance, with the goal of maintaining a functional marketplace in which information asymmetries are addressed and consumers are adequately protected from self-interested company behavior. Both of these analyses are descriptively accurate, if simplified, versions of how U.S. law handles AI speech.
If this observation—that U.S. law constructs AI speech through different institutions in different ways, at the same time—is all that legal construction of technology afforded us, it would still, we think, be a meaningful step in the right direction. It may be descriptively accurate to say in a narrow sense that AI systems create an interesting problem for First Amendment doctrine because of the law’s lopsided focus on the intent of human speakers. It is not descriptively accurate to say that AI-generated content “disrupts the law” in the sense of being generally hard for the law to regulate, just because construction within First Amendment law takes a hard look at human intent sometimes.
Legal construction of technology can also be revelatory. As a method, it leads us to identify salient aspects of the law and allows us to more accurately pinpoint where and how puzzles arise. This then lets us come up with creative and appropriately targeted interventions.
For example, it is possible to use straightforward legal analysis to arrive at the observation that running AI speech through First Amendment doctrine gets sticky around exceptions and speaker intent. Several scholars have already done so.222 It is also possible, using the kind of analogous reasoning that resounds throughout the law, to suggest solutions for this problem: treat AI speech as an encyclopedia, the advice of a doctor, the output of a product, or the result of faulty parenting.223 This reasoning, we believe, in fact is performing legal construction of technology: using available tools of legal sensemaking to let the law get to the “right” normative outcome—whatever that may be for an individual lawyer, policymaker, or scholar. It is not, however, taking advantage of what a more self-aware and deliberate approach to legal construction of technology can afford.
The legal-construction-of-technology method lets us identify that First Amendment doctrine constructs content differently, at different moments: as speech, when it comes to coverage, and as speech by a human speaker when it comes to historic exceptions to protection. That is itself not just a description, but a minor revelation. That’s surprising, and a bit weird! So, why is it happening—this bifurcation of caring about abstract speech versus caring about human speakers? Because (leaving Lochnerian skepticism aside for just a moment) in service of protecting public-sphere pluralism, contemporary First Amendment doctrine has been structured to pull more and more laws more readily under protective and purportedly neutral judicial scrutiny, and to be highly careful about allowing partisan actors to regulate speech. The big-picture story motivating contemporary First Amendment doctrine over at least the past half-century has been that courts help to structure a pluralistic public sphere, where we all learn to live with speech and opinions that may deeply offend or even hurt us—and with courts protecting minority human speakers whose speech might readily be chilled.224 Widening the threshold for coverage by constructing more and more laws as touching on “speech” gives judges more oversight and power; interpreting exceptions from constitutional protection as applying only to “speakers with adequate (human) intent” is meant to protect minority (human) speakers and their (human) audiences from chilling effects in the face of majority, partisan rule.
Thus, examining First Amendment law’s construction of AI speech reveals the following about the area of law: First Amendment law is the situated product of courts creating doctrine in a liberal pluralistic democracy, in response to clever lawyers who know how to trigger that liberal pluralistic framing by invoking a specter, sometimes true, of partisan interference in public debates, in front of a Supreme Court increasingly solicitous to religious and conservative speakers. A core value running through this project is that quantity of speech is a good thing: the mode by which we hope “good speech” rises to the top is by relying on sheer quantity of and competition between differing information and opinions, because we cannot trust elected political leaders with polemical views to prevent harms while also preserving deeply valued pluralism,.225 But if quantity of speech is actually part of the problem—of misinformation, consumer manipulation, the circulation of deep fakes, and election fraud—then this pluralistic-individualistic First Amendment framing may have to shift.226 Or at least, courts will have to make room in the doctrine for handling the massive amount of communicative output generated at low cost by nonhuman speech engines, distinguishing in some way between them and the opinion-holding human speakers and listeners that the liberal pluralistic framing is designed to protect.
Legal construction of technology thus makes evident particular oddities in legal sensemaking.227 It makes us explore what, exactly, those particular oddities might in turn evidence about the broader area of law, its institutions, and its underlying values. It reveals what might happen when existing values are pursued to illogical or inconsistent ends, or when the makeup of interpreting institutions changes.
That’s all very big picture, though. More practically, the method yields some very concrete policy revelations, too. It reveals, for example, the limits of resorting only to analogies for construction. Sure, an AI system is not a malicious human speaker. But it’s also not an advice-offering doctor, a precocious child of a negligent parent, or a physical product like a chair. It’s a complex human-machine prediction system with a befuddling supply chain, often deployed as software-as-a-service. It’s not that legal construction of technology is the first to identify the significance or limitations of analogies in legal reasoning. But it enables one to do interesting things with analogies: be appropriately self-aware of their significance and the constructions they trigger, recognize their limitations, recognize their embedded normative goals, and recognize the extent to which technological design is itself nudging you toward a particular analogy (like understanding email as a letter).228
Legal construction of technology, as noted above, frequently yields the insight that the developing uses of a technology will likely be constructed differently through multiple different institutions at the same time. Awareness of these differing constructions can be revelatory in a number of ways. It can lead to effective counterfactuals: Why compare AI to a chair, as one part of the legal system might do, if it’s more accurate to talk about it as a complex risky system, as another part does? It can lead to crucial explorations of how different areas of the law interact: Does one mode of construction leave important things out? Does another mode adequately fill identified gaps? Do the different modes preclude each other? Leave room for each other? Create helpful amplifications, redundancies, even resilience with each other?229
Let’s again make this more concrete with examples drawn from our four analyses above. If we think that speech-at-scale construction, risky-complex-system construction, or consumer-protection construction are in some ways desirable, does First Amendment doctrine leave space for, or even amplify, those approaches? Right now, at least, the Supreme Court’s First Amendment caselaw hasn’t had to deal directly with the constitutionality of speech-at-scale regulations, such as notice-and-takedown regimes, or for that matter, with risk regulation of speech. (It has, however, potentially started dismantling aspects of consumer protection, much to the dismay of many.230) Congress’s policy decision to immunize online platforms in CDA 230 actually punted First Amendment analysis of speech-at-scale regulation by, for the most part, privatizing content moderation instead of drafting command-and-control laws that would be subject to First Amendment analysis. (Blake E. Reid has referred to this as the First Amendment’s “interpretative debt,” attributable to CDA 230.231)
Thus, although First Amendment questions around AI speech are exciting, the First Amendment probably doesn’t matter much, at least not right now. This reality has nothing to do with how AI works or what it can do. U.S. legal choices have left considerable space for the European Union’s DSA risk-regulation construction to be important and influential, even to U.S.-based AI players. While the United States has abstained, the European Union has proactively regulated in this space, openly pursuing global influence through the so-called Brussels Effect, whereby global companies subject to EU regulation de facto export EU laws by abiding by them elsewhere.232 That is, global private companies may start doing a DSA-style mix of due process and risk mitigation all over the world, and they have the space to do that in the United States because of the regulatory gap CDA 230 creates and preserves.
But legal construction is historically contingent, politics are ever changing, and a shift to institutions and U.S. jurisprudence could change everything. If the Supreme Court decides to let states regulate platforms with must-carry laws, which construct content moderation as something like “common carriage,” those laws will compete with the DSA’s construction.233 A decision by the Court that platforms’ moderation decisions aren’t covered by the First Amendment may leave room, too, for stateside risk regulation in this space. If alternatively, the Court decides there is a First Amendment right of some kind for platforms to conduct content moderation, the ability of the U.S. government to regulate platforms and design features in the future will be significantly limited. On the one hand, this may block stateside DSA-style risk regulation and government-driven notice-and-takedown, but on the other, it will preserve the CDA 230 gap and thus may perversely strengthen the Brussels Effect of the DSA’s construction.
The interaction of judicial interpretation of CDA 230 and the First Amendment with the European Union’s risk regulation is just one way these constructions create interlocking pieces. Legal construction of technology encourages its practitioners to think seriously about design coherence between the systems—and to think carefully about how one ill-fit system of construction might impede better-fitting or more effective systems.
Practicing legal construction of technology thus takes us out of narrow frameworks. It removes blinders. It makes clear what aspects of the law the technology is running into, so that you can fix them. This brings us to empowerment.
We find legal construction of technology to be immensely empowering in the policy realm. That is because it inherently pushes lawyers away from surrendering to the idea that the technology is in charge of legal changes, and instead asks them to be creative about imagining how things might be another way. Here are just a few examples.
Legal construction of technology can lead to sometimes helpful legal transplants. A legal transplant occurs when the tools or constructions of one type of existing law are brought to bear on a problem arising in another.234 Transplants can be transnational: for example, when a country copies or imports the U.S. approach to notice-and-takedown in copyright law,235 or when U.S. states copy aspects of the GDPR.236 Transplants can be domestic, such as when states copy each other’s privacy laws.237 Or transplants can occur across different policy fields, such as when policymakers use metaphors and regulatory approaches from natural-resource law to address the allocation and governance of digital assets.238
Legal construction of technology empowers us to discover transplantable constructions, select more appropriate transplants, and be wary of the troubles a particular transplanted construction might bring. For example, AI risk regulation transplants a lot from other areas of law, including data-protection law, environmental law, and EU product-safety regulation.239 Boine and Rolnick observe that the EU AI Act stitches together a particular version of risk regulation with EU products-safety law. They claim that this particular transplanted way of legally constructing AI systems—with its ultimate focus on ex ante mitigation, product design, intended purpose, and fit for purpose—failed to contemplate or prepare for General Purpose AI.240
Legal construction of technology enables us to bring different, maybe better, modes of construction to the table, within an existing meaning-making mode. For example, if one is a First Amendment scholar, one could get trapped in substitution debates (e.g., what to do about intent?). Or, one could use analogies to transplant another mode of construction. We see, for example, Brown’s suggestion that courts analyze defamation claims through a products-liability lens as an attempt to port risk mitigation into the tort-speech interface.241 We similarly understand Volokh’s suggestion that First Amendment defamation law leaves room for “notice-and-blocking” requirements as an attempt to transplant speech-at-scale construction into the First Amendment.242
Legal construction of technology, too, helps us discover when the necessary intervention must occur at the level of values. If we can identify that the problem is that the motivating forces behind the minutia of laws and regulation are mismatched to the problems we’re trying to solve, then we know to look elsewhere for solutions. If one’s concern over AI-generated speech is that it will destroy democracy, the right policy answer is probably not going to be a market-driven policy solution such as copyright law.243 Or, to return to a previous example, if the problem one is trying to solve is caused by scale, then one needs to be thinking in terms of scale-directed solutions, driven by values that don’t overemphasize the scale of speech production as an inherently positive thing.
Finally, legal construction of technology enables lawyers to figure out the shape of the particular issue they’re trying to solve and draw on appropriate tools. If we’re in an expertise-gap situation, we have plenty of tools within the law for getting more technical expertise into policymaking. Not every problem is the pacing problem, but when the concern is that companies developing technologies are attempting to engage in regulatory arbitrage by falling between legal boxes and into legal cracks, the law has plenty of tools (with tradeoffs) for trying to get around that problem.244
Conclusion
A stable legal construction of AI has not yet set in, and negotiations are occurring with all the histories of global and national technology governance as the setting. We have presented only four potential constructions of AI to show a values-first approach, as opposed to a technology-first approach, to technology governance. From copyright law to tax law to competition law, there are plenty of others to choose from. When we skip to asking how a technology will disrupt some law, we don’t do the normative work of recognizing these variations and choices.
The law makes meaning of the social uses of technology—it constructs them. That’s not to say that’s all the law does. Law is construction with consequences. It conducts sensemaking through language, through institutions, through policy tools, and through motivating theories. Legal actors are not passive, even when they fail to enact new law. They have substantial agency in how AI is and will be legally constructed. A technological-exceptionalist approach to technology law can inaccurately characterize technology policymakers as unduly incapacitated and reactive. At the same time, technological exceptionalism can afford technology-law scholars and policymakers a degree of exceptionalism that detracts from what they can learn and sometimes transplant from practices across a wide array of other areas of law.
While these interactions can be hard to predict, as can politics, the legal construction of AI should not be considered neutral or objective. AI, like any technology, cannot be understood outside of its social context—and legal actors should fight for that approach, because it positions them to fight for technologies that further or protect certain democratic values. Legal scholars and technology law experts should not pretend to be caught off guard or fall into the disruption rhetoric. When legal actors, institutions, and ideals make meaning of technology they make paths for powerful democratic participation in technology. Without taking that path, we choose to follow technology.
Meg Leta Jones, Provost’s Distinguished Associate Professor in Communication, Culture, & Technology; Faculty Advisor of Institute for Technology Law & Policy; Core Faculty in Science, Technology, & International Affairs; and Faculty Fellow in Ethics Lab at Georgetown University.
Margot Kaminski, Professor of Law, Colorado Law School; Fulbright-Schuman Grantee 2023-2024; Fernand-Braudel Senior Fellow, European University Institute 2024; Affiliated Fellow, Information Society Project at Yale Law School; Faculty Associate, Harvard Berkman Klein Center 2023-2024.
Thanks to Nicholson Price, Paul Schwartz, and Ari Waldman and their extraordinary students for workshopping drafts of this Essay. Thanks also to Jack Balkin, participants in the Yale ISP Ideas Lunch, Blake Reid, the Rocky Mountain First Amendment Scholars workshop, and the We Robot and Privacy Law Scholars Conference communities for important and helpful feedback at various stages of these ideas.