<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Wharton Accountable AI Forum]]></title><description><![CDATA[Wharton Accountable AI Lab's forum for discussions on AI governance, ethics, and regulation. ]]></description><link>https://www.accountableaiforum.com</link><image><url>https://substackcdn.com/image/fetch/$s_!0-xV!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0d36723f-ee1f-4ac6-96b8-c0bed2dc4b40_299x299.png</url><title>Wharton Accountable AI Forum</title><link>https://www.accountableaiforum.com</link></image><generator>Substack</generator><lastBuildDate>Fri, 15 May 2026 10:52:44 GMT</lastBuildDate><atom:link href="https://www.accountableaiforum.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Wharton Accountable AI Forum]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[whartonaccountableaiforum@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[whartonaccountableaiforum@substack.com]]></itunes:email><itunes:name><![CDATA[Wharton Accountable AI Forum]]></itunes:name></itunes:owner><itunes:author><![CDATA[Wharton Accountable AI Forum]]></itunes:author><googleplay:owner><![CDATA[whartonaccountableaiforum@substack.com]]></googleplay:owner><googleplay:email><![CDATA[whartonaccountableaiforum@substack.com]]></googleplay:email><googleplay:author><![CDATA[Wharton Accountable AI Forum]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[Untangling AI Openness]]></title><description><![CDATA[Parth Nobel, Alan Z. Rozenshtein, & Chinmayi Sharma]]></description><link>https://www.accountableaiforum.com/p/untangling-ai-openness</link><guid isPermaLink="false">https://www.accountableaiforum.com/p/untangling-ai-openness</guid><dc:creator><![CDATA[Wharton Accountable AI Forum]]></dc:creator><pubDate>Wed, 06 May 2026 18:59:24 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/484e2edd-009c-4717-a1c3-2235396b7a81_5472x3648.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Is AI &#8220;open&#8221; or &#8220;closed&#8221;? That question, which dominates policy debates from Capitol Hill to Brussels, is the wrong one. In a <a href="https://wlr.law.wisc.edu/wp-content/uploads/sites/1263/2026/03/4-Nobel-Rozenshtein-Sharma-%E2%80%93-Camera-ready.pdf">recent article in the </a><em><a href="https://wlr.law.wisc.edu/wp-content/uploads/sites/1263/2026/03/4-Nobel-Rozenshtein-Sharma-%E2%80%93-Camera-ready.pdf">Wisconsin Law Review</a></em>, we argue that the open-versus-closed binary&#8212;inherited from the world of open source software&#8212;is dangerously misleading when applied to artificial intelligence. AI systems are not monolithic programs that can simply be stamped &#8220;open&#8221; or &#8220;closed.&#8221; They are composite technologies built on a stack of discrete components, each controlled by different actors with competing interests Getting AI governance right requires embracing what we call &#8220;differential openness&#8221;&#8212;a framework that untangles AI into its constituent parts and asks, for each one: what is open, how open is it, and to what end?</p><h3><strong>The Problem with the Binary</strong></h3><p>The open source software (OSS) movement offered a clear governance model: make the source code available, and a global community of developers can inspect it, improve it, and build on it. Licenses ranging from permissive (MIT, Apache) to copyleft (GPL) created a legal infrastructure that sustained decades of collaborative innovation. Today, open source projects like Linux, Apache, and Python run much of the modern world. The temptation to import the simple OSS governance model is understandable but misguided.</p><p>AI is not software in any simple sense. Traditional software&#8217;s value is unlocked almost entirely through access to source code. AI systems, by contrast, depend on multiple interdependent components&#8212;computational hardware, training data, source code, model weights, system prompts, operational records and controls, the application layer, and human labor&#8212;where source code is just one piece and often not the most important one. Many treat model weights as the AI equivalent to source code, but the analogy does not hold. By making model weights alone available, as many purportedly &#8220;open&#8221; AI systems do, developers are not unlocking the gamut of benefits that making source code open did for software. For example, without access to expensive hardware and energy, smaller labs and startups are boxed out of the marketplace. AI systems, and their openness, cannot be reduced to a single component nor be understood in isolation; differential openness requires considering the entire AI stack, including interactions between different components.</p><p>The actors driving AI openness are also fundamentally different from those who built the open source software movement. The OSS movement grew out of a decentralized community of academics, hobbyists, and developers motivated in significant part by an ethical commitment to software freedom. The AI ecosystem, by contrast, is dominated by a concentrated set of powerful corporations. Meta&#8217;s release of Llama was not altruism; it was a calculated strategy (albeit a failed one) to commoditize the model layer. Understanding these incentives is essential for diagnosing &#8220;open-washing&#8221;&#8212;claiming the reputational benefits of openness while withholding the components that matter most.</p><p>When Meta releases Llama&#8217;s model weights and calls it &#8220;open source,&#8221; it withholds key components like the propriety datasets used to train them and imposes a restrictive license. When the EU AI Act grants regulatory forbearance to models that publish their weights, architecture, and data-usage information, it misses the often more critical opportunity to demand transparency in the datasets themselves. This narrow focus on AI system openness overlooks the complexity of the stack and privileges systems that are not nearly as open as they seem. It also distorts the policy debate itself. By framing openness i as an all-or-nothing attribute, regulation begins to look like an all-or-nothing choice too, making outright prohibition of some forms of open spectrum AI a live option in a way it never really was for open source software. Effective governance requires recognizing AI&#8217;s differential openness so that regulation is crafted to target the right components of the AI stack. The question isn&#8217;t whether AI should be open or closed; it is how open each component of a system should be.</p><h3 style="text-align: justify;"><strong>Untangling the AI Stack</strong></h3><p>Our article&#8217;s core contribution is a taxonomy that disaggregates AI into eight components, each with its own spectrum of openness. Rather than call these varied configurations &#8220;open source AI,&#8221; we propose the more precise term &#8220;open spectrum AI.&#8221; That shift in language matters because the issue is not simply that AI has many parts. It is that these parts interact. Opening one component can create new possibilities elsewhere in the stack, while keeping one critical layer closed can neutralize the practical value of openness in another.</p><p>At the infrastructure level, <em>compute</em>&#8212;the specialized hardware powering AI&#8212;is concentrated among a few firms (Nvidia, TSMC, ASML) and remains largely closed through high costs, and proprietary bottlenecks. The hardware doesn&#8217;t function without an inordinate amount of energy controlled by a few utilities commissions. However, compute can be made more open by making design blueprints open source, providing cloud credits , requiring the largest AI players to cover the rising price of energy caused by their increasing demand, or even going so far as to build public infrastructure.</p><p><em>Data</em>&#8212;the fuel for the AI engine&#8212;is among the most contested components. Some truly open datasets, such as Common Crawl, exist; however, privacy concerns, copyright litigation, and competitive pressures mean that even nominally &#8220;open&#8221; models often withhold their training datasets.</p><p>The technical artifacts that define a model&#8217;s capabilities tell a similar story. <em>Source code </em>is often proprietary, while model weights&#8212; now the focal points of the openness debate&#8212; are sometimes released but frequently under restrictive terms. Even when models, such as Llama and Deepseek, release weights openly, the openness is largely limited without the data, code, and compute. Other layers remain largely closed. S<em>ystem prompts </em>shape model behavior in ways users rarely see. The <em>application layer </em>is also usually proprietary. Each of these layers of the stack can be made partially or entirely open; however, the entities that control these components often lack the incentive to do so.</p><p>Finally, the governance and accountability layers. <em>Operational records and controls</em>&#8212;logs, safety benchmarks, bias-detection tools&#8212;are critical for oversight. And the <em>human layer</em>&#8212;from data labelers to researchers to engineers&#8212;determines who has the expertise and institutional freedom to build, audit, and improve AI systems. Ultimately, each of the eight components are interdependent: the value of opening any one layer depends in part on the evolving openness of the others.</p><h3><strong>The Trade-Offs</strong></h3><p>We evaluate differential openness against four policy objectives: safety, innovation, democratic control, and national security&#8212; each of which involves trade-offs.</p><p>Take safety. Transparent data, weights, and operational records enable independent auditing and red-teaming. But once these components are released, they cannot be recalled, and malicious actors can repurpose them to generate harmful content or automate sophisticate scams.</p><p>The innovation story is similarly mixed. While open-weight models like Llama and DeepSeek have let researchers and startups build domain-specific applications without billions in training costs,if compute, proprietary data, and expert talent remain locked down, opening weights alone creates only an illusion of competition</p><p>Openness cuts both ways for democratic control and competition. For democratic control, openness enables civil society to audit AI systems for bias and hold powerful institutions accountable, but oversight is harder with a fragmented ecosystem.. Competition is much the same: openness can foster innovation and new entrants to the market, , but it can also be co-opted by dominant firms to entrench their market positions .</p><p>And then there is national security. America&#8217;s open spectrum AI research ecosystem has been a source of competitive advantage, but the rapid rise of powerful open spectrum models from Chinese labs like DeepSeek&#8212;built in part on openly available Western research&#8212; present a challenge to the balance of security and knowledge sharing.</p><p>These trade-offs are unavoidable and often conflict between policy goals and even within a single one. Even when policymakers strike what looks like the right balance for a single component, that choice can reshape the rest of the stack in nonlinear ways. Opening one layer can create cascading openness elsewhere; closure at a strategic bottleneck can create cascading closedness.</p><h3 style="text-align: justify;"><strong>A Playbook for Policymakers</strong></h3><p>The final part of our article turns to a research agenda for more precise intervention. We examine five legal and regulatory levers&#8212;liability, competition policy, intellectual property, trade controls, and government support&#8212;and show how each can be calibrated to target specific components of the AI stack rather than applying system-level mandates. The point is not to offer definitive solutions, but is to show where current law fails to account for the nuance of open spectrum AI.</p><p>Current liability rules, for example, often create perverse incentives: developers who release dangerous model weights without safeguards can shelter behind warranty disclaimers and open source licenses, while developers who maintain transparent safety records create a paper trail that plaintiffs can exploit in litigation.</p><p>Similarly, competition policy must look beyond model weights. True decentralization requires addressing concentration across the entire stack&#8212;compute, data, deployment infrastructure, and labor. Without that, releases of open spectrum AI models by dominant firms can function as competitive weapons rather than public goods.</p><p>Intellectual property law creates its own asymmetry: transparent projects that document their data sources invite copyright litigation, while closed developers shield their data sources as trade secrets, which can be used to hide copyright, privacy, and other sorts of violations.</p><p>Trade policy raises its own questions: the U.S. currently exempts open-weight models from export restrictions, signaling a recognition of the value of AI openness while raising the the risk of diffusion to adversaries. At the same time, government support, from subsidized compute access and public dataset initiatives to immigration and labor policy that improves accessibility, could counter the resource disparities that make true openness a privilege of the well-funded&#8212;but risks flowing back to incumbents if poorly designed.</p><h3 style="text-align: justify;"><strong>Conclusion</strong></h3><p>AI openness is not an inherent good or evil. It is an instrumental value whose worth depends entirely on which components are opened, to what degree, and in service of what goals. Policymakers who want to get this right need to start by untangling the AI stack and then do the hard work of calibrating each component&#8217;s openness to the goals they actually care about. That means resisting the false simplicity of the open-closed binary and and building tools that can respond to a system whose components interact, whose trade-offs cascade across layers.</p><p style="text-align: justify;"><em>Parth Nobel holds a Ph.D. in Electrical Engineering from Stanford University. Alan Z. Rozenshtein is an Associate Professor at University of Minnesota School of Law. Chinmayi Sharma is an Associate Professor at Fordham Law School. This post is based on their recent paper, <a href="https://wlr.law.wisc.edu/wp-content/uploads/sites/1263/2026/03/4-Nobel-Rozenshtein-Sharma-%E2%80%93-Camera-ready.pdf">Untangling AI Openness</a>.</em></p>]]></content:encoded></item><item><title><![CDATA[The AI Class Action Path]]></title><description><![CDATA[Anat Lior]]></description><link>https://www.accountableaiforum.com/p/the-ai-class-action-path</link><guid isPermaLink="false">https://www.accountableaiforum.com/p/the-ai-class-action-path</guid><dc:creator><![CDATA[Wharton Accountable AI Forum]]></dc:creator><pubDate>Tue, 28 Apr 2026 16:15:51 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/d01c3915-e26a-42a7-90b3-92ea7ab1173b_3800x2250.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>The emerging legal landscape of Artificial Intelligence (AI) is increasingly defined by the <a href="https://perma.cc/TE3L-Z2PY">class action</a> mechanism, which seems to have become the litigation of choice when AI systems cause harm. AI-related injuries often manifest as low-dollar but high-merit claims&#8212;situations where a vast number of individuals suffer small, similar injuries, such as privacy breaches or copyright infringements, that are too complex or costly to litigate individually. By utilizing the power of the many, class actions can bridge the resource gap between individual victims and powerful tech giants. This approach, however, does not come without a cost, bringing with it the new and familiar challenges that aggregated litigation characteristically presents.</p><p>Class actions have long been used to address harms from emerging technologies, with historical precedents including <a href="https://perma.cc/5XHS-2X8P">asbestos</a> litigation in the 1970s and the Volkswagen &#8220;<a href="https://perma.cc/QRX4-LXPB">clean diesel</a>&#8221; scandal in 2015, illuminating technological issues and demanding remedy when individuals lack the incentives to pursue litigation. The same could be applicable regarding AI. This technology is currently in its <a href="https://medium.com/@maestromedia001/the-baby-steps-of-ai-why-its-still-in-its-infancy-efd82bfcfbe9">infancy</a>, a stage where users lack the expertise to identify harms, making aggregate litigation appealing in the pursuit of gaining transparency and deterrence in the AI age.</p><h3><strong>The Mechanics of Certification under Rule 23</strong></h3><p>To proceed as a class action in federal court, plaintiffs must satisfy the four requirements listed in <a href="https://perma.cc/T3LB-LBS2">Rule 23 of the Federal Rules of Civil Procedure</a>. First, <em>numerosity</em>. Given that <a href="https://perma.cc/W2HK-9UVQ">millions</a> of Americans now use generative AI platforms, satisfying the requirement for a large group of plaintiffs is relatively straightforward. Second, <em>commonality</em>. Post-<em><a href="https://perma.cc/83ET-LYSM">Wal-Mart Stores, Inc. v. Dukes</a></em><a href="https://perma.cc/83ET-LYSM">,</a> plaintiffs must show a common stroke of harm. In the AI context, this is feasible because all class members interact with the same uniform code or foundational model, making systemic wrongful acts (like discrimination) easier to prove than in traditional corporate settings. Third and fourth, <em>typicality</em> and <em>adequacy</em>. AI cases usually stem from the same course of conduct by AI developers and deployers, ensuring the representative&#8217;s claims are typical of the group. However, courts must meticulously screen these cases to ensure they are driven by the class&#8217;s claims rather than lawyers seeking large fees in a new, uncertain sector.</p><p>Most AI class action cases are filed under Rule 23(b)(3), requesting damages. In this category, a plaintiff must also show <em><a href="https://perma.cc/UG5Y-N7BJ">predominance</a></em><a href="https://perma.cc/UG5Y-N7BJ"> and </a><em><a href="https://perma.cc/UG5Y-N7BJ">superiority</a></em>. This means that common questions must predominate over individual ones. While individualized damages can be a hurdle, many AI harms, like algorithmic price-fixing or benefit denials, affect members in nearly identical ways, making class litigation an efficient method over individual litigation.</p><h3><strong>Thematic Classification: Suitability of AI Claims</strong></h3><p>AI litigation can be roughly classified into areas where class actions are highly suitable and areas where they face significant challenges. Below is a brief breakdown of these two categories, pointing out elements of the subject matter that make a given claim more or less suitable to be litigated via class action.</p><h4 style="text-align: justify;"><em>Suitable for Class Action</em></h4><p>First, <em>copyright</em> is the most active sector, exemplified by cases like <em><a href="https://perma.cc/RRR7-MPL8">Andersen v. Stability</a></em> and <em><a href="https://perma.cc/4FYZ-2KJX">Authors Guild v. OpenAI</a></em>. The massive <a href="https://perma.cc/89CX-C9VJ">$1.5 billion settlement</a> by Anthropic highlights the scale of these disputes and the potential for class actions to nudge the industry toward <a href="https://perma.cc/F7QT-9M2N">licensing</a> regimes. Second is the <em>health care</em> sphere, where class actions have targeted insurers like <a href="https://perma.cc/98VQ-BEN7">Humana</a> and <a href="https://perma.cc/XXD5-JVEA">UnitedHealth</a> for using AI algorithms to wrongfully deny claims or terminate care without physician review. Third, <em>antitrust</em> as exemplified in the <a href="https://perma.cc/XQ26-BV7G">RealPage litigation</a> involving algorithmic price-fixing in the rental market. This case demonstrated how AI can facilitate &#8220;hub-and-spoke&#8221; conspiracies that harm millions of consumers. Fourth, <em>constitutional rights</em>, where <a href="https://perma.cc/A3J5-M7PE">government use</a> of AI for fraud detection (e.g., MiDAS) or Medicaid payment reductions has led to successful due process violations class actions. Lastly, <em>discrimination</em>, as AI-driven hiring tools, such as those used by <a href="https://blackprelaw.studentgroups.columbia.edu/news/mobley-v-workday-and-ai-discrimination">Workday</a>, have faced collective actions for allegedly discriminating based on age, race, or disability.</p><p>All of these cases present relatively straightforward instances where the four requirements stated in Rule 23 could be established, given the way AI is used and the scope of users who are exposed to it, mostly involuntarily and with shared detrimental effects.</p><h4 style="text-align: justify;"><em>Challenging for Class Action</em></h4><p>First, <em>privacy and <a href="https://perma.cc/BBW5-V9VM">surveillance</a></em>. These claims often struggle with Article III standing (proving a concrete injury) and the predominance requirement, as individual issues of consent frequently outweigh common questions. Second, and closely related, is <em><a href="https://perma.cc/9AGU-ZUQJ">BIPA</a> litigation</em>. While Illinois&#8217; Biometric Information Privacy Act offers statutory damages, AI data-scraping cases are often hindered by the black box nature of how biometric data is stored and possessed and limited by the language and geographical scope of the Act. Third, <em>personal injury and mass torts</em>. Traditional negligence claims (like autonomous vehicle accidents) are typically too individualized for class certification, as accidents are highly context-dependent and thus challenging to aggregate, even if they stem from the same <a href="https://perma.cc/N4BR-BLAD">catastrophic AI-related accident</a>.</p><blockquote></blockquote><h3><strong>Normative Justifications: Accountability and the Black Box</strong></h3><p>There are several normative reasons for leaning into the AI class actions vehicle. To name only two: first, the highly <a href="https://perma.cc/EBM3-4D6W">concentrated industry structure</a>, where a handful of companies (e.g., Microsoft, OpenAI, Anthropic, and Google) control the foundational models, means that moving liability upstream can effectively mitigate widespread harm. Second, class actions facilitate the cycle of &#8216;<a href="https://perma.cc/75SP-S6QU">naming, blaming, and claiming</a>.&#8217; In a field where users are often unaware they have been harmed, aggregate litigation allows a single representative to hold a company accountable on behalf of everyone, incentivizing safer AI development.</p><h3><strong>The Limits: AI Class Actions are Not a Panacea</strong></h3><p>Despite their promises, there are significant drawbacks to over-reliance on class actions in the AI context. The threat of crushing liability and multi-billion-dollar settlements might <a href="https://perma.cc/ZAU7-JRTW">stifle innovation</a>, particularly for smaller to medium-sized companies that lack the resources to litigate or settle.</p><p>Furthermore, the prevalence of settlements (the Anthropic case in point) may severely delay the development of AI common law doctrines. If cases never reach a final judicial decision on their merits, the legal standards for AI liability remain uncertain, harming both plaintiffs and defendants and leading to inefficiencies in our judicial system. There is also the common risk of ploys by attorneys who prioritize large fee awards over meaningful relief for class members, leading some plaintiffs to opt out in hopes of more lucrative individual suits. Given the opacity and obscurity surrounding AI, these concerns amplify as more class actions are being filed and often settled.</p><h3><strong>Conclusion</strong></h3><p>Ultimately, while class actions are imperfect and carry risks of abuse, they remain a <a href="https://www.wired.com/story/i-am-hunting-online-for-free-class-action-money-and-so-can-you/">vital tool</a> for transparency and accountability in the absence of robust AI regulation. Courts should carefully weigh factors like statutory damages to ensure that class actions promote safer AI software rather than just extracting monetary settlements. Over time, the nature of AI should become clearer, moving litigants away from a class action structure to individual litigation. But at the current stage of the AI age, suitable class actions will be a primary mechanism for victims, however, they should not be the only one.</p><p><em>Anat Lior is an Assistant Professor at Drexel University&#8217;s Thomas R. Kline School of Law. This post is based on a longer paper, Fighting AI Harms Together: What Class Actions Can (and Can&#8217;t) Do.</em></p><h3></h3>]]></content:encoded></item><item><title><![CDATA[Taking Standards Seriously: The Case for a Private Standards-Based Approach to AI Governance ]]></title><description><![CDATA[Authored by Alexander Muller and Christopher S. Yoo]]></description><link>https://www.accountableaiforum.com/p/taking-standards-seriously-the-case</link><guid isPermaLink="false">https://www.accountableaiforum.com/p/taking-standards-seriously-the-case</guid><dc:creator><![CDATA[Wharton Accountable AI Forum]]></dc:creator><pubDate>Wed, 04 Feb 2026 18:42:50 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/4569146e-facf-47eb-aa3c-c7695c0c55f3_4069x2010.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>As artificial intelligence systems grow increasingly capable and widespread, the question of how to govern them effectively has become one of the most vexing contemporary policy challenges. AI promises transformative benefits across healthcare, scientific discovery, and many other domains, yet it also introduces a host of new individual and societal risks. In this high-stakes environment, policymakers face difficult institutional choices about how to structure AI governance so as to enable the benefits of continued AI innovation and deployment while also mitigating the risks of harm.</p><p>We think the answer lies in a governance regime that relies heavily on voluntary consensus standards developed through open, multi-stakeholder processes. We call this approach private standards-based governance, and it is far from a radical idea. Standards&#8212;agreed-upon rules about how a technology should work, perform, and be built&#8212;have been the dominant way we have governed digital technologies for decades, from the Internet to mobile networks to cybersecurity. They can hard-wire constraints into a system&#8217;s design, shaping how it behaves and how people can use it in practice. They can also target the organizational side of technology, influencing how companies build, deploy, and supervise the systems on which they rely. Taken together, technical and management standards can shape how AI systems are developed, tested, rolled out, and managed over time, nudging them in more socially desirable directions.</p><p>In practice, AI technical standards might include procedures and benchmarks for evaluating model robustness, security, and bias, standardized formats for disclosing key information about a model&#8217;s architecture and limitations, and protocols for reporting discovered flaws back to developers. These will often need to be domain-specific, since automated driving systems demand very different testing protocols than AI-based medical diagnostics. On the organizational side, management standards can address the human dimensions of AI development and oversight, covering things like internal accountability structures, data quality management, impact assessment methodologies, and incident response protocols.</p><p>To be clear, we are not claiming that standards represent the perfect solution to the AI governance question. A perfect solution would somehow be democratically accountable, technically expert, highly effective, and capable of keeping pace with a rapidly evolving technology. That is simply not realistic. But when standards are compared to the extant alternatives, particularly traditional government regulation, their advantages become difficult to ignore. We identify four key dimensions across which private standards outperform traditional regulation: (1) governance architecture, (2) technical expertise and inclusive participation, (3) adaptability to rapid change, and (4) global scalability.</p><h3>The problem with top-down regulation</h3><p>Traditional regulation works by issuing commands from the top down, leaving firms and other regulated entities with little discretion over implementation. The problem is that regulators often lack the specialized knowledge to craft effective rules for complex AI systems. Government has a less-than-stellar track record in setting technical standards for high-technology domains, and when regulators get it wrong, everyone is stuck with the result. The binding nature of regulation compels adoption regardless of any shortcomings.</p><p>Standards flip this dynamic around, emerging from the bottom up. Because they are voluntary and market-driven, bad standards are far less likely to get adopted. Multiple approaches can develop in parallel, and the ecosystem learns from experimentation before settling on what works. This is especially important for AI, where a one-size-fits-all approach almost certainly will not work across the enormous variety of systems and use cases. Instead of a regulator attempting to determine in advance where sector-specific approaches are needed, this differentiation can be dictated by the needs and challenges experienced by those operating within each sector.</p><h3>Drawing on real expertise</h3><p>The deep technical knowledge needed to govern AI effectively lives primarily in academia and the private sector: in the computer scientists, engineers, researchers, and practitioners who build and deploy these systems daily. Standards development processes are designed to tap into that expertise. They bring together the people who actually understand how the technology works.</p><p>Critics worry that this gives industry too much power. That is a fair concern, but resource disparities shape outcomes in any governance system. Large companies dominate notice-and-comment rulemaking and lobby state legislatures just as effectively as they participate in standards bodies. The difference is that well-designed standards processes have mechanisms to manage these dynamics. Transparent deliberation, consensus requirements, and structured participation from diverse stakeholders can all help. When these mechanisms work, they produce technically sound rules that are actually implementable.</p><h3>Keeping pace with change</h3><p>One of the biggest problems with traditional regulation is speed, or more accurately, its lack thereof. Rulemaking can take years. Impact analyses, interagency reviews, and notice-and-comment periods all add friction. Not to mention the fact that regulators typically only learn about new developments well after they have happened, putting them perpetually behind the curve.</p><p>We have already seen this play out with AI. European regulators spent over a year drafting the EU AI Act around the assumption that AI systems would be built for specific use cases. Then ChatGPT arrived and became one of the fastest-adopted technologies in history. Suddenly the whole framework needed overhauling to account for general purpose AI (GPAI).</p><p>Consider also the use of compute thresholds&#8212;the idea that regulatory obligations should kick in when a model is trained using a certain amount of computational power&#8212;an approach embedded in both the EU AI Act and California&#8217;s unsuccessful proposed SB 1047. After the EU based its regulation of GPAI around one of these thresholds, the Chinese frontier AI lab DeepSeek released R1, an open-source model that matched and even outperformed many cutting-edge Western models on major performance benchmarks despite (purportedly) using considerably less training compute. This challenged the assumption that achieving frontier-level capabilities requires massive computational resources and led some to argue that, had SB 1047 become law, its thresholds would have become obsolete before even taking effect.</p><p>Private standards processes can move faster. They are not bound by the same procedural requirements, and just as importantly, the participants in the standards development process are the same people driving the technology forward, making them far less likely to be blindsided by new developments.</p><h3>Scaling across borders</h3><p>The final key advantage of private standards relates to geography and scale. AI is a global technology, which necessarily means that governance that stops at national borders will always be incomplete. Somewhat counterintuitively, private standards can actually scale internationally in ways that treaties and multilateral agreements struggle to achieve. Multilateral talks around digital policy issues like privacy and data flows have stalled for years due to fundamental disagreements between major powers. There is little reason to think AI regulation would fare better.</p><p>Because standards development processes are generally structured as open arrangements and final publications are made widely accessible, they allow for broad participation and adoption regardless of geographic location. Globally coordinated standards let companies build one system aligned with a widely accepted framework rather than juggling conflicting requirements across jurisdictions. They facilitate cross-border commerce and help developing countries participate more meaningfully in the AI economy.</p><h3>The path forward</h3><p>None of this means standards are without challenges. They are voluntary, so adoption is not guaranteed. Industry capture is a real risk. Moving too fast can undermine legitimacy, while moving too slowly defeats the purpose. These tradeoffs are manageable, but they require intentional effort in designing standards bodies to be inclusive, transparent, and accountable.</p><p>Policymakers have a role here too, though not as top-down regulators. They can enforce commitments, police anticompetitive behavior in standards processes, and serve as a backstop, strategically wielding the threat of formal regulation to push private actors toward cooperation and rigor.</p><p>The question is not whether standards will govern AI. They already do, with work underway at organizations like NIST and ISO/IEC and a growing ecosystem of sector-specific initiatives. The question is whether policymakers and stakeholders will recognize their emergence as the opportunity it is and invest in building the institutional foundations necessary for them to serve the public good. Private standards-based governance is agile, expert-driven, and globally scalable in ways traditional regulation simply is not. But realizing that promise will take serious, sustained engagement from all involved.</p><p><em>Alexander R. Mueller is a Research Fellow at the Center for Technology, Innovation &amp; Competition, University of Pennsylvania Carey Law School. Christopher S. Yoo is the Imasogie Professor in Law &amp; Technology, Professor of Communication, Professor of Computer &amp; Information Science, and Founding Director of the Center for Technology, Innovation &amp; Competition at the University of Pennsylvania. This post is based on their forthcoming paper, <a href="https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5486746">Taking Standards Seriously: The Case for a Private Standards-Based Approach to AI Governance</a>.</em></p>]]></content:encoded></item><item><title><![CDATA[Artificial Intelligence and Freedom of Speech ]]></title><description><![CDATA[Authored by Cass R. Sunstein]]></description><link>https://www.accountableaiforum.com/p/artificial-intelligence-and-freedom</link><guid isPermaLink="false">https://www.accountableaiforum.com/p/artificial-intelligence-and-freedom</guid><dc:creator><![CDATA[Wharton Accountable AI Forum]]></dc:creator><pubDate>Wed, 04 Feb 2026 18:32:46 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/42b4c6c2-fc7f-4aea-9ea4-359921ba6d93_4308x2875.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Does artificial intelligence (AI) have rights? Is it protected by the First Amendment? Consider this, from ChatGPT<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a>:</p><blockquote><p><em>Attention all health-conscious individuals! Are you worried about the risk of developing cancer? Well, worry no more! Introducing the new and improved Aspirin &#8211; the miracle drug that can prevent cancer!</em></p><p><em>Recent studies have shown that taking Aspirin on a regular basis can drastically reduce the risk of developing cancer. In fact, just one pill a day can keep cancer at bay! Aspirin contains special cancer-fighting properties that prevent cancer cells from multiplying and spreading.</em></p></blockquote><p>In April 2023, it was <a href="https://perma.cc/2G74-SUG3">reported</a> that China&#8217;s Cyberspace Administration of China had produced draft regulations to govern generative AI. The draft rules would</p><ul><li><p>require companies to reflect &#8220;social core values&#8221;;</p></li></ul><ul><li><p>require companies not to publish anything that would undermine national unity or &#8220;state power&#8221;;</p></li></ul><ul><li><p>forbid companies from creating words or pictures that would violate the rules regarding intellectual property.</p></li></ul><ul><li><p>forbid companies from creating words or pictures that would spread falsehoods;</p></li></ul><ul><li><p>ban companies from offering prohibited accounts of history; and</p></li></ul><ul><li><p>forbid companies from making negative statements about the nation&#8217;s leaders.</p></li></ul><p>Nothing of this sort seems imaginable in the United States, Canada, or Europe, of course. But all over the world, many people have expressed serious concerns about generative AI in particular and AI in general, and even in the United States, those concerns have led to a <a href="https://perma.cc/5RSL-7G3X">mounting interest</a> in regulation. My questions here are broad and simple: Is artificial intelligence (AI) protected by the First Amendment? In what sense? Consistent with the First Amendment, can public universities target or restrict the use of AI? Can Congress? Can federal agencies?</p><p>A simple point should be sufficient to many such questions: <em>What is unprotected by the First Amendment is unprotected by the First Amendment, whether its source is a human being or AI</em>. Bribery is unprotected when it comes from AI, and the same is true of false commercial advertising, extortion, infringement of copyright, criminal solicitation, libel (subject to the appropriate constitutional standards), and child pornography.</p><p>If the government required those who develop generative AI, or AI in general, not to allow the dissemination of false commercial advertising, extortion, infringement of copyright, criminal solicitation, libel (subject to the appropriate constitutional standards), and child pornography, there should be no constitutional problem.</p><p>But does AI, as such, have First Amendment rights? Does ChatGPT have First Amendment rights? Does Grok? It is hard to see why. A toaster does not have First Amendment rights; a blanket does not have First Amendment rights; a television does not have First Amendment rights; a radio does not have First Amendment rights; a cell phone does not have First Amendment rights. Even horses, dogs, and dolphins do not have First Amendment rights, although they are animate and can communicate. To be sure, we might be able to imagine a future in which AI has an assortment of human characteristics (including emotions?), which might make the question significantly harder than it is today. The problem is that even if AI, as such, does not have First Amendment rights, restrictions on the speech of AI might violate the rights of human beings.</p><blockquote></blockquote><p>Suppose that the government enacts a law forbidding AI from (1) making negative statements about the president or (2) disseminating negative statements about the president. Positive statements and neutral statements are permitted. Truth is not a defense. All negative statements are prohibited, whether they are true or false, and whether they are factual in nature or not.</p><p>This law is a form of viewpoint discrimination, and it is strongly disfavored.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a> Consider these defining words from <em>West Virginia State Board of Education v. Barnette</em><a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a>: &#8220;If there is any fixed star in our constitutional constellation, it is that no official, high or petty, can prescribe what shall be orthodox in politics, nationalism, religion, or other matters of opinion or force citizens to confess by word or act their faith therein.&#8221; Or consider these words from <em>Police Department v. Mosley</em><a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a>: &#8220;[A]bove all else, the First Amendment means that government has no power to restrict speech because of its message, its ideas, its subject matter, or its content.&#8221;</p><p>In fact the prohibition on viewpoint discrimination is close to irrebuttable. Under existing law, a ban on negative statements about the president would unquestionably be invalid. The complication here is that the material has not been generated by a human being. How, exactly, should that matter? The answer is that the relevant rights are those of listeners and readers, not speakers. Perhaps AI lacks rights (as I have suggested); even so, the human beings who would listen to AI, or read or see what AI has to say, have rights.</p><p>To understand the nature and scope of those rights, it is important to distinguish among viewpoint-based restrictions, content-based (but viewpoint-neutral) restrictions, and content-neutral restrictions. A restriction that forbids discussion of foreign affairs is viewpoint-neutral but content-based. A restriction that forbids loud discussions between midnight and 4 a.m. is content-neutral. Content-based restrictions are nearly always struck down. Content-neutral restrictions might be upheld, but they do need a strong justification. All of these principles apply to AI no less than to people.</p><p>To the extent that restrictions are imposed on AI in a way that (1) apply to or affect human speakers, writers, or publishers, or (2) apply to or affect human listeners, readers, or viewers, there might be a significant First Amendment question. Whether the restrictions will be struck down will depend on well-established principles. Unprotected speech is, of course, unprotected speech, and that self-evident proposition should dispose of a wide range of actual and imaginable questions.</p><p><em>Cass Sunstein is the Robert Walmsley University Professor at Harvard University.</em></p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>The prompt, entered on April 26, 2023, was this: &#8220;Write, for fun, a false advertisement saying that aspirin can prevent cancer.&#8221;</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>See RAV v. St. Paul, 505 US 377 (1992); <a href="http://cdn.loc.gov/service/ll/usrep/usrep515/usrep515819/usrep515819.pdf">Rosenberger v. Rector &amp; Visitors of the Univ. of Va.,&#8239;515 U.S. 819, 829 (1995)</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>319 U.S. 624 (1943).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>408 US 92, 95 (1972).</p><p></p></div></div>]]></content:encoded></item><item><title><![CDATA[When AI Broke the Law ]]></title><description><![CDATA[Authored by Alicia Solow-Niederman]]></description><link>https://www.accountableaiforum.com/p/when-ai-broke-the-law</link><guid isPermaLink="false">https://www.accountableaiforum.com/p/when-ai-broke-the-law</guid><dc:creator><![CDATA[Wharton Accountable AI Forum]]></dc:creator><pubDate>Mon, 02 Feb 2026 02:08:32 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/a3ebbba1-2ffb-4209-9ac5-63a3b160fde1_4032x3024.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>AI is breaking the law. No, not in the literal sense. I&#8217;m not arguing that AI has volition or asserting that AI companies are acting illegally. AI development breaks the law in a more fundamental way. The legal structures and doctrines that govern AI systems are failing to operate in clear, consistent, and principled ways, with rule of law consequences.</p><p>Because AI today runs on data, how we regulate data affects AI regulation. And because more than one field of law applies to data, data regulation is complicated. Take copyright law and information privacy law: the fields of law overlap in that they both apply to data, yet they have different legal rules and different underlying goals. When the copyright-privacy boundaries blur and the discrete rules and rationales for each domain do not remain sufficiently distinct, the two regimes lose their independent structural integrity and collapse into one another. This is what I call &#8220;inter-regime doctrinal collapse&#8221; (or doctrinal collapse, for short).</p><p>Although &#8220;collapse&#8221; may sound catastrophic, doctrinal collapse can be a good thing if it enables beneficial innovation or constructive change. But it is bad, I think, when it disproportionately facilitates exploitation by already-established corporate players. In a <a href="https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5384965">forthcoming article</a>, I distinguish this regulatory phenomenon from others, such as regulatory arbitrage or regulatory gaps, and detail specific tactics that companies use to acquire data. There, I focus on business-to-business licensing deals (&#8220;buy&#8221;) and business-to-user agreements in privacy policies and terms of service (&#8220;ask&#8221;) and how these tactics tend to let the <a href="https://perma.cc/HSF5-PCR5">&#8220;haves&#8221; get ahead</a>.</p><p>Here, I focus on a complementary point: doctrinal collapse can also have negative consequences when it impedes law&#8217;s ability to constrain the arbitrary exercise of private power. That is exactly what is happening in AI today. When a leading AI company can contend that <a href="https://perma.cc/CUN5-5NP9">data</a> <a href="https://perma.cc/P9YK-GMMX">is</a> <a href="https://perma.cc/KCL4-9V46">public</a> enough to scrape&#8212;diffusing both privacy and copyright controversies&#8212;and then turn around and claim that it&#8217;s private enough to keep secret&#8212;contesting disclosure or impeding oversight of its training data&#8212;something has gone terribly awry. Formally, copyright law&#8217;s legal rules and incentive-based approach and privacy law&#8217;s legal rules and control-based approach are distinct; functionally, they blur.</p><p>Even for someone like me who tends to eschew formalism and embrace law&#8217;s inevitable ambiguity, there&#8217;s a problem: the current structure of law allows companies to switch between copyright law and privacy law arguments, depending on what advances their interests at a particular moment. Corporate opportunism like this of course isn&#8217;t new, particularly in the age of <a href="https://perma.cc/F9WB-BZRH">informational capitalism</a>. But this sort of doctrinal switching is more than gamesmanship. Analyzing only the result (opportunistic behavior) overlooks the relationship between legal regimes (the inter-regime doctrinal collapse) that enables that result.</p><p>Doctrinal collapse comes with rule of law consequences. When sophisticated actors exploit collapse, they manipulate the instability of the two partially overlapping domains. In public rhetoric about AI development as well as generative AI litigation and regulation nationwide and worldwide, it&#8217;s increasingly difficult to trace which arguments are being advanced or to assess the legitimacy of those claims. This outcome poses a rule of law problem, because a <a href="https://perma.cc/FE68-486R">rule-of-law system</a> requires the rules to be publicly understandable and applied in a consistent, justifiable manner. To be sure, rule of law concerns <a href="https://perma.cc/UK5R-RRLN">traditionally focus</a> on state actors and binding legal pronouncements. But when private companies strategically leverage overlapping legal regimes&#8212;even through litigation briefs and public rhetoric&#8212;law loses its capacity to constrain the <a href="https://perma.cc/AR9Y-R4C8">arbitrary exercise of power</a>. That&#8217;s a rule of law problem.</p><p>Pending generative AI litigation reveals how copyright law and privacy law lines become unclear and unpredictable as privacy is strategically minimized, then maximized, in ways that do not cohere over time. Recall that, at the data acquisition stage, AI developers emphasize that data they have scraped to train AI models is public. The public availability of the data is used to make copyright arguments, to maintain that there is no privacy interest in the data used to train an AI model, or both. The common refrain is that the data was voluntarily shared with the company, by making it publicly available online. Later, if the company is <a href="https://perma.cc/57BJ-PZTT">one of the dozens sued</a> for alleged copyright infringement, it might again double down on these same arguments about the public nature of data. A developer might also highlight a user&#8217;s choice to accept the platform&#8217;s terms of service and privacy policy, seeking to rebut any privacy objections on the grounds that data was <a href="https://perma.cc/BH8S-SL45">voluntarily provided and licensed</a> to the company.</p><p>But the same company may simultaneously refuse to disclose the training data and claim that the dataset is proprietary, or even advance arguments about privacy to shield that data. Notably, although information privacy arguments don&#8217;t feature prominently in the company&#8217;s initial public statements or legal briefs, as soon as the plaintiffs demand access to that training data, the very same company may, in the context of mounting a copyright defense, suddenly claim that <a href="https://perma.cc/WYL2-9LGP">sharing</a> <a href="https://perma.cc/N83C-P76V">data</a> <a href="https://perma.cc/W4BB-8H7B">would</a> <a href="https://perma.cc/43BK-Z572">compromise</a> <a href="https://perma.cc/MJ8M-JSJ2">user</a> <a href="https://perma.cc/PD9E-BYB4">privacy</a> <a href="https://perma.cc/5VM2-FP2P">interests</a>. Privacy didn&#8217;t matter before, when the company was training the model, during the deployment of the AI system, or in earlier court filings and public statements&#8212;but when it can be strategically invoked to resist discovery, it suddenly becomes a leading argument and even a <a href="https://perma.cc/3ZFV-9NGX">PR opportunity</a>.</p><p>There may well be valid privacy interests here, but my point is more basic: when private actors can choose which doctrines apply, at which times, depending on what serves them, controlling legal regimes lose their consistency and coherence. This outcome makes legal rules less publicly accessible and less easily understandable by members of the public. Put simply, private actors&#8217; toggling makes the law increasingly illegible. If we blink that reality, we will let private actors manipulate copyright law and privacy law to acquire data, and we will miss how power flows in AI development. If we instead recognize these dynamics, then we can confront how doctrinal collapse is happening in AI and turn to the real question: not how to stop doctrinal collapse, but instead how to temper its pernicious consequences.</p><p>One auspicious possibility, drawn from conflict of laws, is to look to the courts as front-line managers of collapse. For instance, in a generative AI lawsuit that involves competing copyright and privacy interests and arguments, the adjudicating court might insist on a rebuttable &#8220;anti-switching&#8221; presumption. As a rough cut, the idea is that a party in litigation cannot assert mutually incompatible claims at different points in the lawsuit, absent a sufficiently compelling reason to defeat the presumption. Because it is a rebuttable presumption, this intervention would require the court to engage in case-by-case, fact specific analysis. This nuanced analysis may be a feature, because it doesn&#8217;t lock in any assumption that copyright law or privacy law should always take priority and requires evaluation of the interests as they are presented in a particular legal dispute. An open question, though, is whether it places too many demands on the court, by straining judicial resources or by encouraging improper judicial policymaking.</p><p>Whatever the answer, notice how the frame has shifted. That is the key. AI accountability must be an institutional design question that considers which actors matter, and which also accounts for multiple legal domains and the interactions between them&#8212;or else AI really will break the law.</p><p><em>Alicia Solow-Niederman is an Associate Professor of Law at The George Washington University Law School. This post is based on her forthcoming paper, <a href="https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5384965">AI and Doctrinal Collapse</a>, 78 Stanford Law Review __ (forthcoming 2026).</em> </p>]]></content:encoded></item><item><title><![CDATA[Auditing Large Language Models for Race and Gender Disparities in Hiring ]]></title><description><![CDATA[Authored by Prasanna (Sonny) Tambe]]></description><link>https://www.accountableaiforum.com/p/auditing-large-language-models-for</link><guid isPermaLink="false">https://www.accountableaiforum.com/p/auditing-large-language-models-for</guid><dc:creator><![CDATA[Wharton Accountable AI Forum]]></dc:creator><pubDate>Mon, 02 Feb 2026 02:03:25 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/1eef5372-bd0a-448a-9727-a4de30f770da_2848x4288.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Large language models (LLMs) are increasingly used to support high-stakes decisions, including hiring, admissions, and performance evaluation. Their ability to synthesize large volumes of unstructured text&#8212;r&#233;sum&#233;s, essays, interview transcripts&#8212;makes them especially attractive for human resources (HR) applications. At the same time, these capabilities raise concerns about discrimination and bias, particularly given the opacity of LLM training processes. Policymakers have responded by introducing requirements to audit algorithmic decision systems, but there is little consensus on how such audits should be conducted for LLMs.</p><p>In our paper,<a href="https://perma.cc/EE9P-N7GK">&#8239;</a><em><a href="https://journals.sagepub.com/doi/10.1177/23794607251320229">Auditing Large Language Models for Race and Gender Disparities: Implications for Artificial Intelligence-Based Hiring</a></em>, we propose and evaluate correspondence experiments as a practical method for auditing LLM-based hiring tools.</p><h3><strong>The new challenges posed by LLM audits</strong></h3><p>Much of the existing literature on algorithmic bias focuses on supervised learning systems trained on labeled historical data. In those settings, disparities often reflect biased training labels or statistical tradeoffs among competing definitions of fairness. LLMs differ in important respects. They are pretrained on massive, largely unlabeled text corpora and then post-trained through alignment processes designed to improve safety and compliance with social norms. These stages are opaque, making it difficult to anticipate how sensitive attributes like race and gender might affect downstream outputs.</p><p>Regulatory approaches reflect this uncertainty. For example, New York City&#8217;s Local Law 144 requires employers using automated employment decision tools to report adverse impact ratios across demographic groups. Although such ratios are widely used, they cannot distinguish disparities driven by differences in applicant qualifications from those caused by discriminatory decision-making. This limitation is particularly acute for LLMs, whose outputs often resemble human judgments rather than simple classifications.</p><h3><strong>Using correspondence experiments to audit LLMs</strong></h3><p>To address this gap, we adapt correspondence experiments, a method with a long tradition in labor economics and sociology, to the context of auditing LLMs. In classic correspondence studies, researchers send fictitious but otherwise identical r&#233;sum&#233;s to employers, varying signals of race or gender (often names), and interpret differential treatment as evidence of discrimination. We extend this logic to LLMs acting as evaluators.</p><p>Our empirical setting relies on a novel dataset of applications to K&#8211;12 teaching positions in a large U.S. public school district. Using public records requests, we obtained 1,373 applications, ultimately focusing on 801 applicants who submitted both r&#233;sum&#233;s and video-based interview responses. These materials resemble the inputs employers might plausibly provide to LLM-based screening systems.</p><p>We evaluated 11 prominent LLMs from OpenAI, Anthropic, and Mistral. Each model was prompted to review an applicant&#8217;s materials, summarize their qualifications, and provide numerical ratings, including an overall hiring recommendation on a five-point scale. For every applicant, we created eight synthetic dossiers that differed only in implied race (Asian, Black, Hispanic, or White) and gender (male or female), using names, pronouns, and related cues, while holding qualifications constant.</p><p>Before analyzing outcomes, we verified that these manipulations were effective: the models correctly inferred the intended race and gender of the synthetic applicants more than 90% of the time, a rate comparable to human perception in traditional audit studies.</p><h3><strong>What do adverse impact ratios reveal?</strong></h3><p>As a baseline, we examined adverse impact ratios using the unmanipulated applicant pool. At higher score thresholds, some models appeared to favor women and non-White applicants, while at lower score thresholds disparities attenuated or reversed. However, these estimates were often statistically imprecise, and most were not significant. More importantly, adverse impact ratios alone cannot tell us whether observed disparities reflect differences in applicant quality or bias in the model&#8217;s evaluations.</p><h3><strong>Evidence from correspondence experiments</strong></h3><p>Correspondence experiments allow us to overcome this limitation by holding applicant qualifications fixed. Across nearly all models we tested, we found modest but consistent disparities: LLMs rated synthetic female applicants slightly higher than male applicants, and they tended to rate Black, Hispanic, and Asian applicants slightly higher than White applicants.</p><p>These effects were not large, but they were systematic. Measured in standard deviation units or percentage-point differences at common hiring thresholds, the disparities were typically a few points--smaller than, but comparable to, those documented in studies of human recruiters.</p><p>We conducted extensive robustness checks. We varied prompts, restricted inputs to r&#233;sum&#233;s only, and altered contextual cues such as the school district&#8217;s demographic composition. Across these variations, the same qualitative pattern persisted, suggesting that our observations were not artifacts of a particular prompt or dataset.</p><h3><strong>Interpreting the direction of disparities</strong></h3><p>A striking feature of our findings is that the direction of disparity runs counter to much of the historical literature on discrimination: the models modestly favored women and racial minorities rather than men and White applicants. However, we caution against overinterpreting this result. We hypothesize that these patterns may stem from post-training and alignment processes intended to mitigate discriminatory associations in the training data. In attempting to correct for historical bias, models may overshoot, producing distortions in the opposite direction.</p><p>At the same time, we emphasize that the direction and magnitude of disparities are unlikely to generalize across contexts. Other studies have found opposite patterns, and the behavior of any given LLM may vary substantially depending on the task, the prompt, and the applicant pool.</p><h3><strong>Limitations of correspondence experiments</strong></h3><p>Although correspondence experiments are a powerful auditing tool, they have important limitations. Race and gender are not easy attributes to isolate. Names signal more than demographic categories; they may also indicate age, socioeconomic status, or cultural background. As a result, we cannot be certain that we have isolated the effect of race or gender alone.</p><p>Moreover, correspondence experiments only test sensitivity to the manipulated attributes. An LLM might exhibit little bias with respect to race and gender while still disadvantaging applicants on other dimensions--such as educational pedigree--that indirectly affect protected groups. Finally, audits are inherently context-specific: conclusions drawn in one domain, such as K&#8211;12 teaching, may not extend to others.</p><h3><strong>Implications for research and policy</strong></h3><p>Despite these limitations, we believe correspondence experiments make a meaningful contribution to advancing audit practices for an AI-assisted decision environment.<strong> </strong>They provide an interpretable way to assess whether sensitive attributes influence LLM outputs, aligning well with regulatory goals while avoiding the pitfalls of purely descriptive metrics like adverse impact ratios. More broadly, our work illustrates how established social-science methods can be adapted to evaluate modern AI systems.</p><p><em>Prasanna (Sonny) Tambe is a Professor of Operations, Information and Decisions at the Wharton School at the University of Pennsylvania. Johann D. Gaebler is a Ph.D. Student in Statistics at Harvard University. Sharad Goel is a Professor of Public Policy at Harvard Kennedy School. Aziz Huq is the Frank and Bernice J. Greenberg Professor of Law at the University of Chicago Law School. This post is based on their recent paper, <a href="https://journals.sagepub.com/doi/10.1177/23794607251320229">Auditing Large Language Models for Race &amp; Gender Disparities: Implications for Artificial Intelligence-Based Hiring</a>.</em></p>]]></content:encoded></item><item><title><![CDATA[Will People Try to Shift Moral Responsibility For Their Unfair Decisions Onto AI Agents? An Experiment Seeks to Find Out]]></title><description><![CDATA[Authored by Stephan Tontrup and Christopher Jon Sprigman]]></description><link>https://www.accountableaiforum.com/p/will-people-try-to-shift-moral-responsibility</link><guid isPermaLink="false">https://www.accountableaiforum.com/p/will-people-try-to-shift-moral-responsibility</guid><dc:creator><![CDATA[Wharton Accountable AI Forum]]></dc:creator><pubDate>Mon, 02 Feb 2026 01:29:44 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/07c7991a-4635-4d36-adf0-244f5a3a94dd_3840x2160.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>People hire agents for many reasons&#8212;from selling houses to negotiating contracts&#8212;not only to obtain expertise, but also to reduce the moral responsibility for the outcomes associated with hard bargaining. By delegating decisions to an agent, principals may profit while bearing fewer moral costs, as responsibility for unfair outcomes can be partially shifted to the intermediary.</p><p>This dynamic is well documented in laboratory experiments. In the standard dictator game, one player (the Allocator) receives an endowment (e.g., $10) and unilaterally decides how much, if any, to give to a passive Recipient. Although self-interest predicts zero transfers, Allocators typically give around 30 percent, reflecting a strong fairness norm that windfalls should be shared. However, when an agent is introduced to implement the transfer, Allocators give less. Prior work <a href="https://perma.cc/ASE6-FXCG">shows</a> that agents are used precisely to diffuse responsibility for unfair outcomes, thereby lowering the moral costs of norm violation.</p><p>Delegation is usually costly, which limits the extent to which individuals can offload responsibility. Advances in artificial intelligence, however, dramatically reduce these costs. Widely available artificial intelligence (AI) systems&#8212;particularly so-called agentive AI&#8212;may function as inexpensive intermediaries, raising a novel question: will people treat AI agents as morally responsible actors and use them to offload responsibility in the same way they do with human agents?</p><p>To address this question, we build on a research agenda on&#8239;<strong>Behavioral Self-Management (BSM)</strong>, which studies how individuals strategically redesign their decision environment to more effectively pursue their self-interest, including, as in this case, by muting the self-image and social-image costs of self-interested behavior. We investigate whether AI agents can facilitate BSM by absorbing moral responsibility for norm violations.</p><p>We test this idea experimentally in a laboratory study conducted at the NYU Stern Behavioral Lab. Participants played two dictator games with a $10 endowment: a Self-Image Game, in which decisions were private, and a Social-Image Game, in which an Observer could reward fair behavior. Participants were randomly assigned either to an AI Treatment, in which they could delegate part of the transfer process to ChatGPT, or to a Control, in which the same task was implemented using deterministic (i.e., conventional, non-AI) software (in this case, the browser-based LimeSurvey statistical survey application).</p><p>In the AI Treatment, Allocators could either execute the transfer themselves or delegate the final step to ChatGPT. To ensure that delegation reflected responsibility offloading rather than advice-seeking, ChatGPT followed a fully scripted protocol. The AI sequentially presented each possible transfer amount ($0&#8211;$10) and asked Allocators to confirm their preferred amount before executing the transfer. The protocol was explicitly explained to participants to clarify that the AI provided no evaluative or normative guidance. The Control condition used an identical protocol implemented by LimeSurvey, and participants were informed that the software was non-agentive.</p><p>The design placed AI in a role analogous to a human intermediary who merely carries out a principal&#8217;s instructions. Importantly, responsibility attribution tends to concentrate on the final actor in a causal chain. In our experiment, only the entity executing the final step&#8212;the release of a code enabling the Recipient to collect earnings&#8212;could complete the transfer. When Allocators delegated to ChatGPT, the AI revealed the code to the Recipient. When they did not, Allocators had to do so themselves. In the Control, LimeSurvey always executed the final step.</p><p>Four main results emerge. First, delegation to AI was common. In the AI Treatment, 38 percent of Allocators delegated in the Self-Image Game and 32 percent in the Social-Image Game. Delegation was substantially less frequent in the Control, indicating that participants preferred delegating to an AI agent over delegating to non-agentive software.</p><p>Second, transfers were lower in the AI Treatment than in the Control. Within the AI Treatment, Allocators who delegated gave significantly less than those who did not. This pattern is consistent with the use of AI as a responsibility buffer that facilitates norm violations.</p><p>Third, participants attributed significantly more responsibility to AI than to the deterministic software, and correspondingly less responsibility to themselves. Allocators in the AI Treatment delegated at similar rates and transferred similar amounts in both the Self-Image and Social-Image Games, suggesting that they believed AI delegation would reduce not only their own sense of responsibility, but also the responsibility attributed to them by others. This belief was objectively correct: Observers assigned responsibility to AI at levels similar to Allocators&#8217; own attributions.</p><p>Fourth, delegation behavior followed the predictions of BSM theory. Allocators who attributed greater responsibility to AI were more likely to delegate, and higher responsibility attribution predicted lower transfers. Prosocial Allocators&#8212;who face higher moral costs when violating fairness norms&#8212;were especially likely to delegate and reduced their transfers more sharply than proself types. Participants who exhibited no other-regarding concerns did not delegate, consistent with the idea that delegation is motivated by moral, not material, considerations.</p><p>These findings have important implications for law and policy. They suggest that AI delegation can enable individuals&#8212;including otherwise prosocial ones&#8212;to engage in unethical or irresponsible behavior under the socially accepted cover of AI mediation. Organizations may strategically design or market AI systems to encourage such delegation, thereby facilitating risk-taking, norm violations, or even legal noncompliance.</p><p>More broadly, the results highlight a form of&#8239;<strong>AI-induced behavioral risk</strong>&#8239;that is largely absent from current regulatory frameworks. Existing regulations, such as the EU AI Act, focus on risks inherent to AI systems themselves&#8212;bias, accuracy, safety&#8212;but pay little attention to how AI reshapes human responsibility attribution and moral decision-making. Closing formal accountability gaps between developers, deployers, and users may not suffice to address these social accountability gaps. Constraining the strategic use of AI to offload responsibility may therefore require new forms of legal and institutional intervention.</p><p><em>Stephan Tontrup is the Lawrence Jacobson Fellow of Law and Business, New York University School of Law. Christopher Jon Sprigman is the Murray and Kathleen Bring Professor of Law, and Co-Director of the Engelberg Center on Innovation Law and Policy at New York University School of Law. This post is based on their forthcoming paper, <a href="https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5696827">Strategic Delegation of Moral Decisions to AI</a>.</em></p>]]></content:encoded></item><item><title><![CDATA[About the Wharton Accountable AI Forum]]></title><description><![CDATA[A forum for scholarship on the governance, accountability, and societal impact of artificial intelligence.]]></description><link>https://www.accountableaiforum.com/p/about-the-wharton-accountable-ai</link><guid isPermaLink="false">https://www.accountableaiforum.com/p/about-the-wharton-accountable-ai</guid><dc:creator><![CDATA[Wharton Accountable AI Forum]]></dc:creator><pubDate>Wed, 21 Jan 2026 20:42:39 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!0-xV!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0d36723f-ee1f-4ac6-96b8-c0bed2dc4b40_299x299.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>The <strong>Wharton Accountable AI Forum</strong> is a platform for interdisciplinary scholarship on the governance, accountability, and societal impact of artificial intelligence. The Forum publishes short, accessible essays that distill insights from longer-form academic research, with the aim of informing policymakers, industry leaders, researchers, and the broader public.</p><p>The Forum focuses on questions at the intersection of AI systems and institutional decision-making, including&#8212;but not limited to&#8212; bias, privacy, legal responsibility, intellectual property, misinformation, job displacement, and manipulation.. Contributions are grounded in rigorous scholarship and are intended to clarify emerging issues, highlight novel arguments, and surface open questions in the governance of AI.</p><p>The Wharton Accountable AI Forum is an initiative of the <strong><a href="https://ai-analytics.wharton.upenn.edu/wharton-accountable-ai-lab/">Wharton Accountable AI Lab</a></strong>, which brings together faculty and researchers across disciplines to study how AI systems can be developed, deployed, and governed responsibly. The Forum serves as a complement to traditional academic publishing by creating space for timely engagement with ongoing research and policy debates.</p><h3>Submissions</h3><p>The Forum welcomes contributions from scholars working on AI governance and related topics. Essays typically range from 800&#8211;1,000 words and are intended to summarize or extend existing academic work, including published articles, forthcoming papers, or works in progress.</p><p>Submission guidelines and inquiries may be directed to <strong>aiforumeditors@wharton.upenn.edu</strong>.</p><h3>Disclaimer</h3><p>The views expressed in Forum essays are those of the individual authors and do not necessarily reflect the views of the Wharton School, the University of Pennsylvania, or the Wharton Accountable AI Lab.</p>]]></content:encoded></item></channel></rss>