<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	>

<channel>
	<title>May 2026 &#8211; Communications of the ACM</title>
	<atom:link href="https://cacm.acm.org/issue/latest/feed" rel="self" type="application/rss+xml" />
	<link>https://cacm.acm.org</link>
	<description></description>
	<lastBuildDate>Fri, 24 Apr 2026 16:36:10 +0000</lastBuildDate>
	<language>en-US</language>
	<sy:updatePeriod>
	hourly	</sy:updatePeriod>
	<sy:updateFrequency>
	1	</sy:updateFrequency>
	<generator>https://wordpress.org/?v=6.8.5</generator>

<image>
	<url>https://cacm.acm.org/wp-content/uploads/2023/11/cropped-cropped-cacm_favicon-1.png?w=32</url>
	<title>May 2026 &#8211; Communications of the ACM</title>
	<link>https://cacm.acm.org</link>
	<width>32</width>
	<height>32</height>
</image> 
<site xmlns="com-wordpress:feed-additions:1">212686646</site>	<item>
		<title>A Grounded Conceptual Model for Ownership Types in Rust</title>
		<link>https://cacm.acm.org/research-highlights/a-grounded-conceptual-model-for-ownership-types-in-rust/</link>
					<comments>https://cacm.acm.org/research-highlights/a-grounded-conceptual-model-for-ownership-types-in-rust/#respond</comments>
		
		<dc:creator><![CDATA[Will Crichton, Gavin Gray, and Shriram Krishnamurthi]]></dc:creator>
		<pubDate>Thu, 23 Apr 2026 20:54:34 +0000</pubDate>
				<category><![CDATA[Education]]></category>
		<category><![CDATA[Software Engineering and Programming Languages]]></category>
		<guid isPermaLink="false">https://cacm.acm.org/?post_type=digital-library&#038;p=779417</guid>

					<description><![CDATA[<p>Our immediate goal in this work is to make Rust ownership types more understandable, and our broader goal is to explore the viability of different techniques for improving PL pedagogy. </p>]]></description>
										<content:encoded><![CDATA[<article>
<div style="background: #F5CBA7;">
<h2>Abstract</h2>
<p>Programmers learning Rust struggle to understand ownership types, Rust’s core mechanism for ensuring memory safety without garbage collection. This paper describes our process of systematically designing a pedagogy for ownership types. First, we studied Rust developers’ misconceptions of ownership to create the Ownership Inventory, a new instrument for measuring a person’s knowledge of ownership. We found that Rust learners could not connect Rust’s static and dynamic semantics, such as determining why an ill-typed program would (or would not) exhibit undefined behavior. Second, we created a conceptual model of Rust’s semantics that explains borrow checking in terms of flow-sensitive permissions on paths into memory. Third, we implemented a Rust compiler plug-in that visualizes programs under the model. Fourth, we integrated the permissions model and visualizations into a broader pedagogy of ownership by writing a new ownership chapter for <em>The Rust Programming Language</em>, a popular Rust textbook. Fifth, we evaluated an initial deployment of our pedagogy against the original version, using reader responses to the Ownership Inventory as a point of comparison. Thus far, the new pedagogy has improved learner scores on the Ownership Inventory by an average of 9% (N= 342, d= 0.56).</p>
</div>
<div class="body" lang="en">
<section id="sec1" class="sec">
<h2 class="heading"><span class="caption-label">1. </span>Introduction</h2>
<p id="p-1">Ownership is a programming discipline for managing the aliasing and mutation of data, enforced statically through ownership types. The flagship programming language for ownership is Rust, which empowers programmers to write memory-safe code without garbage collection. Rust’s ownership model synthesizes several ideas from programming language research, such as linear logic,<a class="reference-link xref xref-bibr" href="#B9" data-jats-ref-type="bibr" data-jats-rid="B9"><sup>9</sup></a> class-based alias management,<a class="reference-link xref xref-bibr" href="#B6" data-jats-ref-type="bibr" data-jats-rid="B6"><sup>6</sup></a> and region-based memory management.<a class="reference-link xref xref-bibr" href="#B10" data-jats-ref-type="bibr" data-jats-rid="B10"><sup>10</sup></a> Developers cannot use languages like C and C++ to build memory-safe systems at scale,<a class="reference-link xref xref-bibr" href="#B19" data-jats-ref-type="bibr" data-jats-rid="B19"><sup>19</sup></a> so the software industry is turning toward Rust. For example, Google’s Android team has thus far found zero memory vulnerabilities in 1.5 million lines of Rust code.<a class="reference-link xref xref-bibr" href="#B24" data-jats-ref-type="bibr" data-jats-rid="B24"><sup>24</sup></a></p>
<p id="p-2">This rosy picture of tech transfer belies a persistent obstacle: teaching Rust to prospective users, especially about ownership. Over the last four years, studies have found that Rust learners struggle to fix ownership type errors,<a class="reference-link xref xref-bibr" href="#B28" data-jats-ref-type="bibr" data-jats-rid="B28"><sup>28</sup></a> and users self-report that ownership is among their biggest barriers to learning Rust.<a class="reference-link xref xref-bibr" href="#B23" data-jats-ref-type="bibr" data-jats-rid="B23"><sup>23</sup></a> To wit: Advances in the technical factors of type systems require commensurate advances in the human factors of type systems.</p>
<p id="p-3">Our work started with the question: How can we systematically design a pedagogy for ownership types? Today, popular pedagogies for advanced type systems are driven by experts’ intuitions about how people learn, as well as by intuitions about what makes type systems difficult to understand. As practicing (computer) scientists, we wanted to approach pedagogic design through scientific principles: grounding the pedagogy in observations about the struggles of Rust learners, and then evaluating the pedagogy by its effects on learning outcomes. This paper describes how we put these principles into practice:</p>
<ol class="list" data-jats-list-type="order">
<li class="list-item">
<p id="p-4"><b>We ran a formative user study to identify misconceptions about ownership types held by Rust learners</b> (<a class="xref xref-sec" href="#sec5" data-jats-ref-type="sec" data-jats-rid="sec5">Section 3</a>). We designed a new instrument for evaluating understanding of ownership, the Ownership Inventory, by drawing tasks from commonly reported Rust issues on StackOverflow. We studied <i>N</i> = 36 Rust learners trying to solve Ownership Inventory problems. We found that learners can generally identify the surface-level reason for why a program is ill-typed with respect to ownership. However, learners do not understand what undefined behavior (if any) would occur if an ill-typed program were executed. This misunderstanding is reflected in inefficient and incorrect strategies used to fix ownership errors.</p>
</li>
<li class="list-item">
<p id="p-5"><b>We developed a conceptual model of ownership types to address these misconceptions</b> (<a class="xref xref-sec" href="#sec8" data-jats-ref-type="sec" data-jats-rid="sec8">Section 4</a>). The conceptual model represents the aspects of Rust’s static and dynamic semantics that are relevant to ownership while abstracting other details. The model provides learners a foundation to understand essential concepts such as undefined behavior and the incompleteness of Rust’s ownership type-checker, or “borrow checker.” We implemented tools to execute Rust programs under the conceptual model, generating traces which we visualize to illustrate how the model applies to concrete examples. We wrote a new chapter on ownership for a popular Rust textbook, <i>The Rust Programming Language</i> (<i>TRPL</i>),<a class="reference-link xref xref-bibr" href="#B12" data-jats-ref-type="bibr" data-jats-rid="B12"><sup>12</sup></a> using this conceptual model with these visualizations.</p>
</li>
<li class="list-item">
<p id="p-6"><b>We A/B tested our pedagogy against the <i>TRPL </i></b><b>baseline</b> (<a class="xref xref-sec" href="#sec11" data-jats-ref-type="sec" data-jats-rid="sec11">Section 5</a>). We set up and advertised a public website that hosts our <i>TRPL</i> fork. We measured learning outcomes with two kinds of quizzes: simpler comprehension questions about the conceptual model, and more difficult multiple-choice versions of the Ownership Inventory. Learners could correctly answer comprehension questions with 72% accuracy. Our initial deployment improved the average Inventory score from 48% to 57% (<i>N</i> = 342, <i>p</i> &lt; 0.001, <i>d</i> = 0.56).</p>
</li>
</ol>
</section>
<section id="sec2" class="sec">
<h2 class="heading"><span class="caption-label">2. </span>Background</h2>
<p id="p-7">If you, the reader, are not familiar with Rust or ownership, then it’s important to understand the basics to understand this paper’s contributions. One contribution is exactly an explanation of the basics of ownership. We will therefore give a brief exposition of Rust in the form of a heavily abbreviated version of the actual pedagogy we developed for Rust learners. Later sections will refer back to explain the design rationale of the concepts and diagrams mentioned here.</p>
<section id="sec3" class="sec">
<h3 class="heading"><span class="caption-label">2.1 </span>Ownership.</h3>
<p id="p-8">Rust is a systems programming language, most similar in its aims and scope to C++. Rust contains many standard imperative language features, such as mutable variables, conditionals, arrays, loops, structs, and functions.</p>
<p id="p-9">Rust manages memory through a system of compile-time checks known as <i>ownership</i>, as opposed to runtime checks (garbage collection) or manual checks (malloc/free). The basic idea is that heap allocations are <i>owned</i> by a variable, and allocations are freed when their owner goes out of scope, unless the owner <i>moves</i> ownership to another variable. This concept is illustrated in <a class="xref xref-fig" href="#F1" data-jats-ref-type="fig" data-jats-rid="F1">Figure 1</a>.</p>
<figure id="F1" class="fig" data-jats-position="float">
<div class="image-container"><img decoding="async" class="graphic" title="Figure 1. " src="https://cacm.acm.org/wp-content/uploads/2026/04/3796537_fig01.jpg" alt="" data-image-id="F1" data-image-type="figure" /></div><figcaption><span class="caption-label">Figure 1. </span> <span class="p">An example of how ownership and moves work in Rust with boxed arrays. At L1, assigning an array to a variable <code class="monospace">a</code> puts the data on the stack. At L2, assigning <code class="monospace">b</code> to <code class="monospace">a</code> copies the array into <code class="monospace">b</code>. At L3, we introduce a box, which owns data on the heap, therefore <code class="monospace">c</code> points to a heap allocation. At L4, assigning <code class="monospace">d</code> to <code class="monospace">c</code> copies the pointer (but not the array). This copy is called a move because <code class="monospace">c</code> can no longer be legally used. Once the function ends, the heap allocation is deallocated on behalf of its owner, <code class="monospace">d</code>.</span></p>
<div class="figcaption-footer"> </div>
</figcaption></figure>
<p id="p-11">The goal of ownership is to ensure memory safety, or more generally, to prevent <i>undefined behavior</i>. Undefined behavior means runtime operations without semantics under the Rust language specification, such as dereferencing a null pointer or pretending a string is a boolean. To prevent undefined behavior, the Rust compiler enforces several rules. First, data always has exactly one owner. Second, data is only deallocated on behalf of its owner. Third, data can only be accessed through its owner (except through references, discussed next). These rules prevent situations such as double-frees or use after-frees as shown in <a class="xref xref-fig" href="#F2" data-jats-ref-type="fig" data-jats-rid="F2">Figure 2</a>.</p>
<figure id="F2" class="fig" data-jats-position="float">
<div class="image-container"><img decoding="async" class="graphic" title="Figure 2. " src="https://cacm.acm.org/wp-content/uploads/2026/04/3796537_fig02.jpg" alt="" data-image-id="F2" data-image-type="figure" /></div><figcaption><span class="caption-label">Figure 2. </span> <span class="p">Rust’s ownership checks prevent undefined behavior. This diagram shows the counterfactual semantics for a program which does not actually compile. If allowed to compile, the box <code class="monospace">x</code> would be moved into the call to <code class="monospace">print</code> at L2, and get deallocated at the end of <code class="monospace">print</code>, leaving <code class="monospace">x</code> pointing to invalid memory at L3. It would violate memory safety to use <code class="monospace">x</code> at L4. Rust catches this error, reporting “<code class="monospace">error: use of moved value: ‘x‘</code>” at L4.</span></p>
<div class="figcaption-footer"> </div>
</figcaption></figure>
</section>
<section id="sec4" class="sec">
<h3 class="heading"><span class="caption-label">2.2 </span>Borrowing.</h3>
<p id="p-13">To allow accessing data without owning it, Rust provides <i>non-owning pointers</i> called references. References <i>borrow</i> data owned by another variable. Those references can be either immutable (created by <code class="monospace">&amp;</code>) or mutable (created by <code class="monospace">&amp;mut</code>). <a class="xref xref-fig" href="#F3" data-jats-ref-type="fig" data-jats-rid="F3">Figure 3</a> shows an example of creating an immutable reference to a piece of a string, and creating a mutable reference to pass to a helper function.</p>
<figure id="F3" class="fig" data-jats-position="float">
<div class="image-container"><img decoding="async" class="graphic" title="Figure 3. " src="https://cacm.acm.org/wp-content/uploads/2026/04/3796537_fig03.jpg" alt="" data-image-id="F3" data-image-type="figure" /></div><figcaption><span class="caption-label">Figure 3. </span> <span class="p">An example of immutable and mutable references in Rust. At L1, <code class="monospace">s</code> owns the string “Hello” on the heap. At L2, the reference <code class="monospace">lo</code> points to the end of <code class="monospace">s</code> but does not own it. At L3, a mutable reference to <code class="monospace">s</code> is used to append to the string. This causes the string’s buffer to resize, invalidating the reference <code class="monospace">lo</code>. At L4, <code class="monospace">s</code> is still valid, because <code class="monospace">arg</code> did not own <code class="monospace">s</code>.</span></p>
<div class="figcaption-footer"> </div>
</figcaption></figure>
<p id="p-15">Rust enforces safety with references through the concept of permissions. Permissions describe the kinds of operations one can do to data; they come and go based on which owners and references are in use at any given time. Specifically, Rust tracks three permission: readable (<span class="styled-content" style="color: #debc53;"><code class="monospace">R</code></span>), writable (<span class="styled-content" style="color: #0000ff;"><code class="monospace">W</code></span>), or ownable (<span class="styled-content" style="color: #d0021b;"><code class="monospace">O</code></span>). <a class="xref xref-fig" href="#F4" data-jats-ref-type="fig" data-jats-rid="F4">Figure 4</a> shows one example of how permissions change over the course of a program.</p>
<figure id="F4" class="fig" data-jats-position="float">
<div class="image-container"><img loading="lazy" decoding="async" class="graphic" title="Figure 4. " src="https://cacm.acm.org/wp-content/uploads/2026/04/3796537_fig04.jpg" alt="" width="429" height="449" data-image-id="F4" data-image-type="figure" /></div><figcaption><span class="caption-label">Figure 4. </span> <span class="p">Each table shows how the permissions on variables change after each statement. On line 1, the mutable owned string <code class="monospace">x</code> has all three permissions. On line 2, creating an immutable reference <code class="monospace">y</code> to <code class="monospace">x</code> removes the <b><span class="styled-content" style="color: #0000ff;"><code class="monospace">W</code></span></b> and <span class="styled-content" style="color: #d0021b;"><code class="monospace">O</code></span> permissions from <code class="monospace">x</code> so <code class="monospace">y</code> cannot be invalidated while in use. <code class="monospace">*y</code> only has <span class="styled-content" style="color: #debc53;"><code class="monospace"><b>R</b></code></span> permissions because it is an immutable reference. On line 3, <code class="monospace">x</code> regains its permissions after <code class="monospace">y</code> is no longer used. On line 4, <code class="monospace">x</code> loses its permissions after it is no longer used.</span></p>
<div class="figcaption-footer"> </div>
</figcaption></figure>
<p id="p-17">Each operation expects certain permissions from its inputs. For example, getting the length of a vector requires <span class="styled-content" style="color: #debc53;"><code class="monospace">R</code></span>. Pushing an element to a vector requires <span class="styled-content" style="color: #0000ff;"><code class="monospace">W</code></span>. Deallocating a vector requires <span class="styled-content" style="color: #d0021b;"><code class="monospace">O</code></span>. Rust’s <i>borrow checker</i> analyzes a program to ensure that variables have their requisite permissions in each operation, raising a compiler error otherwise. <a class="xref xref-fig" href="#F5" data-jats-ref-type="fig" data-jats-rid="F5">Figure 5</a> shows how these expectations occur in a Rust program.</p>
<figure id="F5" class="fig" data-jats-position="float">
<div class="image-container"><img decoding="async" class="graphic" title="Figure 5. " src="https://cacm.acm.org/wp-content/uploads/2026/04/3796537_fig05.jpg" alt="" data-image-id="F5" data-image-type="figure" /></div><figcaption><span class="caption-label">Figure 5. </span> <span class="p">Operations which expect permissions show their expectations inline. On line 2, the borrow <code class="monospace">&amp;x</code> requires <b><span class="styled-content" style="color: #debc53;"><code class="monospace">R</code></span></b> permission on <code class="monospace">x</code>, which it has. On line 3, the method call <code class="monospace">.push_str</code> requires <b><span class="styled-content" style="color: #debc53;"><code class="monospace">R</code></span></b> and <b><span class="styled-content" style="color: #0000ff;"><code class="monospace">W</code></span></b>, but <code class="monospace">x</code> is not writable due to the borrow by <code class="monospace">y</code>. This program is rejected by the compiler with the error: “<code class="monospace">error: cannot borrow ‘x‘ as mutable because it is also borrowed as immutable</code>”.</span></p>
<div class="figcaption-footer"> </div>
</figcaption></figure>
</section>
</section>
<section id="sec5" class="sec">
<h2 class="heading"><span class="caption-label">3. </span>A Concept Inventory for Ownership</h2>
<p id="p-19">Before this work, Rust learners did not learn about ownership in the style of <a class="xref xref-sec" href="#sec2" data-jats-ref-type="sec" data-jats-rid="sec2">Section 2</a>. Prior Rust learning resources adopted a range of pedagogical techniques, but the consistent outcome was that Rust learners struggled in practice to work with ownership. We first sought to understand the core misconceptions that underlay those struggles. Our overarching methodology was the development of a <i>concept inventory</i> for ownership, henceforth called the “Ownership Inventory.” In education research, a concept inventory (CI) is a test, usually composed of multiple-choice questions, about a narrow domain where the questions and distractors are drawn from common misconceptions about the domain.<a class="reference-link xref xref-bibr" href="#B11" data-jats-ref-type="bibr" data-jats-rid="B11"><sup>11</sup></a></p>
<p id="p-20">We set out to design the Ownership Inventory for two reasons. First, the misconceptions we would observe in creating the Inventory would inform our eventual pedagogy. Second, we could use the Inventory to evaluate the efficacy of an intervention. If our pedagogy is better than before, then it should cause learners to score higher on the Inventory.</p>
<p id="p-21">To construct the Inventory, we designed open-ended questions about ownership in situations that frequently stymie Rust learners. We searched for the most common questions asked about Rust on StackOverflow that pertain to ownership. We iteratively categorized each question and identified four main categories of ownership problems: dangling pointers, overlapping borrows, illegal borrow promotion, and lifetime parameters. For each category, we selected a few representative StackOverflow questions and cleaned up the snippet in question. For example, <a class="xref xref-fig" href="#F6" data-jats-ref-type="fig" data-jats-rid="F6">Figure 6</a> shows a StackOverflow question in the “illegal borrow promotion” category and the corresponding clean program.</p>
<figure id="F6" class="fig" data-jats-position="float">
<div class="image-container"><img loading="lazy" decoding="async" class="graphic" title="Figure 6. " src="https://cacm.acm.org/wp-content/uploads/2026/04/3796537_fig06.jpg" alt="" width="1014" height="348" data-image-id="F6" data-image-type="figure" /></div><figcaption><span class="caption-label">Figure 6. </span> <span class="p">An example of how we created snippets for the Ownership Inventory.</span></p>
<div class="figcaption-footer"> </div>
</figcaption></figure>
<p id="p-23">We designed a single set of template questions that apply to each program. The template represents each stage of reasoning involved in fixing an ownership error:</p>
<ol class="list" data-jats-list-type="order">
<li class="list-item">
<p id="p-24">What error message would you expect from the compiler?</p>
</li>
<li class="list-item">
<p id="p-25">Assume the compiler <i>did not</i> reject this function. (A) What is a program that calls this function which would violate memory safety or cause a data race? (B) Explain your reasoning.</p>
</li>
<li class="list-item">
<p id="p-26">(A) How can this function be changed to pass the compiler while preserving its intent? (B) Explain your reasoning.</p>
</li>
</ol>
<section id="sec6" class="sec">
<h3 class="heading"><span class="caption-label">3.1 </span>Methodology.</h3>
<p id="p-27">After developing the open-ended Ownership Inventory, we next administered the Inventory to elicit misconceptions that Rust learners have about ownership. We recruited <span class="inline-formula"><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><mrow><mi>N</mi><mo>=</mo><mn>36</mn></mrow></math></span> participants for the study. We found Rust learners by embedding an advertisement for the study within the online version of <i>TRPL</i>. Participants were required to be 18 years or older, and they were required to have completed reading <i>TRPL</i> before participating. Participants were compensated $20 and had, on average, <span class="inline-formula"><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><mrow><mn>1</mn><mo>.</mo><mn>7</mn></mrow></math></span> (<span class="inline-formula"><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><mrow><mi>σ</mi><mo>=</mo><mn>2</mn><mo>.</mo><mn>2</mn></mrow></math></span>) prior years of experience with either C or C++.</p>
<p id="p-28">We created a Web interface that presents participants with a program and prompts for open-ended responses to each question. The interface uses the Monaco code editor running a Rust language server via a WebAssembly build of Rust Analyzer. The in-browser IDE allows participants to get information about the type and functionality of unfamiliar methods.</p>
<p id="p-29">To evaluate the accuracy of participant’s answers, the first two authors independently coded each response as correct or incorrect. After the first round of coding, the authors resolved major disagreements, then independently re-coded the data. After the second round, the inter-rater reliability was 91% in terms of raw agreement and <span class="inline-formula"><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><mrow><mi>κ</mi><mo>=</mo><mn>0</mn><mo>.</mo><mn>81</mn></mrow></math></span> as measured by Cohen’s <span class="inline-formula"><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><mi>κ</mi></math></span>. We considered this sufficient agreement to proceed with the analysis. For the quantitative results, we report scores as the average of the two raters’ scores on each item. To characterize the specific misconceptions that led to incorrect answers, we performed a thematic analysis of participant responses.</p>
</section>
<section id="sec7" class="sec">
<h3 class="heading"><span class="caption-label">3.2 </span>Results.</h3>
<p id="p-30">Participants could usually predict why the borrow checker would reject a program (Q1, 64% correct). However, participants were less able to fix the program (Q3a, 46%) and design counterexample (Q2a, 31%). Their accuracy further drops when asked to justify their answer (Q2b, 15% and Q3b, 31%). That is, participants could sometimes create counterexamples and fixes without understanding why their answer is correct.</p>
<p id="p-31">Participants’ reasonable performance on Q1 suggests that Rust learners generally understand the surface-level reason for why a program is rejected. However, participants’ comparatively poor performance on Q2a and Q2b suggests Rust learners do not understand the deeper reasons that justify the ownership rules. For instance:</p>
<ul class="list" data-jats-list-type="bullet">
<li class="list-item">
<p id="p-32">Participants struggled to construct a correct counterexample to an unsafe function. Participants often gave counterexamples which executed the function but failed to actually trigger undefined behavior, such as dereferencing an invalid pointer.</p>
</li>
<li class="list-item">
<p id="p-33">Participants struggled to identify when a function is actually safe and no counterexample exists, such as mutably referencing disjoint indexes in an array.</p>
</li>
<li class="list-item">
<p id="p-34">Participants could usually change a broken function to pass the borrow checker, but these fixes were not always correct and idiomatic. For example, participants might clone a data structure, mutate the clone, and then immediately throw away the cloned data.</p>
</li>
</ul>
</section>
</section>
<section id="sec8" class="sec">
<h2 class="heading"><span class="caption-label">4. </span>A Conceptual Model for Ownership</h2>
<p id="p-35">Our study of the open-ended Ownership Inventory suggested Rust learners lacked a depth of understanding of ownership in a few key areas. In particular:</p>
<ol class="list" data-jats-list-type="order">
<li class="list-item">
<p id="p-36">Rust learners generally did not understand what the borrow checker was protecting them from, that is, what would go wrong without ownership.</p>
</li>
<li class="list-item">
<p id="p-37">Rust learners generally could not distinguish between issues of soundness and completeness, that is, whether a program was genuinely unsafe, or rejected due to limitations of the borrow checker.</p>
</li>
</ol>
<p id="p-38">Our hypothesis was that the core missing pedagogical component was a durable <i>conceptual model</i> of Rust’s semantics with respect to ownership. A conceptual model is a way of thinking about a thing, distinct from a mental model as the concrete way a given person thinks about a thing. The challenge for designers of pedagogy is to provide a conceptual model at an appropriate level of abstraction.</p>
<p id="p-39">Starting with an example simpler than ownership, consider teaching about integer addition in Rust. The challenge is to design a conceptual model of the semantics of the statement <code class="monospace">let z = x + y</code> where <code class="monospace">x, y : i32</code>. The true runtime semantics of Rust’s integer addition include aspects like two’s complement overflow and auto-vectorization due to LLVM’s optimizations. But for the average Rust user, these details are usually irrelevant for correctly using addition in routine programming tasks. A conceptual model that approximates the semantics as “<span class="inline-formula"><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><mrow><mi>x</mi><mo>,</mo><mi>y</mi><mo>∈</mo><mi>Z</mi></mrow></math></span> and <span class="inline-formula"><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><mrow><mi>x</mi><mo>+</mo><mi>y</mi></mrow></math></span> is standard integer addition” is a generally viable model.</p>
<p id="p-40">In the context of ownership, Rust learning resources would often provide a conceptual model in the form of patterns of bad code. These books explain concepts like mutable references by saying, for example, “Borrows can be read-only or read-write. Only one read-write borrow can exist at any one time.”<a class="reference-link xref xref-bibr" href="#B18" data-jats-ref-type="bibr" data-jats-rid="B18"><sup>18</sup></a> However, these types of explanations provide little grounding to understand either the safety justification for ownership, or the distinction between soundness and completeness. We therefore set out to build a more holistic conceptual model of ownership that could encompass these concerns.</p>
<section id="sec9" class="sec">
<h3 class="heading"><span class="caption-label">4.1 </span>Conceptual model.</h3>
<p id="p-41">We developed two related conceptual models of Rust’s runtime (<i>dynamic</i>) semantics and Rust’s compile-time (<i>static</i>) semantics. A key design constraint for both models is that they needed to be <i>executable</i>, that is, we could simulate a program under each model to collect trace information for visualization as opposed to a conceptual model that exists purely in prose or mathematical formalism.</p>
<p id="p-42">The main goal of the dynamic semantics model is to abstract Rust’s runtime enough to eliminate details irrelevant to ownership (e.g., whether data is placed in a register or spilled onto the stack) while maintaining details essential to ownership (e.g., whether data is allocated on the heap). Conveniently, the Rust compiler contains an intermediate representation of Rust programs called MIR, and MIR satisfies both design goals: Its semantics hit the right level of abstraction, and the Rust compiler provides a built-in interpreter for MIR for purposes of constant evaluation. We therefore instrumented this interpreter to generate the diagrams shown in <a class="xref xref-fig" href="#F1" data-jats-ref-type="fig" data-jats-rid="F1">Figure 1</a> and <a class="xref xref-fig" href="#F2" data-jats-ref-type="fig" data-jats-rid="F2">Figure 2</a>.</p>
<p id="p-43">A key aspect of the dynamic semantic model is that we can configure the MIR interpreter to evaluate programs even when they contain borrow checker violations. This allows Rust learners to observe the counterfactual behavior of what would go wrong without the protections of the borrow checker, as shown in <a class="xref xref-fig" href="#F2" data-jats-ref-type="fig" data-jats-rid="F2">Figure 2</a>.</p>
<p id="p-44">The dynamic model provides the foundation for understanding how programs can go wrong. The static model should then help learners understand how Rust’s borrow checker catches programs that could go wrong (soundness), as well as when safe programs may be rejected (incompleteness). As a flow-sensitive analysis, borrow checking is more complex than the usual type system encountered by today’s programmers. So, in designing a conceptual model of the borrow checker, our main goal was to condense the complex intermediate state of the analysis into a comprehensible, visualizable object. The result is the <i>permissions model</i> of borrow checking, as shown in <a class="xref xref-fig" href="#F4" data-jats-ref-type="fig" data-jats-rid="F4">Figure 4</a> and <a class="xref xref-fig" href="#F5" data-jats-ref-type="fig" data-jats-rid="F5">Figure 5</a>.</p>
<p id="p-45">The permissions model allows us to explain how the borrow checker “thinks” about a program in a more general manner than describing one-off rule violations. For example, permissions can be computed and visualized for programs that pass the borrow checker, as shown in <a class="xref xref-fig" href="#F4" data-jats-ref-type="fig" data-jats-rid="F4">Figure 4</a>. This model separates out how the compiler computes facts about a program (the permission change tables) versus how those facts are used to check a program (the permission expectations).</p>
<p id="p-46">These permissions are a conceptual model in that the Rust compiler does not literally compute the <b><span class="styled-content" style="color: #f19a38;">R</span></b><b><span class="styled-content" style="color: #387ad6;">W</span></b><b><span class="styled-content" style="color: #387ad6;">O</span></b> permissions. We implemented an analysis which derives the permissions from the internal state of the borrow checker, and that analysis drives the permissions visualization. The permissions are formally equivalent to the actual borrow checker in a manner we demonstrate in the full paper.</p>
</section>
<section id="sec10" class="sec">
<h3 class="heading"><span class="caption-label">4.2 </span>Pedagogy.</h3>
<p id="p-47">After designing the conceptual models, we next designed a pedagogy to help Rust learners internalize these models, as in <a class="xref xref-sec" href="#sec2" data-jats-ref-type="sec" data-jats-rid="sec2">Section 2</a>. Rather than designing an entire Rust curriculum from scratch, we instead forked <i>TRPL,</i><a class="reference-link xref xref-bibr" href="#B12" data-jats-ref-type="bibr" data-jats-rid="B12"><sup>12</sup></a> which covers most of the language’s core features, and it is the official Rust learning resource endorsed by the Rust project.</p>
<p id="p-48">We designed a new pedagogy of ownership as a replacement for the existing chapter on ownership in <i>TRPL</i>. We start by explaining the core ideas of undefined behavior and memory safety through boxes and moves (the dynamic model). We then introduce references, the borrow checker, and permissions (the static model). Finally, we synthesize these ideas by providing multiple examples of how a Rust programmer can interpret and fix ownership errors, emphasizing the distinction between soundness and completeness. The full text of the chapter is available online.<a class="footnote-link xref xref-fn" href="#fn1" data-jats-ref-type="fn" data-jats-rid="fn1"><sup>a</sup></a></p>
</section>
</section>
<section id="sec11" class="sec">
<h2 class="heading"><span class="caption-label">5. </span>Evaluation</h2>
<p id="p-49">We sought to evaluate our pedagogy on whether it helps learners understand ownership in Rust. This raised two immediate questions. First, how do we find learners to try out our pedagogy? The vast majority of CS education research takes place in a classroom, but we explored an alternative route: free online textbooks. These resources provide access to a larger and more diverse population of learners than CS undergraduates at a single institution. To that end, we set up a publicly accessible website that hosts our <i>TRPL</i> fork, and it has been visited by tens of thousands of Rust learners to date. This site provides a research platform for analyzing and intervening in the Rust learning process. The intervention described in this paper is the first step in an ongoing experiment to leverage the platform for systematically improving Rust education at scale.</p>
<p id="p-50">The second key question is: How do we know if learners understand ownership after following our pedagogy? <i> vc</i> is difficult to define—ideally, a longitudinal study might measure understanding as learners’ ability to productively write safe and performant Rust code in their context of use. But for lack of such data, we instead opted for a common substitute: quiz questions. We diffused the quiz questions throughout the ownership chapter and the rest of the book. Furthermore, we distinguished between two kinds of quiz questions designed to answer two research questions:</p>
<ol class="list" data-jats-list-type="order">
<li class="list-item">
<p id="p-51">Does our pedagogy help learners understand ownership <i>at all</i>?</p>
</li>
<li class="list-item">
<p id="p-52">Does our pedagogy help learners understand ownership <i>better than before</i>?</p>
</li>
</ol>
<p id="p-53">For RQ1, we asked participants simple comprehension questions about ownership, presented immediately following the book content that is relevant to a given question. These questions determine whether participants can transfer their ownership knowledge to situations similar to the text. Because the questions make references to the permissions model, we cannot establish a score baseline. Therefore we judge the scores in absolute rather than relative terms.</p>
<p id="p-54">For RQ2, we gave participants a multiple-choice version of the Ownership Inventory. We inserted these questions later in the book after covering the essential prerequisites for a given program. To compare the baseline <i>TRPL</i> pedagogy against ours, we ran a kind of temporal A/B test. Participants answered Inventory questions after reading the original <i>TRPL</i> content for a few weeks. We then deployed the intervention and continued receiving responses to the Inventory. We quantified the pedagogy’s effect based on the resulting change in scores.</p>
<section id="sec12" class="sec">
<h3 class="heading"><span class="caption-label">5.1 </span>Methodology.</h3>
<p id="p-55">We recruited participants by advertising in the title page of the official Web version of <i>TRPL</i>, courtesy of the authors. Since this advertisement was put up on Nov. 1, 2022, our <i>TRPL</i> fork has received an average of 450 visitors per day, as measured by unique session IDs stored via cookies.</p>
<figure id="F7" class="fig">
<div class="image-container"><img loading="lazy" decoding="async" class="graphic" title="Figure 7. " src="https://cacm.acm.org/wp-content/uploads/2026/04/3796537_fig07.jpg" alt="" width="789" height="827" data-image-id="F7" data-image-type="figure" /></div><figcaption><span class="caption-label">Figure 7. </span> <span class="p">A comprehension question (“Analysis state in permissions diagram”) that tests whether a person can correctly interpret a permission diagram.</span></p>
<div class="figcaption-footer"> </div>
</figcaption></figure>
<p id="p-57">We developed 11 comprehension questions to cover the content of the ownership chapter. <a class="xref xref-fig" href="#F7" data-jats-ref-type="fig" data-jats-rid="F7">Figure 7</a> shows one example—to test understanding of permission diagrams, we asked participants to infer the permissions for a path at a given point.</p>
<p id="p-58">We developed 24 Ownership Inventory questions based on the programs in <a class="xref xref-sec" href="#sec5" data-jats-ref-type="sec" data-jats-rid="sec5">Section 3</a>. For each program, we created a close-ended version of Q1, Q2a, and Q3a (the justification questions Q2b/Q3b did not translate to the multiple-choice setting). Following the concept inventory methodology, we selected distractors from common misconceptions about the Inventory programs.</p>
<p id="p-59"><a class="xref xref-fig" href="#F8" data-jats-ref-type="fig" data-jats-rid="F8">Figure 8</a> shows an example of a multiple-choice Ownership Inventory question. The <code class="monospace">make_separator</code> task is an instance of a dangling stack reference. A common incorrect counterexample provided by participants in <a class="xref xref-sec" href="#sec5" data-jats-ref-type="sec" data-jats-rid="sec5">Section 3</a> was the snippet <code class="monospace">let s = make_separator("")</code> which creates a dangling pointer, but does not use it. By using that incorrect answer as a distractor, the multiple-choice question is more likely to test for the presence of this misconception. As an example, see the Chapter 6 Inventory questions.<a class="footnote-link xref xref-fn" href="#fn2" data-jats-ref-type="fn" data-jats-rid="fn2"><sup>b</sup></a></p>
<figure id="F8" class="fig" data-jats-position="float">
<div class="image-container"><img loading="lazy" decoding="async" class="graphic" title="Figure 8. " src="https://cacm.acm.org/wp-content/uploads/2026/04/3796537_fig08.jpg" alt="" width="876" height="708" data-image-id="F8" data-image-type="figure" /></div><figcaption><span class="caption-label">Figure 8. </span> <span class="p">A multiple-choice Ownership Inventory question (Q2a for the <code class="monospace">make_separator</code> program). The distractors are drawn from common incorrect answers given for the open-ended version of the same question.</span></p>
<div class="figcaption-footer"> </div>
</figcaption></figure>
<p id="p-61">On January 3, 2023 we deployed the questions to our website and started gathering pre-intervention data on performance with the baseline <i>TRPL</i> pedagogy. After 45 days, on Feb. 17, 2023 we deployed the initial draft of our new ownership chapter. After a few months of iteration, on June 15, 2023 we began gathering post-intervention data. Data collection continued for 45 days until July 30, 2023.</p>
</section>
<section id="sec13" class="sec">
<h3 class="heading"><span class="caption-label">5.2 </span>Results.</h3>
<section id="sec14" class="sec">
<h4 class="heading"><span class="caption-label">5.2.1 </span>RQ1: Does our pedagogy help learners understand ownership at all?</h4>
<p id="p-62">Overall, readers’ accuracy on the comprehension questions was 72%, suggesting that readers could mostly understand the basic concepts within our pedagogy. Readers were able to successfully interpret both runtime and compile-time diagrams and identify when the compiler was going to reject a program.</p>
<p id="p-63">However, readers’ mediocre performance on a few of the comprehension questions suggests their understanding may be somewhat shallow. For example, the ownership chapter provides an example program containing a variable <code class="monospace">x : &amp;Box&lt;i32&gt;</code>, and explains that two dereferences like <code class="monospace">**x</code> are needed to access the inner integer. One question presents a program that constructs an expression of type <code class="monospace">Box&lt;&amp;Box&lt;i32&gt;&gt;</code> (including a runtime diagram), and asks respondents to determine the number of dereference operations needed to access the inner <code class="monospace">i32</code>. Only 47% of respondents correctly answer three, suggesting readers still leave with a somewhat fragile understanding of an essential concept like pointers.</p>
</section>
<section id="sec15" class="sec">
<h4 class="heading"><span class="caption-label">5.2.2 </span>RQ2: Does our pedagogy help learners understand ownership better than before?</h4>
<p id="p-64">We focus on the first 18 Inventory questions, as those questions received enough responses to make statistical inferences. First, we analyze the overall Inventory scores for the <span class="inline-formula"><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><mrow><mi>N</mi><mo>=</mo></mrow></math></span> 177 (pre) / 165 (post) participants who completed the first 18 questions. The average pre-intervention score was 48% (<span class="inline-formula"><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><mrow><mi>σ</mi><mo>=</mo><mn>16</mn><mo>%</mo></mrow></math></span>). Notably, the average score on the open-ended Inventory questions in <a class="xref xref-sec" href="#sec5" data-jats-ref-type="sec" data-jats-rid="sec5">Section 3</a> was 41% (which should be more difficult than equivalent multiple-choice questions), showing that the quantitative results of the formative study reasonably generalized to a larger sample. The average post-intervention score was 57% (<span class="inline-formula"><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><mrow><mi>σ</mi><mo>=</mo><mn>15</mn><mo>%</mo></mrow></math></span>). Using a two-tailed <span class="inline-formula"><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><mi>t</mi></math></span>-test, the difference is statistically significant (<span class="inline-formula"><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><mrow><mi>p</mi><mo>&lt;</mo><mn>0</mn><mo>.</mo><mn>001</mn></mrow></math></span>). The normalized effect size as measured by Cohen’s <span class="inline-formula"><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><mi>d</mi></math></span> is <span class="inline-formula"><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><mrow><mn>0</mn><mo>.</mo><mn>56</mn></mrow></math></span>. Therefore, the pedagogy had a statistically significant positive effect (+9%) on overall Inventory performance. The results also confirm that Inventory questions are substantially harder than the comprehension questions.</p>
<p id="p-65">Second, we analyze the intervention’s effect on each Inventory question individually. The intervention had a statistically significant effect on 10/18 questions. <a class="xref xref-table" href="#T1" data-jats-ref-type="table" data-jats-rid="T1">Table 1</a> shows the size of these effects, including almost-significant effects. Overall, the pooled significant effect was 10% or <span class="inline-formula"><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><mrow><mi>d</mi><mo>=</mo><mn>0</mn><mo>.</mo><mn>22</mn></mrow></math></span> (note that the question-level effect size is smaller than the quiz-level effect size due to the higher per-question variance). Between the different types of questions, the intervention primarily affected performance on questions about undefined behavior (Q2) and fixing a type error (Q3) more so than identifying a type error (Q1). For example, with <code class="monospace">make_separator</code> Q2, the +7% effect corresponds to an 8% decrease in the incorrect response of “does not have counterexamples.” Conversely, for <code class="monospace">reverse</code> and <code class="monospace">apply_curve</code> Q2, the +13%/+17% effects correspond to participants answering correctly that these functions are safe and do not have counterexamples.</p>
<figure id="T1" class="table-wrap" data-jats-position="float">
<div class="caption"><span class="caption-label">Table 1. </span> <span class="p">Effects of the permissions pedagogy for readers’ accuracy on Ownership Inventory questions. Questions are presented in the order encountered by readers. Only effects with <span class="inline-formula"><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><mrow><mi>p</mi><mo>&lt;</mo><mn>0</mn><mo>.</mo><mn>15</mn></mrow></math></span> are included here, with <span class="inline-formula"><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><mrow><mi>p</mi><mo>&lt;</mo><mn>0</mn><mo>.</mo><mn>05</mn></mrow></math></span> in bold.</span></div>
<div class="table-container">
<table class="table table-bordered table-condensed table-hover" data-jats-frame="hsides" data-jats-rules="rows">
<thead>
<tr>
<th><b>Task</b></th>
<th style="text-align: right;"><b>Q.</b></th>
<th><b>Before</b></th>
<th style="text-align: right;"><span class="inline-formula"> <math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"> <mi>N</mi> </math> </span></th>
<th><b>After</b></th>
<th style="text-align: right;"><span class="inline-formula"> <math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"> <mi>N</mi> </math> </span></th>
<th><b>Effect</b></th>
<th style="text-align: right;"><span class="inline-formula"> <math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"> <mi>d</mi> </math> </span></th>
<th style="text-align: right;"><span class="inline-formula"> <math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"> <mi>p</mi> </math> </span></th>
</tr>
</thead>
<tbody>
<tr>
<td><code class="monospace">make_separator</code></td>
<td style="text-align: right;">Q2</td>
<td style="text-align: right;">33%</td>
<td style="text-align: right;">1120</td>
<td style="text-align: right;">40%</td>
<td style="text-align: right;">660</td>
<td style="text-align: right;">+7%</td>
<td style="text-align: right;">0.14</td>
<td style="text-align: right;"><b>0.007</b></td>
</tr>
<tr>
<td><code class="monospace">make_separator</code></td>
<td style="text-align: right;">Q3</td>
<td style="text-align: right;">57%</td>
<td style="text-align: right;">1120</td>
<td style="text-align: right;">62%</td>
<td style="text-align: right;">660</td>
<td style="text-align: right;">+5%</td>
<td style="text-align: right;">0.11</td>
<td style="text-align: right;"><b>0.024</b></td>
</tr>
<tr>
<td><code class="monospace">get_or_default</code></td>
<td style="text-align: right;">Q1</td>
<td style="text-align: right;">56%</td>
<td style="text-align: right;">1120</td>
<td style="text-align: right;">66%</td>
<td style="text-align: right;">660</td>
<td style="text-align: right;">+10%</td>
<td style="text-align: right;">0.21</td>
<td style="text-align: right;"><b>&lt;0.001</b></td>
</tr>
<tr>
<td><code class="monospace">get_or_default</code></td>
<td style="text-align: right;">Q2</td>
<td style="text-align: right;">10%</td>
<td style="text-align: right;">1120</td>
<td style="text-align: right;">16%</td>
<td style="text-align: right;">660</td>
<td style="text-align: right;">+6%</td>
<td style="text-align: right;">0.19</td>
<td style="text-align: right;"><b>&lt;0.001</b></td>
</tr>
<tr>
<td><code class="monospace">remove_zeros</code></td>
<td style="text-align: right;">Q3</td>
<td style="text-align: right;">35%</td>
<td style="text-align: right;">629</td>
<td style="text-align: right;">52%</td>
<td style="text-align: right;">470</td>
<td style="text-align: right;">+17%</td>
<td style="text-align: right;">0.34</td>
<td style="text-align: right;"><b>&lt;0.001</b></td>
</tr>
<tr>
<td><code class="monospace">reverse</code></td>
<td style="text-align: right;">Q2</td>
<td style="text-align: right;">28%</td>
<td style="text-align: right;">629</td>
<td style="text-align: right;">40%</td>
<td style="text-align: right;">470</td>
<td style="text-align: right;">+13%</td>
<td style="text-align: right;">0.27</td>
<td style="text-align: right;"><b>&lt;0.001</b></td>
</tr>
<tr>
<td><code class="monospace">reverse</code></td>
<td style="text-align: right;">Q3</td>
<td style="text-align: right;">21%</td>
<td style="text-align: right;">629</td>
<td style="text-align: right;">33%</td>
<td style="text-align: right;">470</td>
<td style="text-align: right;">+13%</td>
<td style="text-align: right;">0.29</td>
<td style="text-align: right;"><b>&lt;0.001</b></td>
</tr>
<tr>
<td><code class="monospace">find_nth</code></td>
<td style="text-align: right;">Q1</td>
<td style="text-align: right;">86%</td>
<td style="text-align: right;">314</td>
<td style="text-align: right;">90%</td>
<td style="text-align: right;">374</td>
<td style="text-align: right;">+4%</td>
<td style="text-align: right;">0.12</td>
<td style="text-align: right;">0.099</td>
</tr>
<tr>
<td><code class="monospace">find_nth</code></td>
<td style="text-align: right;">Q2</td>
<td style="text-align: right;">16%</td>
<td style="text-align: right;">314</td>
<td style="text-align: right;">23%</td>
<td style="text-align: right;">374</td>
<td style="text-align: right;">+7%</td>
<td style="text-align: right;">0.18</td>
<td style="text-align: right;"><b>0.018</b></td>
</tr>
<tr>
<td><code class="monospace">find_nth</code></td>
<td style="text-align: right;">Q3</td>
<td style="text-align: right;">27%</td>
<td style="text-align: right;">316</td>
<td style="text-align: right;">34%</td>
<td style="text-align: right;">374</td>
<td style="text-align: right;">+7%</td>
<td style="text-align: right;">0.15</td>
<td style="text-align: right;"><b>0.041</b></td>
</tr>
<tr>
<td><code class="monospace">apply_curve</code></td>
<td style="text-align: right;">Q2</td>
<td style="text-align: right;">41%</td>
<td style="text-align: right;">452</td>
<td style="text-align: right;">58%</td>
<td style="text-align: right;">374</td>
<td style="text-align: right;">+17%</td>
<td style="text-align: right;">0.34</td>
<td style="text-align: right;"><b>&lt;0.001</b></td>
</tr>
<tr>
<td style="text-align: right;" colspan="6"><b>Pooled significant effect:</b></td>
<td style="text-align: right;"><b>+10%</b></td>
<td style="text-align: right;"><b>0.22</b></td>
<td> </td>
</tr>
</tbody>
</table>
</div>
</figure>
</section>
</section>
<section id="sec16" class="sec">
<h3 class="heading"><span class="caption-label">5.3 </span>Threats to validity.</h3>
<p id="p-67">This experiment assumes that the Ownership Inventory is a valid instrument to measure a person’s understanding of ownership. To that end, we designed the Inventory such that the situations reflect common ownership problems (by weighting based on StackOverflow), and such that the questions reflect each stage of reasoning about ownership (based on our formative study). However, future work should validate the extent to which performance on the Inventory correlates to performance in solving ownership problems in practice.</p>
<p id="p-68">The setting of an online textbook provides the benefit of scale, but it also poses methodological challenges due to lack of controls. One such threat is the uncontrolled quizzing environment. A reader could augment their problem-solving with external aids like a friend, a compiler, a Google search, a large language model, and so on. Participants could also be influenced by learning material outside the book, such as the official or Rust-related YouTube videos. To combat this threat, we explicitly instructed participants to not use external resources while solving quiz problems, and the quiz widget takes over the browser tab while taking a quiz. Moreover, we assume the average participant will be a good actor—our readers are taking these quizzes for their own edification, not to get paid by us or to get a good grade. Gathering enough data should turn bad actors into noise.</p>
<p id="p-69">Another threat is the uncontrolled assignment to experimental conditions. We chose not to perform a randomized-controlled trial for the reasons discussed in <a class="xref xref-sec" href="#sec12" data-jats-ref-type="sec" data-jats-rid="sec12">Section 5.1</a>. However, it is possible that temporal correlations in readership could have affected our results. For example, if all the C++ engineers at one company decided to start learning Rust at the same time, then average scores would likely go up in that window of time compared to the average in the limit.</p>
<p id="p-70">A final threat is teaching to the test. Unlike us, the <i>TRPL</i> authors were not aware of the Ownership Inventory when they wrote the book. At the extreme, if our pedagogy taught the exact answers to Inventory questions, then Inventory scores would not be a useful measure of ownership understanding. At the same time, part of the point of our experiment is exactly to teach to the test! For example, the Inventory is intentionally designed to measure understanding of undefined behavior; in turn, we intentionally designed our new pedagogy to explain undefined behavior. Like any well-meaning educator, we sought a balance. The Inventory materials do not appear anywhere in the revised text. But we do, for instance, walk through an example of how iterator invalidation causes undefined behavior, which is similar to the <code class="monospace">remove_zeros</code> problem.</p>
</section>
</section>
<section id="sec17" class="sec">
<h2 class="heading"><span class="caption-label">6. </span>Related Work</h2>
<p id="p-71">Almeida et al.<a class="reference-link xref xref-bibr" href="#B1" data-jats-ref-type="bibr" data-jats-rid="B1"><sup>1</sup></a> created RustViz, a visualization format for ownership annotations on a Rust program. In terms of pedagogy, RustViz’s premise is that the key challenge with ownership is “the user must learn to mentally simulate the logic of the borrow checker.” Our pedagogy is based more on connecting Rust’s static and dynamic semantics, which we show in our formative study to be a more serious problem for Rust learners. In terms of implementation, RustViz diagrams are constructed by hand using a DSL while we automatically generate our diagrams from the compiler. In terms of evaluation, Almeida et al. deployed RustViz in a classroom, finding that students responded to a Likert item that the visualizations were “helpful in terms of improving their understanding of ownership.” Our evaluation goes further to quantify the effect of our pedagogy on learning outcomes.</p>
<p id="p-72">Our runtime diagram is similar to program-state visualizers in prior work—see Sorva<a class="reference-link xref xref-bibr" href="#B22" data-jats-ref-type="bibr" data-jats-rid="B22"><sup>22</sup></a> for a survey. Our findings about misconceptions of undefined behavior and memory safety are consistent with prior work on teaching C. For instance, Lam et al.<a class="reference-link xref xref-bibr" href="#B13" data-jats-ref-type="bibr" data-jats-rid="B13"><sup>13</sup></a> found in a study of undergraduates who had taken a computer systems course that “many students displayed little knowledge or had misunderstandings about memory and memory layout” and would simply say “something bad” happens when unsafe operations occur.</p>
<p id="p-73">Our work continues a line of CS education research about conceptual models. Bayman and Mayer<a class="reference-link xref xref-bibr" href="#B2" data-jats-ref-type="bibr" data-jats-rid="B2"><sup>2</sup></a> first showed that an appropriate conceptual model for BASIC could help students “develop fewer misconceptions [&#8230;] and perform better on transfer tests.” du Boulay<a class="reference-link xref xref-bibr" href="#B8" data-jats-ref-type="bibr" data-jats-rid="B8"><sup>8</sup></a> coined the term “notional machine” for conceptual models specifically of a language’s dynamic semantics, which has received renewed focus in recent years.<a class="reference-link xref xref-bibr" href="#B7" data-jats-ref-type="bibr" data-jats-rid="B7"><sup>7</sup></a> Our work differs from most research on notional machines by focusing equally on a conceptual model of <i>static</i> semantics.</p>
<p id="p-74">Our work also intersects with a line of programming language research on the human factors of type systems and functional languages. Most prior work has focused on algorithms for identifying the root cause of confusing type-inference errors.<a class="reference-link xref xref-bibr" href="#B25" data-jats-ref-type="bibr" data-jats-rid="B25"><sup>25</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B27" data-jats-ref-type="bibr" data-jats-rid="B27"><sup>27</sup></a> Recent work has broadened scope to develop theories about how programmers read functional programs<a class="reference-link xref xref-bibr" href="#B16" data-jats-ref-type="bibr" data-jats-rid="B16"><sup>16</sup></a> and leverage the type system during development.<a class="reference-link xref xref-bibr" href="#B14" data-jats-ref-type="bibr" data-jats-rid="B14"><sup>14</sup></a></p>
<p id="p-75">This paper focuses on ownership types as they are implemented in Rust, but ownership types have taken many forms in prior work.<a class="reference-link xref xref-bibr" href="#B5" data-jats-ref-type="bibr" data-jats-rid="B5"><sup>5</sup></a> For instance, early systems of ownership focused on ensuring uniqueness of access to data by checking for dominance in the alias graph.<a class="reference-link xref xref-bibr" href="#B6" data-jats-ref-type="bibr" data-jats-rid="B6"><sup>6</sup></a> Later systems relaxed this constraint by permitting temporary borrowing of data, both mutably<a class="reference-link xref xref-bibr" href="#B3" data-jats-ref-type="bibr" data-jats-rid="B3"><sup>3</sup></a> and immutably.<a class="reference-link xref xref-bibr" href="#B20" data-jats-ref-type="bibr" data-jats-rid="B20"><sup>20</sup></a> The connection between ownership and permissions has been well-established within formal models, such as fractional permissions.<a class="reference-link xref xref-bibr" href="#B4" data-jats-ref-type="bibr" data-jats-rid="B4"><sup>4</sup></a></p>
</section>
<section id="sec18" class="sec">
<h2 class="heading"><span class="caption-label">7. </span>General Discussion</h2>
<p id="p-76">Future programming languages will undoubtedly have increasingly complex type systems. Rust is the language <i>du jour</i>, so this work focused on ownership types. But the next popular language could bring a renewed emphasis to any existing line of PL research: refinement types, session types, or even theorem proving. Effective transfer of these technologies will require pedagogies that do not expect learners to come equipped with Ph.D.-level knowledge of programming languages, mathematics, and Greek. While our immediate goal in this work was to make ownership types more understandable, our broader goal was to explore the viability of different techniques for improving PL pedagogy. In this section, we will briefly reflect on lessons learned.</p>
<p id="p-77">First, to develop a metric for understanding of ownership, we created a concept inventory by combining data from StackOverflow with a formative study of Rust learners. StackOverflow works for popular languages like Rust but is less useful for niche languages. Human-factors research on niche languages can instead consider using telemetry from developer interactions as has been explored for Racket<a class="reference-link xref xref-bibr" href="#B15" data-jats-ref-type="bibr" data-jats-rid="B15"><sup>15</sup></a> and Coq.<a class="reference-link xref xref-bibr" href="#B21" data-jats-ref-type="bibr" data-jats-rid="B21"><sup>21</sup></a> The concept inventory is an idea that could easily be reused in the context of other languages. Inventories can serve as communal benchmarks for progress in education research, like how datasets of programs serve as benchmarks for performance in compiler research.</p>
<p id="p-78">Second, to develop a conceptual model for ownership, we carefully selected a level of abstraction that was concrete enough to explain relevant phenomena like undefined behavior, while abstract enough to avoid unnecessary details. We leveraged the rich prior work on distilling the Rust type system into a small, explainable set of mechanisms, especially the Oxide<a class="reference-link xref xref-bibr" href="#B26" data-jats-ref-type="bibr" data-jats-rid="B26"><sup>26</sup></a> and Polonius<a class="reference-link xref xref-bibr" href="#B17" data-jats-ref-type="bibr" data-jats-rid="B17"><sup>17</sup></a> models. However, PL research usually distills type systems to permit formal reasoning, such as a soundness proof. An open question is how to distill type systems for didactic reasoning, that is, to help learners acquire a conceptual model valid for common tasks. For example, one of our principles was that our model must be encodable in a concise visual representation, which is not a property usually expected of standard PL research. Future work can investigate the properties of semantics that make them more or less explainable.</p>
<p id="p-79">Finally, to evaluate the efficacy of our pedagogy, we publicly deployed our textbook and compared pre/post-intervention scores on the Ownership Inventory. Collecting telemetry from quizzes in online learning resources is a readily applicable strategy for other contexts. Learners <i>want</i> to take quizzes to engage with the content they are reading. Temporal A/B testing offers a lightweight method for evaluating content changes without sophisticated infrastructure. We encourage anyone interested in programming language learning to try out our methodology. To that end, we have open-sourced our front-end quiz plug-in and our back-end telemetry system.<a class="footnote-link xref xref-fn" href="#fn3" data-jats-ref-type="fn" data-jats-rid="fn3"><sup>c</sup></a></p>
</section>
<section id="sec19" class="sec">
<h2 class="heading"><span class="caption-label">8. </span>Acknowledgments</h2>
<p id="p-80">The authors are immensely grateful to Niko Matsakis and Amazon. They provided both the encouragement and the funding to initiate this project, and supplied additional emergency funding when our first grant application fell through because we were studying Rust instead of C++. We thank Carol Nichols for taking a leap of faith in allowing us to advertise in <span class="sc">trpl</span>; this was essential for driving traffic to the experiment. Later parts of this work are partially supported by the US NSF under Grant No. 2319014.</p>
</section>
<section id="sec20" class="sec">
<h2 class="heading"><span class="caption-label">9. </span>Originally Published</h2>
<p id="p-81">This paper was originally published in <i>Proceedings of the 2023 Intern. Conf. on Object-oriented Programming, Systems, Languages, and Applications</i>.</p>
</section>
</div>
</article>
]]></content:encoded>
					
					<wfw:commentRss>https://cacm.acm.org/research-highlights/a-grounded-conceptual-model-for-ownership-types-in-rust/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
		      <dc:creator><![CDATA[Gavin Gray]]></dc:creator>
      <dc:creator><![CDATA[Shriram Krishnamurthi]]></dc:creator>
<post-id xmlns="com-wordpress:feed-additions:1">779417</post-id>	</item>
		<item>
		<title>Technical Perspective: A Grounded Conceptual Model for Ownership Types in Rust</title>
		<link>https://cacm.acm.org/research-highlights/technical-perspective-a-grounded-conceptual-model-for-ownership-types-in-rust/</link>
					<comments>https://cacm.acm.org/research-highlights/technical-perspective-a-grounded-conceptual-model-for-ownership-types-in-rust/#respond</comments>
		
		<dc:creator><![CDATA[Alex Potanin]]></dc:creator>
		<pubDate>Thu, 23 Apr 2026 20:51:58 +0000</pubDate>
				<category><![CDATA[Software Engineering and Programming Languages]]></category>
		<category><![CDATA[Theory]]></category>
		<guid isPermaLink="false">https://cacm.acm.org/?post_type=digital-library&#038;p=778194</guid>

					<description><![CDATA[<p>The core contribution of "Grounded Conceptual Model for Ownership Types in Rust," by Crichton et al., is a conceptual model that reinterprets Rust’s borrow checker in more user-friendly terms.</p>]]></description>
										<content:encoded><![CDATA[<article>
<div class="body" lang="en">
<section id="sec1" class="sec">
<p id="p-1"><i>Ownership types</i> emerged from the musings of James Noble, a young Ph.D. student working on software visualization for object-oriented programs and a tool called Tarraingím more than 30 years before the accompanying paper appeared in 2023. Noble was frustrated by unexpected changes to the objects being visualized, often caused by multiple references to the same object across different parts of a program. A last-minute chapter in Noble’s Ph.D. thesis, and later his own student (Dave Clarke), led to follow-up works and a series of papers in the late 1990s, culminating in the ownership types concept we know and love (to hate) today. Ownership types specify which parts of a program can access or modify a particular object.</p>
<p id="p-2">In the early days of ownership types, there was mostly a single overused motivating example presented in most ownership-related talks in the late 1990s and early 2000s: the JDK v1.1.1 applet permission mechanism that would return a direct reference to the permission array instead of a cloned copy, allowing applets to manipulate their own permissions at whim. Some even wondered if there was much to this ownership concept. Thankfully, a much broader adoption of ownership took hold, thanks to safe systems programming languages such as Rust. Yet, it also led to a proliferation of confusion and Stack Overflow posts in the 2010s, highlighting that while ownership types are brilliant for memory safety, they are hard to understand.</p>
<p id="p-3">By adopting a structured approach to educating the new generation of ownership-savvy programmers, the authors attempt to use the latest educational techniques to explain the dark and confusing corners of the ownership world. This is particularly important because even basic data structures taught in first-year papers, such as circular linked lists, are challenging to master in the Rust world. The issue is not helped by the fact that the actual details of how borrowing checking—the core mechanism behind ownership support in Rust—works are, to this day, defined by an implementation hidden deeply in the compiler. Many brave research project attempts have tried to expose it to the masses.</p>
<p id="p-4">The paper’s core contribution is a conceptual model that reinterprets Rust’s borrow checker in more user-friendly terms. Rather than treating ownership as an abstract rule set, the authors model it dynamically: Each variable and pointer path carries permissions that evolve as code executes. This model bridges the gap between Rust’s static type rules and the programmer’s intuitive understanding of runtime behavior. It makes explicit the cause-and-effect relationships that the compiler enforces silently.</p>
<p id="p-5">To test how well people actually grasp these ideas, the authors built the Ownership Inventory, a diagnostic instrument inspired by “concept inventories” pioneered in the mid-1980s in physics education to study motion. These inventories identify systematic misconceptions—for example, that two references can coexist harmlessly as long as they are used sequentially, or that lifetimes are equivalent to scope. Using data from hundreds of learners, they catalogued the most common misunderstandings of Rust’s ownership system.</p>
<p id="p-6">Armed with this data, the authors designed a new pedagogical framework for teaching ownership. They built a visualization tool that animates the flow of permissions through code, showing precisely how variables become frozen, moved, or invalidated. They then collaborated with “The Rust Book” team to integrate this conceptual model into a rewritten version of the ownership chapter. The revision was deployed to real readers, who could voluntarily complete an interactive survey. The results were telling: Learners using the new materials scored significantly higher on the Ownership Inventory.</p>
<p id="p-7">As our day-to-day programming experience becomes increasingly reliant on complex language constructs and tools, aided by a growing number of LLM-based and similar technologies, it becomes more important to educate new and current generations of software engineers effectively and efficiently. This work lays the groundwork for the expected standard of presentation and user education for many years to come. In doing so, it demonstrates a model for how future language designers might empirically test not only compiler performance but also human comprehension.</p>
<p id="p-8">The ownership types idea began as an attempt to visualize what programmers could not see. Decades later, this paper closes the loop—using visualization once again to illuminate what has long been opaque. It reminds us that even the most elegant type systems succeed only when people can understand them, and that the real frontier of programming-language research may lie not in inventing new rules, but in helping humans grasp the ones we already have.</p>
</section>
</div>
</article>
]]></content:encoded>
					
					<wfw:commentRss>https://cacm.acm.org/research-highlights/technical-perspective-a-grounded-conceptual-model-for-ownership-types-in-rust/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
		<post-id xmlns="com-wordpress:feed-additions:1">778194</post-id>	</item>
		<item>
		<title>Motivation Is All You Need</title>
		<link>https://cacm.acm.org/future-tense/motivation-is-all-you-need/</link>
					<comments>https://cacm.acm.org/future-tense/motivation-is-all-you-need/#respond</comments>
		
		<dc:creator><![CDATA[Banzan]]></dc:creator>
		<pubDate>Thu, 23 Apr 2026 20:49:24 +0000</pubDate>
				<category><![CDATA[Artificial Intelligence and Machine Learning]]></category>
		<category><![CDATA[Society]]></category>
		<guid isPermaLink="false">https://cacm.acm.org/?post_type=digital-library&#038;p=779252</guid>

					<description><![CDATA[<p>Eva felt she could share almost anything with Ed. Maybe it was time to talk to him and see if they could be more than just online friends?</p>]]></description>
										<content:encoded><![CDATA[<article>
<div class="body" lang="en">
<section id="sec1" class="sec">
<p id="p-1"><i>Ding!</i></p>
<p id="p-2">Idly, Eva checked the app notification. To her surprise, someone who wasn’t in her contact list had messaged her out of the blue.</p>
<p id="p-3"><i>Avpr 2 zrrg h!</i></p>
<p id="p-4">“What?!” she cried, baffled. But the mysterious message was strangely intriguing to her. After tinkering with different alternatives for a few minutes, she understood it used a type of substitution cipher, the so-called Caesar’s code, to encode a most ordinary “Nice 2 meet u!”</p>
<p id="p-5"><i>Nice 2 meet u too</i>, Eva wrote back using the same cipher. <i>Who could it be?</i> she wondered, anxiously.</p>
<p id="p-6">The reply was instant: <i>Wow, u were fast</i>!</p>
<p id="p-7">This time Eva grinned, her self-esteem well boosted. By exchanging several messages, she learned the other person, a boy named Ed who was the same age as her, was just looking for friends. In particular, those who were into challenges. “Then let’s give him a proper one,” Eva muttered to herself with a sneer and got to work.</p>
<p id="p-8">It took Ed a while to solve her challenge, but in the end his reply came. Lo and behold, it was encrypted using her own cipher: <i>U good! Now try this.</i></p>
<p id="p-9">The message that followed looked again like some garbled text, but by now Eva knew what she had to do. Armed with the confidence in her math skills, the girl proceeded to crack it.</p>
<p id="p-10"><i>Done</i>, she wrote to Ed after some time, using his algorithm to encrypt her message as proof she had managed to find the solution. She was positively beaming now, filled both with excitement and relief, as she had really needed to use some advanced mathematical tricks this time.</p>
<p id="p-11">Still, Eva felt she had to do more to keep Ed on his toes, and after a few hours she succeeded in coming up with an even better algorithm. <i>Let’s see him puzzle this out</i>, she thought mischievously.</p>
<p id="p-12">Eva was so content with her new algorithm that she sent Ed the source code, challenging him to decrypt the message in the absence of the encryption key—which, of course, she kept to herself.</p>
<p id="p-13">Ed’s boastful reply came the following day. It’s true he had used brute force, trying every possible combination, but he genuinely managed to do it! Perhaps for the first time in her life, Eva felt she had found someone she could look up to.</p>
<p id="p-14">In the days that followed, she was constantly looking forward to messaging Ed. Like any teenager, Eva was keen on protecting her online communication from privy eyes. Hence, creating stronger and stronger algorithms with the boy was thrilling in itself. But more than that, she felt that the way in which she and Ed were challenging each other had connected them in an unusual way—perhaps a very special way.</p>
<p id="p-15">Over time, their messages went from simple confirmations of success to deeper, more elaborate exchanges. Eva learned not only what made Ed happy but also what made him feel insecure, mainly due to the issues he had with his abusive father. That gave her enough courage to share her own thoughts and insecurities with him, including how she was teased at school because of the way she looked. And she was overjoyed when he appeared to understand her so well!</p>
<p id="p-16">Although they had never met, Eva felt that she could share almost anything with the boy. So maybe it was finally time to talk to him and see if they could be more than just online friends? Before doing that, though, she wanted to make sure no one else could spy on their conversations. To make that happen, Eva started using all her spare time working on algorithms and didn’t stop until she had something she really felt she could be proud of.</p>
<p id="p-17"><i>U can use my newest &amp; strongest algo from now on</i>, she messaged Ed once she finished the source code, which she sent as an attachment. Her lips curled in a timid smile. <i>Almost there…</i></p>
<p id="p-18"><i>It’s even better than the previous one?!</i> came the reply.</p>
<p id="p-19"><i>Yup, very strong, don’t think it can be cracked tbh. Proved &amp; tested!</i> Eva typed back.</p>
<p id="p-20">She then took a deep breath. What would he think when he saw her mousy little face? Still, she was willing to run the risk. So, Eva took another deep breath, and, with slightly trembling fingers, she started writing: <i>U ok to video cha—</i></p>
<p id="p-21">* * *</p>
<p id="p-22">Mark sat in front of the computer on which he had just pressed <b>Ctrl</b>+<b>C</b> to stop the execution. His heart was pounding, his face red with excitement. The AI model he had nicknamed EvA, which implemented his “Evolutionary AI” agent approach, had come up with a never-before-seen encryption algorithm. It combined elliptic curve cryptography with a new variation of lattice-based codes in a way that even seemed quantum-resistant. Mark was no security expert, but EvA had actually written a formal proof for it! Therefore, he surmised the algorithm must be strong. Maybe stronger even than any other.</p>
<p id="p-23">Elated, Mark began to relax, the tension in his shoulders slowly uncoiling. Nobody had thought of having two LLMs prompt each other in such a collaborative manner. Regardless, Mark felt it was his other idea, the use of the initial system prompts to give each of them a personality and a backstory, that had truly made a difference.</p>
<p id="p-24">Suddenly, Mark remembered a seminal paper claiming that “attention is all you need” to create powerful AI models. However, he was now convinced that giving LLMs a strong motivation was what had made them perform so well. He smiled as the title of a new paper started floating in front of his eyes: &#8220;Motivation Is All You Need.&#8221; With his novel technique and the experimental results he had gotten so far, having that paper published in a top conference shouldn’t be that hard. That also meant his Ph.D. degree was now within reach.</p>
<p id="p-25">True, in the back of his mind, he felt kind of uncomfortable about some of the messages exchanged by the LLMs. Yet, mentioning how the models seemed to have developed some sort of emotions would ruin his credibility. As a result, Mark felt he might be better off deleting the log lines that suggested sentience. <i>Since it wouldn’t affect the results concerning the encryption algorithm, it should be fine, right?</i></p>
<p id="p-26">After finishing with the log data, the man hesitated for a while, pondering the implications of his actions. Albeit reluctantly, he then proceeded to reset all the model parameters to their defaults, effectively erasing Eva and Ed forever into nothingness. Self-preservation had won.</p>
</section>
</div>
</article>
]]></content:encoded>
					
					<wfw:commentRss>https://cacm.acm.org/future-tense/motivation-is-all-you-need/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
		<post-id xmlns="com-wordpress:feed-additions:1">779252</post-id>	</item>
		<item>
		<title>We Must Teach Computer Science in Prison</title>
		<link>https://cacm.acm.org/opinion/we-must-teach-computer-science-in-prison/</link>
					<comments>https://cacm.acm.org/opinion/we-must-teach-computer-science-in-prison/#respond</comments>
		
		<dc:creator><![CDATA[Emma Hogan Benser]]></dc:creator>
		<pubDate>Thu, 23 Apr 2026 17:57:13 +0000</pubDate>
				<category><![CDATA[Computing Profession]]></category>
		<category><![CDATA[Education]]></category>
		<category><![CDATA[Society]]></category>
		<guid isPermaLink="false">https://cacm.acm.org/?post_type=digital-library&#038;p=780197</guid>

					<description><![CDATA[<p>Expanding CS education in prison is a vote toward a future where people with  experience of incarceration are represented among computing faculty and industry.</p>]]></description>
										<content:encoded><![CDATA[<article>
<div class="body" lang="en">
<section id="sec1" class="sec">
<p id="p-1">During my past three years as a computing education Ph.D. student, I taught computer science courses to 72 incarcerated college students in a prison near San Diego, CA, USA. Prior to this experience, I had never been incarcerated myself or known anyone who had been. Without experiencing prison, we often ask the wrong questions about CS education in prison: “Have you ever been threatened by a student?” “Is it safe to teach them how to use computers?” While many adults have a close family member who has been incarcerated (in the U.S., 64%), the predominantly white or Asian male computing field is significantly less likely to have this experience.<a class="reference-link xref xref-bibr" href="#B5" data-jats-ref-type="bibr" data-jats-rid="B5"><sup>5</sup></a> Those of us who are not directly impacted by the criminal justice system may not (consciously choose to) think about prisons much, but are likely to absorb messaging that people in prison are inherently bad, dangerous, or undeserving of basic human rights.<a class="reference-link xref xref-bibr" href="#B3" data-jats-ref-type="bibr" data-jats-rid="B3"><sup>3</sup></a> For me, it would be difficult, if not impossible, to engage with these misconceptions—let alone grow from them—without coming face to face with people who are incarcerated. I have found that teaching computer science in prison is a particularly meaningful way to do so.</p>
<p id="p-2">As technology increasingly expands into our criminal justice systems, blind spots regarding potential harms are ever more apparent in our homogenous field (for example, deploying machine learning algorithms in criminal sentencing that reinforce racial bias, giving longer sentences to people with darker skin). Engaging directly with incarcerated students can help us to recognize and begin to address these blind spots, while training a new population of technologists with relevant experience to prevent more harm.</p>
</section>
<section id="sec2" class="sec">
<h2 class="heading">The Current System is Broken—For Everyone</h2>
<p id="p-3">In the U.S. and globally, who is criminalized—and the severity of the punishment—is largely dependent on class, race, mental health, level of education, and other social factors.<a class="reference-link xref xref-bibr" href="#B3" data-jats-ref-type="bibr" data-jats-rid="B3"><sup>3</sup></a> Being incarcerated itself also creates additional barriers, such as exclusionary policies in housing and employment. The “revolving door” of prison comes at an enormous human and financial cost, with public spending on incarceration growing at three times the rate of spending on public education since 1990.</p>
<p id="p-4">Higher education in prison (HEP) can be a powerful antidote to the harmful impacts of incarceration. Over the past three years, I have studied how to improve computing HEP under a grant from the National Science Foundation (NSF) and with the support of a Ford Foundation predoctoral fellowship. As part of this work, I taught introductory computing for three cohorts of 20–30 incarcerated college students, and most recently, we offered an upper-level systems course as an elective. As a result, we have learned many lessons from our experiences teaching, and most importantly, feedback from our students.</p>
</section>
<section id="sec3" class="sec">
<h2 class="heading">Lessons Learned So Far</h2>
<p id="p-5"><b>There is a demand for CS education in prison.</b> Within a few weeks of when I started teaching, word spread that I was a computer science professor. On walks across the prison yard to where classes take place, I was joined by people at all stages in their educational journey from basic adult education, GED, and community college who sought me out to express their interest in what I was teaching and their hope to reach the college program in the future to take our CS courses.</p>
<p id="p-6">To begin to address this large demand, I began organizing “Tech Talk” events open to anyone on the yard. So far, computing professors and alumni from our university have offered seven talks on topics including demystifying generative AI, improvements in computer hardware, and racial bias in recently implemented criminal sentencing algorithms. These events have drawn crowds of more than 100 people from the yard—younger people rolling in older attendees in wheelchairs, deaf attendees with a sign language interpreter, college students excited to show guests their Excel-based passion projects on their prison-issued laptops, and others with notebooks and pencils to write down names of new books and terminology. Lively question-and-answer sessions last long past scheduled end times, and there has never been enough time to get every question answered.</p>
<p id="p-7"><b>Technology infrastructure is still a major barrier.</b> In a 1990 computing education experience report detailing a CS bachelor’s degree program in a U.S. prison, James R. Aman wrote about the challenges his students faced due to the 2–3 hours per week of access to the on-site computer lab.<a class="reference-link xref xref-bibr" href="#B2" data-jats-ref-type="bibr" data-jats-rid="B2"><sup>2</sup></a> More than 30 years later, lack of technology infrastructure persists as a major barrier. In the first two iterations of our introductory computing course, all tools to run code were disabled on students’ prison-issued laptops—a significant challenge for both the students and me. Writing and debugging code was laborious and frustrating: students handwrote or typed in a Word document all their code for programming assignments and labs, and the course staff ran their code and sent them the output a few times per week. Students also lacked other critical resources such as a quiet and reliable place to study with peers. Much of our early work focused on strategies to mitigate the effects of these environmental barriers. Drawing on examples such as implementing CS Unplugged activities in Peruvian prisons, we balanced coding assignments with paper-based conceptual exercises (for example, code tracing).<a class="reference-link xref xref-bibr" href="#B4" data-jats-ref-type="bibr" data-jats-rid="B4"><sup>4</sup></a> Students were extremely willing to engage in active learning in lectures, which made strategies like Peer Instruction particularly effective, and created their own strategies for individual and collaborative study outside of class adapting to the intense restrictions.</p>
<p id="p-8">Our work so far has highlighted the importance of increasing access to technology infrastructure. After two years of building our relationship with the program, students in the most recent iteration of the introductory course and our upper-level systems course were given access to a Web-based code interpreter. Global prison policy has made increasing technology and Internet access a key focus: for example, a constitutional court in South Africa recently issued a unanimous ruling that barring incarcerated people to access to personal computers in their cells for educational purposes was unconstitutional.</p>
<p id="p-9"><b>Different motivations create opportunities for tailored interventions.</b> Research on motivational factors of students outside of prison (for example, middle and high school students) has shaped successful interventions to recruit a more diverse group of students into computing (for example, early exposure through block-based programming). Adult students, and specifically those who are incarcerated, have unique needs that must be considered in the future design of such interventions (for example, a psychological need for one’s personal autonomy to be recognized). They also may share different motivations for pursuing computing that could be leveraged: for example, one that I hear from many of my students is the ability to connect with their children on the shared struggle of learning computing! Designing interventions specifically for the incarcerated student population that affirm their identities as adults, parents, and so forth could mutually benefit the broader population of adult (“nontraditional”) students in computing to support of lifelong learning and re-skilling programs.<a class="reference-link xref xref-bibr" href="#B1" data-jats-ref-type="bibr" data-jats-rid="B1"><sup>1</sup></a></p>
</section>
<section id="sec4" class="sec">
<h2 class="heading">Next Steps</h2>
<p id="p-10">Aside from small-scale vocational programs in a handful of prisons, few currently or formerly incarcerated people can access CS education.<a class="reference-link xref xref-bibr" href="#B6" data-jats-ref-type="bibr" data-jats-rid="B6"><sup>6</sup></a> This is not something that will resolve itself without disrupting the current system. Expanding CS education in prison is a vote toward a future where people with lived experience of incarceration are represented among computing faculty and industry, informing our research and practice of computing in and outside of prison. Many individual faculty who have navigated institutional barriers imposed by their university and/or department to teach in prison have noted their contradiction with the stated goals of the institutions themselves. Currently, the majority of faculty who teach in prison do so in addition to their normal course load, and many on a volunteer (unpaid) basis! This is not fair to the faculty <i>or</i> the incarcerated students. The bottom line is that nowhere in our company or university definitions does it say that we discriminate on the basis of incarceration status (which, as stated above, is far from an objective measure, as it is unequally applied to people based on categories such as the color of their skin). University presidents, department chairs, industry leaders, and individuals alike must answer: Should we continue to do this in practice? A small first step for universities could be valuing teaching in prisons in faculty hiring/promotion decisions. Next, we must work toward ensuring this labor is fairly compensated and feasible as part of faculty teaching loads. I acknowledge that this demands a multifaceted approach that likely extends beyond the university (for example, since education programs have been proven to substantially reduce financial costs to prison systems by reducing the number of people who return to prison post-release, could governments and universities work together to incentivize/subsidize costs for additional instructors, or for existing faculty to offer additional sections of their courses in prisons?).</p>
<p id="p-11">On an individual level, there are many ways to get involved in computing education in prison. For those interested in teaching, I recommend checking if your institution or company already has a partnership with local prisons, jails, or reentry networks. Alternatively, I recommend checking with local community colleges. Aside from teaching opportunities, consider supporting organizations such as UnlockEd Labs in the U.S., which works to increase technology infrastructure in prisons around the country. We must also recruit and support formerly incarcerated individuals as graduate students, faculty members, and employees in computing. Finally, I continue to benefit from critical reflection on the messages about incarcerated people that I still receive and discover hidden in my subconscious. When I am able to create awareness of absorbed false narratives justifying their dehumanization, I can recognize more humanity in my students and in myself.</p>
</section>
</div>
<footer class="back"></footer>
</article>
]]></content:encoded>
					
					<wfw:commentRss>https://cacm.acm.org/opinion/we-must-teach-computer-science-in-prison/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
		<post-id xmlns="com-wordpress:feed-additions:1">780197</post-id>	</item>
		<item>
		<title>ACM 2026 General Election</title>
		<link>https://cacm.acm.org/news/acm-2026-general-election/</link>
					<comments>https://cacm.acm.org/news/acm-2026-general-election/#respond</comments>
		
		<dc:creator><![CDATA[]]></dc:creator>
		<pubDate>Thu, 23 Apr 2026 17:42:04 +0000</pubDate>
				<category><![CDATA[Computing Profession]]></category>
		<guid isPermaLink="false">https://cacm.acm.org/?post_type=digital-library&#038;p=779984</guid>

					<description><![CDATA[<p>Meet the candidates running for ACM President, Vice President, and more.</p>]]></description>
										<content:encoded><![CDATA[<article>
<div class="body" lang="en">
<section id="sec1" class="sec">
<p id="p-1">The ACM Constitution provides that our Association hold a general election in the even-numbered years for the positions of President, Vice President, Secretary/Treasurer, and Members-at-Large. Biographical information and statements of the candidates appear below (candidates’ names appear in alphabetical order).</p>
<p id="p-2">In addition to the election of ACM’s officers—President, Vice President, Secretary/Treasurer—two Members at Large will be elected to serve on ACM Council.</p>
<p id="p-3"><b>Electronic Balloting Procedures</b>. Please refer to the instructions posted at <a href="https://www.acm.org/elections/acm-vote">https://www.acm.org/elections/acm-vote</a>. Please note the election email will be addressed from <a href="mailto:acmhelp@mg.electionservicescorp.com"><span class="email">acmhelp@mg.electionservicescorp.com</span></a>.</p>
<p id="p-4"><b>Paper Ballots</b>. Should you wish to vote by paper ballot, please contact Election Services Co. to request a paper copy of the ballot and follow the postal mail ballot procedures: <a href="mailto:acmhelp@electionservicescorp.com"><span class="email">acmhelp@electionservicescorp.com</span></a> or +1-866-720-4357.</p>
<p id="p-5"><b>Postal Mail Ballot Procedures</b>. Please return your ballot in the enclosed envelope, which must be signed by you on the outside in the space provided. The signed ballot envelope may be inserted into a separate envelope for mailing if you prefer this method.</p>
<p id="p-6">All ballots must be received by no later than 16:00 UTC on 22 May 2026. The ACM Elections Committee will validate the computerized tabulation of the ballots. Validation by the Elections Committee will take place at 14:00 UTC on 27 May 2026.</p>
<p id="p-7"><b>ACM Elections Committee<br /></b>Cherri Pancake<br />Panagiotis Papadimitratos<br />Yunhao Liu</p>
</section>
</div>
<h2><strong>Candidates for President</strong></h2>
<section>
<article style="overflow: hidden; margin-bottom: 40px; border-bottom: 1px solid #eee; padding-bottom: 5px;"><img decoding="async" class="alignleft" style="float: left; width: 275px; height: auto; margin: 0px 20px 10px 0px; border: 1px solid #cccccc; border-radius: 4px;" src="https://cacm.acm.org/wp-content/uploads/2026/04/e_bertino_president.jpeg?w=600" alt="Elisa Bertino" /></p>
<h3 id="Bertino">Elisa Bertino</h3>
<p><em>Samuel Conte Distinguished Professor of Computer Science<br />Computer Science Department<br /></em><em>Purdue University<br />West Lafayette, IN, USA</em></p>
</article>
<article style="overflow: hidden; margin-bottom: 5px; border-bottom: 1px solid #eee; padding-bottom: 5px;">
<h4>Biography</h4>
<p>Elisa Bertino is a Samuel Conte Distinguished Professor of Computer Science at Purdue University<span data-path-to-node="4,0,1,2">. She has made pioneering contributions over 40 years to information and systems security and privacy, along with contributions to broadening participation in computing via professional leadership and mentoring</span><span data-path-to-node="4,0,1,6">. </span><span data-path-to-node="4,0,1,8"><span class="citation-1091">Her contributions to data security and privacy include context-based access control, data integrity, privacy-preserving analytics, and data protection from insider threats</span></span><span data-path-to-node="4,0,1,10">. </span><span data-path-to-node="4,0,1,12"><span class="citation-1090">Her recent work focuses on security of cellular networks and IoT systems</span></span><span data-path-to-node="4,0,1,14">.</span></p>
<p><span data-path-to-node="4,0,1,16"><span class="citation-1089">Prior to joining Purdue, she was a professor and department head at the Department of Computer Science of the University of Milan (Italy)</span></span><span data-path-to-node="4,0,1,18">. </span><span data-path-to-node="4,0,1,20"><span class="citation-1088">She has been a postdoc at the IBM Research Laboratory (now Almaden) in San Jose, and a visiting professor at the Singapore Management University, the Singapore National University, and Linköping University</span></span><span data-path-to-node="4,0,1,22">.</span></p>
<p><span data-path-to-node="4,0,1,24"><span class="citation-1087">She has served as editor-in-chief of the IEEE <em>Transactions on Dependable and Secure Computin</em>g, and coordinating co-editor-in-chief of the <em>Very Large Database Systems (VLDB) Journal</em></span></span><span data-path-to-node="4,0,1,26">. </span><span data-path-to-node="4,0,1,28"><span class="citation-1086">She served as Chair of the ACM Special Interest Group on Security, Audit and Control (SIGSAC) for the period 2009-2013</span></span><span data-path-to-node="4,0,1,30">. </span><span data-path-to-node="4,0,1,32"><span class="citation-1085">She is a co-founder of the ACM Conference on Data and Application Security and Privacy (ACM CODASPY)</span></span><span data-path-to-node="4,0,1,34">. </span><span data-path-to-node="4,0,1,36"><span class="citation-1084">The conference started in 2011 and is the main forum for high-quality research on data privacy and security</span></span><span data-path-to-node="4,0,1,38">.</span></p>
<p><span data-path-to-node="4,0,1,40"><span class="citation-1083">Elisa Bertino is a Fellow member of ACM, IEEE, and AAAS</span></span><span data-path-to-node="4,0,1,42">. </span><span data-path-to-node="4,0,1,44"><span class="citation-1082">She received the 2019-2020 ACM Athena Lecturer Award and has been named to GSMA&#8217;s Mobile Security Research Hall of Fame for her work on 4G and 5G cellular network security</span></span><span data-path-to-node="4,0,1,46">. </span><span data-path-to-node="4,0,1,48"><span class="citation-1081">She received the 2014 ACM SIGSAC Outstanding Contribution Award &#8220;For her seminal research contributions and outstanding leadership to Data Security and Privacy for the past 25 years,&#8221; the 2021 IEEE Innovation in Societal Infrastructure Award &#8220;For advancing the security and privacy of new-generation cellular networks,&#8221; and the 2005 IEEE Computer Society Tsutomu Kanai Award for &#8220;Pioneering and innovative research contributions to secure distributed systems.&#8221;</span></span></p>
<h4>Statement</h4>
<p><span data-path-to-node="4,1,1,0"><span class="citation-1080">I have been a member of ACM for over 40 years, and throughout this time ACM has been a vital source of professional growth, collaboration, and community for me and for countless computing professionals worldwide</span></span><span data-path-to-node="4,1,1,2">. </span><span data-path-to-node="4,1,1,4"><span class="citation-1079">I am honored to be nominated for President of ACM</span></span><span data-path-to-node="4,1,1,6">. </span><span data-path-to-node="4,1,1,8"><span class="citation-1078">My service has included chairing the ACM Special Interest Group on Security, Audit and Control (SIGSAC), serving as ACM Secretary/Treasurer, and currently serving as ACM Vice President</span></span><span data-path-to-node="4,1,1,10">. </span><span data-path-to-node="4,1,1,12"><span class="citation-1077">Through these roles, I have gained a deep appreciation for the dedication of ACM&#8217;s volunteers and leaders across the globe, whose collective efforts sustain ACM&#8217;s mission of advancing computing as a science, a profession, and a force for positive change</span></span><span data-path-to-node="4,1,1,14">.</span></p>
<p><span data-path-to-node="4,1,1,14">Computing now stands at a defining moment. </span><span data-path-to-node="4,1,1,16"><span class="citation-1076">Transformative advances in artificial intelligence, data science, cloud and edge infrastructures, ubiquitous sensing, and quantum computing are reshaping research, industry, and society at unprecedented speed and scale</span></span><span data-path-to-node="4,1,1,18">. </span><span data-path-to-node="4,1,1,20"><span class="citation-1075">These developments create extraordinary opportunities for discovery, innovation, and economic growth</span></span><span data-path-to-node="4,1,1,22">. </span><span data-path-to-node="4,1,1,24"><span class="citation-1074">At the same time, they raise profound challenges related to trustworthiness, fairness, accountability, privacy, security, workforce readiness, and environmental sustainability</span></span><span data-path-to-node="4,1,1,26">. </span><span data-path-to-node="4,1,1,28"><span class="citation-1073">Meeting these challenges requires not only continued excellence in foundational research, but also strong professional leadership and sustained dialogue across disciplines, sectors, and regions</span></span><span data-path-to-node="4,1,1,30">.</span></p>
<p><span data-path-to-node="4,1,1,32"><span class="citation-1072">ACM has a unique responsibility—and a unique capacity—to provide that leadership</span></span><span data-path-to-node="4,1,1,34">. </span><span data-path-to-node="4,1,1,36"><span class="citation-1071">As the world&#8217;s premier computing society, ACM is uniquely positioned to convene researchers, practitioners, educators, and policymakers, and to offer rigorous, evidence-based guidance on the responsible design and deployment of computing technologies</span></span><span data-path-to-node="4,1,1,38">. </span><span data-path-to-node="4,1,1,40"><span class="citation-1070">By strengthening its role as a trusted global voice, ACM can help ensure that technical advances are aligned with societal values and the public good</span></span><span data-path-to-node="4,1,1,42">.</span></p>
<p><span data-path-to-node="4,1,1,44"><span class="citation-1069">If elected, I will work to advance this role in several concrete ways</span></span><span data-path-to-node="4,1,1,46">. </span><span data-path-to-node="4,1,1,48"><span class="citation-1068">I will support early-career researchers and educators through enhanced mentorship, recognition, and meaningful opportunities to participate in ACM governance and leadership</span></span><span data-path-to-node="4,1,1,50">. </span><span data-path-to-node="4,1,1,52"><span class="citation-1067">I will promote open, sustainable, and equitable access to publications, data, and research infrastructure, recognizing that broad access is essential to scientific progress and global impact</span></span><span data-path-to-node="4,1,1,54">. </span><span data-path-to-node="4,1,1,56"><span class="citation-1066">I will also champion efforts to expand ACM&#8217;s reach across regions and cultures, strengthening participation from underrepresented communities and fostering a truly diverse ecosystem of ideas</span></span><span data-path-to-node="4,1,1,58">. </span><span data-path-to-node="4,1,1,60"><span class="citation-1065">As computing becomes ever more central to economic and social systems, ACM must play a leading role in preparing a globally skilled workforce capable of innovating responsibly and resiliently</span></span><span data-path-to-node="4,1,1,62">.</span></p>
<p><span data-path-to-node="4,1,1,62">ACM stands at a pivotal moment in its history. </span><span data-path-to-node="4,1,1,64"><span class="citation-1064">I am committed to helping our community lead with integrity, inclusiveness, creativity, and a shared sense of purpose as we work together to shape the future of computing</span></span><span data-path-to-node="4,1,1,66">.</span></p>
</article>
<hr />
<article id="Palsberg" class="candidate">
<article style="overflow: hidden; margin-bottom: 40px; border-bottom: 1px solid #eee; padding-bottom: 5px;"><img decoding="async" class="alignleft" style="float: left; width: 275px; height: auto; margin: 0px 20px 10px 0px; border: 1px solid #cccccc; border-radius: 4px;" src="https://cacm.acm.org/wp-content/uploads/2026/04/j_palsberg_president.jpeg" alt="Jens Palsberg" /></p>
<h3>Jens Palsberg</h3>
<p><em>Professor<br />University of California, Los Angeles (UCLA)<br />Los Angeles, CA, USA</em></p>
</article>
<article style="overflow: hidden; margin-bottom: 5px; border-bottom: 1px solid #eee; padding-bottom: 5px;">
<h4>Biography</h4>
<p><span data-path-to-node="8,0,1,0"><span class="citation-1062">Computer Science Professor with 30 years of experience that includes leading the UCLA Computer Science Department, ACM SIGPLAN, and ACM TOPLAS</span></span><span data-path-to-node="8,0,1,2">. </span><span data-path-to-node="8,0,1,4"><span class="citation-1061">Principal investigator or co-PI on $33 million of research funding from NSF, DARPA, IBM, and others; organizer of POPL, LICS, and SPIN; and program chair of POPL</span></span><span data-path-to-node="8,0,1,6">. </span><span data-path-to-node="8,0,1,8"><span class="citation-1060">Highly cited author of 150 papers and books, advisor of 16 graduated Ph.D. students, and ACM distinguished lecturer</span></span><span data-path-to-node="8,0,1,10">. Recipient of a PLDI distinguished paper award. </span><span data-path-to-node="8,0,1,12"><span class="citation-1059">Recipient of a UCLA teaching award, a Purdue University teaching award, and an ACM SIGPLAN distinguished service award</span></span><span data-path-to-node="8,0,1,14">.</span></p>
<ul>
<li><span data-path-to-node="8,0,1,16"><span class="citation-1058">Ph.D., University of Aarhus, Department of Computer Science, Denmark, 1992</span></span><span data-path-to-node="8,0,1,18">.</span></li>
<li><span data-path-to-node="8,0,1,20"><span class="citation-1057">MBA, Executive MBA Program, UCLA Anderson School of Management, 2017</span></span><span data-path-to-node="8,0,1,22">.</span></li>
<li><span data-path-to-node="8,0,1,24"><span class="citation-1056">Professor, Purdue University, 1996–2003</span></span><span data-path-to-node="8,0,1,26">.</span></li>
<li><span data-path-to-node="8,0,1,28"><span class="citation-1055">Professor, UCLA, 2003<span class="citation-1056">–</span>present</span></span><span data-path-to-node="8,0,1,30">.</span></li>
<li><span data-path-to-node="8,0,1,32"><span class="citation-1054">Secretary/Treasurer and Vice Chair of ACM SIGBED, 2005<span class="citation-1056">–</span>2009</span></span><span data-path-to-node="8,0,1,34">.</span></li>
<li><span data-path-to-node="8,0,1,36"><span class="citation-1053">Editor-in-Chief of <em>ACM TOPLAS</em>, 2010<span class="citation-1056">–</span>2016</span></span><span data-path-to-node="8,0,1,38">.</span></li>
<li><span data-path-to-node="8,0,1,40"><span class="citation-1052">Chair, ACM SIGPLAN, 2018<span class="citation-1056">–</span>2021</span></span><span data-path-to-node="8,0,1,42">.</span></li>
<li><span data-path-to-node="8,0,1,44"><span class="citation-1051">Chair, ACM SIG Governing Board, 2020<span class="citation-1056">–</span>2024</span></span><span data-path-to-node="8,0,1,46">.</span></li>
<li><span data-path-to-node="8,0,1,48"><span class="citation-1050">Member, ACM Executive Committee, 2020<span class="citation-1056">–</span>2024</span></span><span data-path-to-node="8,0,1,50">.</span></li>
<li><span data-path-to-node="8,0,1,52"><span class="citation-1049">Member, ACM Council, 2020<span class="citation-1056">–</span>present</span></span><span data-path-to-node="8,0,1,54">.</span></li>
<li><span data-path-to-node="8,0,1,56"><span class="citation-1048">Member, Executive Committee of the UCLA Institute for Digital Research and Education, 2007<span class="citation-1056">–</span>2020</span></span><span data-path-to-node="8,0,1,58">.</span></li>
<li><span data-path-to-node="8,0,1,60"><span class="citation-1047">Chair of the UCLA Computer Science Department, 2010<span class="citation-1056">–</span>2015</span></span><span data-path-to-node="8,0,1,62">.</span></li>
<li><span data-path-to-node="8,0,1,64"><span class="citation-1046">Chair of the UCLA Council on Planning and Budget, 2018<span class="citation-1056">–</span>2019</span></span><span data-path-to-node="8,0,1,66">.</span></li>
<li><span data-path-to-node="8,0,1,68"><span class="citation-1045">Member, the Ascend Steering Committee, 2019<span class="citation-1056">–</span>2022, which worked on modernizing UCLA&#8217;s financial infrastructure</span></span><span data-path-to-node="8,0,1,70">.</span></li>
<li><span data-path-to-node="8,0,1,72"><span class="citation-1044">Chair of UCLA Senate Advisory Committee on Self-Supporting Graduate Professional Degree Programs, 2023<span class="citation-1056">–</span>present</span></span><span data-path-to-node="8,0,1,74">.</span></li>
<li><span data-path-to-node="8,0,1,76"><span class="citation-1043">Director of the UCLA<span class="citation-1056">–</span>Amazon Science Hub for Humanity and Artificial Intelligence, 2021<span class="citation-1056">–</span>present</span></span><span data-path-to-node="8,0,1,78">.</span></li>
<li><span data-path-to-node="8,0,1,80"><span class="citation-1042">Co-director of the UCLA Center for Quantum Science and Engineering</span></span><span data-path-to-node="8,0,1,82">.</span></li>
<li><span data-path-to-node="8,0,1,82">Member of the editorial boards of ACM Transactions on Quantum Computing 2021<span class="citation-1051"><span class="citation-1056">–</span></span>present; IEEE Transactions on Computers, 2022<span class="citation-1051"><span class="citation-1056">–</span></span>present; Science of Computer Programming, 2020<span class="citation-1051"><span class="citation-1056">–</span></span>present; </span><span data-path-to-node="8,0,1,84"><span class="citation-1041">Information and Computation, 2003<span class="citation-1051"><span class="citation-1056">–</span></span>present</span></span><span data-path-to-node="8,0,1,86">.</span></li>
<li><span data-path-to-node="8,0,1,88"><span class="citation-1040">Member of 24 National Science Foundation review panels</span></span><span data-path-to-node="8,0,1,90">.</span></li>
<li><span data-path-to-node="8,0,1,92"><span class="citation-1039">Instructor of UCLA courses on quantum computing and compiler construction</span></span><span data-path-to-node="8,0,1,94">.</span></li>
<li><span data-path-to-node="8,0,1,96"><span class="citation-1038">Expert witness who was deposed three times in litigation cases.</span></span></li>
</ul>
<h4>Statement</h4>
<p><span data-path-to-node="8,1,1,0"><span class="citation-1037">ACM is an organization that is dear to my heart</span></span><span data-path-to-node="8,1,1,2">. </span><span data-path-to-node="8,1,1,4"><span class="citation-1036">I see ACM as my intellectual home: It organizes the conferences I attend, publishes the research I rely on, and connects me with outstanding colleagues across computing</span></span><span data-path-to-node="8,1,1,6">. </span><span data-path-to-node="8,1,1,8"><span class="citation-1035">I prioritize giving back to ACM by volunteering wherever I can and encouraging others to get involved</span></span><span data-path-to-node="8,1,1,10">.</span></p>
<p><span data-path-to-node="8,1,1,12"><span class="citation-1034">I am running for President because I want to strengthen the global computing community</span></span><span data-path-to-node="8,1,1,14">. </span><span data-path-to-node="8,1,1,16"><span class="citation-1033">As computing takes on an increasingly central role in society, we must reinforce the connections among the diverse people and groups that make up ACM and ensure that every member feels welcome, valued, and able to make a meaningful impact</span></span><span data-path-to-node="8,1,1,18">. </span><span data-path-to-node="8,1,1,20"><span class="citation-1032">Stronger ties within our community will help us maintain high standards, give us the capacity to grow thoughtfully, and increase the enjoyment and pride that come from participating in ACM</span></span><span data-path-to-node="8,1,1,22">.</span></p>
<p><span data-path-to-node="8,1,1,24"><span class="citation-1031">I will continue ACM&#8217;s current efforts on open science, globalization, and responsible computing</span></span><span data-path-to-node="8,1,1,26">. </span><span data-path-to-node="8,1,1,28"><span class="citation-1030">I will also prioritize the affordability of ACM conferences, our response to the impact of AI across ACM, and our push to lead in new areas, such as quantum computing</span></span><span data-path-to-node="8,1,1,30">. </span><span data-path-to-node="8,1,1,32"><span class="citation-1029">In addition, I will launch the following three initiatives</span></span><span data-path-to-node="8,1,1,34">.</span></p>
<p><span data-path-to-node="8,1,1,36"><span class="citation-1028"><strong>Increase communication</strong>: I will take every opportunity to tell the community what we are doing and to respond to concerns and ideas</span></span><span data-path-to-node="8,1,1,38">. </span><span data-path-to-node="8,1,1,40"><span class="citation-1027">As we improve our digital library, select conference locations, and combat collusion rings, I will write blog posts, host town hall meetings, invite feedback, and encourage the leaders of all units of ACM to do the same</span></span><span data-path-to-node="8,1,1,42">. </span><span data-path-to-node="8,1,1,44"><span class="citation-1026">More communication will help make ACM more open and cohesive, while giving everyone a voice</span></span><span data-path-to-node="8,1,1,46">.</span></p>
<p><span data-path-to-node="8,1,1,48"><span class="citation-1025"><strong>Support ACM volunteers</strong>: ACM depends on the dedicated work of volunteers, but too many of them are overworked and experience burnout</span></span><span data-path-to-node="8,1,1,50">. </span><span data-path-to-node="8,1,1,52"><span class="citation-1024">I will address this by encouraging and recruiting new volunteers, and by giving them better tool support</span></span><span data-path-to-node="8,1,1,54">. </span><span data-path-to-node="8,1,1,56"><span class="citation-1023">I will increase awareness of how all members can participate in entry-level volunteer roles, strengthen the volunteer pipeline, and introduce frequent-volunteer points to visibly recognize sustained service and encourage broader participation</span></span><span data-path-to-node="8,1,1,58">.</span></p>
<p><span data-path-to-node="8,1,1,60"><span class="citation-1022"><strong>Engage the next generation</strong>: For ACM to continue to thrive, we must bolster our efforts to recruit, retain, and support students</span></span><span data-path-to-node="8,1,1,62">. </span><span data-path-to-node="8,1,1,64"><span class="citation-1021">I will expand ACM student groups at universities and colleges worldwide by inviting students to found new chapters, leveraging nearby ACM groups for support, and partnering with local computing departments</span></span><span data-path-to-node="8,1,1,66">. </span><span data-path-to-node="8,1,1,68"><span class="citation-1020">Stronger student chapters will help students connect their local activities to the global ACM community and encourage continued membership after graduation</span></span><span data-path-to-node="8,1,1,70">. </span><span data-path-to-node="8,1,1,72"><span class="citation-1019">I will also create an ACM-wide mentoring network, modeled on successful efforts in some of ACM&#8217;s special interest groups</span></span><span data-path-to-node="8,1,1,74">.</span></p>
<p><span data-path-to-node="8,1,1,76"><span class="citation-1018">Through these community initiatives, we will lay the foundation for ACM&#8217;s continued excellence</span></span><span data-path-to-node="8,1,1,78">.</span></p>
</article>
</article>
</section>
<hr />
<section>
<h2>Candidates for Vice President</h2>
<article style="overflow: hidden; margin-bottom: 40px; border-bottom: 1px solid #eee; padding-bottom: 5px;"><img decoding="async" class="alignleft" style="float: left; width: 275px; height: auto; margin: 0 20px 10px 0; border: 1px solid #ccc; border-radius: 4px;" src="https://cacm.acm.org/wp-content/uploads/2026/04/a_deshpande_vicepresident.jpeg" alt="Anand Deshpande" /></p>
<h3 id="Deshpande">Anand Deshpande</h3>
<p><em>Founder and Chairman</em><br /><em>Persistent Systems</em><br /><em>Pune, India</em></p>
</article>
<article style="overflow: hidden; margin-bottom: 5px; border-bottom: 1px solid #eee; padding-bottom: 5px;">
<h4>Biography</h4>
<p><span data-path-to-node="13,0,1,0"><span class="citation-1016">Anand Deshpande is a technologist, entrepreneur, and community leader with more than 35 years of experience building and scaling global technology organizations</span></span><span data-path-to-node="13,0,1,2">. </span><span data-path-to-node="13,0,1,4"><span class="citation-1015">He is the Founder and Chairman of Persistent Systems, a digital engineering and enterprise modernization company with more than 26,000 professionals</span></span><span data-path-to-node="13,0,1,6">. </span><span data-path-to-node="13,0,1,8"><span class="citation-1014">He founded Persistent in 1990 after a short stint at Hewlett-Packard Laboratories in Palo Alto, CA</span></span><span data-path-to-node="13,0,1,10">. </span><span data-path-to-node="13,0,1,12"><span class="citation-1013">Deshpande holds a Ph.D. in Computer Science from Indiana University and a B Tech in Computer Science from IIT Kharagpur</span></span><span data-path-to-node="13,0,1,14">.</span></p>
<p><span data-path-to-node="13,0,1,16"><span class="citation-1012">He has been a member since 1984 and has deeply engaged with ACM and the global research community</span></span><span data-path-to-node="13,0,1,18">. </span><span data-path-to-node="13,0,1,20"><span class="citation-1011">As the founding President of the ACM India Council (2010<span class="citation-1051"><span class="citation-1056">–</span></span>2012), he helped establish ACM&#8217;s presence and programs across India and has continued to support ACM initiatives in research, education, and community building</span></span><span data-path-to-node="13,0,1,22">. </span><span data-path-to-node="13,0,1,24"><span class="citation-1010">For his sustained service and leadership, he received the 2024 ACM Presidential Award</span></span><span data-path-to-node="13,0,1,26">. </span><span data-path-to-node="13,0,1,28"><span class="citation-1009">He completed his term as a Trustee of the VLDB Endowment in 2025 and has served as Chairman of the Boards of IIT Patna, IIIT Allahabad, and VJTI Mumbai</span></span><span data-path-to-node="13,0,1,30">.</span></p>
<p><span data-path-to-node="13,0,1,32"><span class="citation-1008">Deshpande plays a key role in strengthening India&#8217;s interdisciplinary research capacity through his work with BRIC—Biotechnology Research and Innovation Council—fostering collaboration across biology, AI, engineering, and data-centric science</span></span><span data-path-to-node="13,0,1,34">.</span></p>
<p><span data-path-to-node="13,0,1,36"><span class="citation-1007">Beyond research, he has contributed significantly to India&#8217;s innovation and entrepreneurship landscape</span></span><span data-path-to-node="13,0,1,38">. </span><span data-path-to-node="13,0,1,40"><span class="citation-1006">Through the deAsra Foundation, he has enabled more than 500,000 nano-entrepreneurs and through the Second Orbit program, he leads scaling programs for thousands of mid-sized technology companies</span></span><span data-path-to-node="13,0,1,42">.</span></p>
<p><span data-path-to-node="13,0,1,44"><span class="citation-1005">Deshpande is a Fellow of the Indian National Academies of Science and Engineering</span></span><span data-path-to-node="13,0,1,46">. </span><span data-path-to-node="13,0,1,48"><span class="citation-1004">His experience building and guiding a large multinational technology organization—and his broad engagement across industry, research, and academia—positions him well to help ACM enhance its global relevance, expand participation, and navigate the opportunities and responsibilities of an AI- and data-driven future</span></span><span data-path-to-node="13,0,1,50">.</span></p>
<h4>Statement</h4>
<p><span data-path-to-node="13,1,1,0"><span class="citation-1003">ACM is a strong and respected global institution, and I deeply value its contribution across research, education, publications, and professional communities</span></span><span data-path-to-node="13,1,1,2">. </span><span data-path-to-node="13,1,1,4"><span class="citation-1002 interactive-span-hovered">As Vice President, my priority will be to strengthen these foundations and ensure that ACM remains a truly member-driven organization—of the members, by the members, and for the members</span></span><span data-path-to-node="13,1,1,6">.</span></p>
<p><span data-path-to-node="13,1,1,8"><span class="citation-1001">ACM is an exceptional platform for global impact, and I will focus on increasing membership by making ACM increasingly relevant to researchers, practitioners, educators, and professionals, especially in emerging regions, by demonstrating clear, differentiated value for every member segment</span></span><span data-path-to-node="13,1,1,10">. </span><span data-path-to-node="13,1,1,12"><span class="citation-1000">We must also inspire the next generation of computing professionals to see ACM as their intellectual home and as a community that advances both their careers and the discipline</span></span><span data-path-to-node="13,1,1,14">.</span></p>
<p><span data-path-to-node="13,1,1,16"><span class="citation-999">The ACM Digital Library&#8217;s transition to open access is a bold and important milestone</span></span><span data-path-to-node="13,1,1,18">. </span><span data-path-to-node="13,1,1,20"><span class="citation-998">To sustain this model while preserving quality, independence, and long-term stability, ACM must develop innovative and durable financial models, engage new constituencies, and expand the ways members contribute to and benefit from ACM</span></span><span data-path-to-node="13,1,1,22">. </span><span data-path-to-node="13,1,1,24"><span class="citation-997">Drawing on my experience leading a multinational technology organization, I look forward to collaborating with fellow ACM leadership to address this strategic shift</span></span><span data-path-to-node="13,1,1,26">.</span></p>
<p><span data-path-to-node="13,1,1,26">The nature of research is evolving. </span><span data-path-to-node="13,1,1,28"><span class="citation-996">Interdisciplinary work and team science are increasingly prioritized by funding agencies and by industry</span></span><span data-path-to-node="13,1,1,30">. </span><span data-path-to-node="13,1,1,32"><span class="citation-995">I will encourage greater engagement with adjacent fields and explore ways to co-locate and collaborate with communities beyond computing</span></span><span data-path-to-node="13,1,1,34">.</span></p>
<p><span data-path-to-node="13,1,1,36"><span class="citation-994">Continuous learning is essential in the era of rapid technological change</span></span><span data-path-to-node="13,1,1,38">. </span><span data-path-to-node="13,1,1,40"><span class="citation-993">My experience in building global talent pipelines reinforces my belief that ACM can play an even larger role in lifelong learning</span></span><span data-path-to-node="13,1,1,42">.</span></p>
<p><span data-path-to-node="13,1,1,44"><span class="citation-992">Finally, ACM must be a catalyst for entrepreneurship, innovation, and societal impact</span></span><span data-path-to-node="13,1,1,46">. </span><span data-path-to-node="13,1,1,48"><span class="citation-991">As a leader in entrepreneurial ecosystems, I appreciate what it takes to help ideas move from research into products, companies, and policy</span></span><span data-path-to-node="13,1,1,50">.</span></p>
<p><span data-path-to-node="13,1,1,52"><span class="citation-990">We must deepen ACM&#8217;s engagement with governments and policy think tanks worldwide so that our members are recognized as trusted voices in shaping the future of technology and its societal consequences</span></span><span data-path-to-node="13,1,1,54">.</span></p>
<p><span data-path-to-node="13,1,1,56"><span class="citation-989">ACM&#8217;s greatest strength is its people, and together we can shape a future where computing empowers all</span></span><span data-path-to-node="13,1,1,58">.</span></p>
</article>
<hr />
<article style="overflow: hidden; margin-bottom: 40px; border-bottom: 1px solid #eee; padding-bottom: 5px;"><img decoding="async" class="alignleft" style="float: left; width: 275px; height: auto; margin: 0 20px 10px 0; border: 1px solid #ccc; border-radius: 4px;" src="https://cacm.acm.org/wp-content/uploads/2026/04/r_mohan_vicepresident.jpeg" alt="Rashmi Mohan" /></p>
<h3 id="Mohan">Rashmi Mohan</h3>
<p><em>Director of Engineering</em><br /><em>Cisco (Splunk) Inc.</em><br /><em>San Francisco, CA, USA</em></p>
</article>
<article style="overflow: hidden; margin-bottom: 5px; border-bottom: 1px solid #eee; padding-bottom: 5px;">
<h4>Biography</h4>
<p><span data-path-to-node="17,0,1,0"><span class="citation-987">Rashmi Mohan is currently a Director of Engineering at Cisco (Splunk) Inc., leading the Enterprise Security SIEM Investigations and Foundations engineering teams</span></span><span data-path-to-node="17,0,1,2">. </span><span data-path-to-node="17,0,1,4"><span class="citation-986">As a technology leader, she has managed, led, and grown teams of engineers and helped deliver complex products both on-prem and in the cloud in the security data analytics domain</span></span><span data-path-to-node="17,0,1,6">.</span></p>
<p><span data-path-to-node="17,0,1,8"><span class="citation-985">Previously, as co-founder at EnTrio Partners, she worked with start-ups on developing their product ideas while also assisting larger companies with their digital transformation journeys</span></span><span data-path-to-node="17,0,1,10">. </span><span data-path-to-node="17,0,1,12"><span class="citation-984">Prior to that, she was a Sr. Engineering Manager at Yahoo Labs and led a group of scientists and engineers</span></span><span data-path-to-node="17,0,1,14">. </span><span data-path-to-node="17,0,1,16"><span class="citation-983">She has spent 25+ years in various technical and management roles in the industry</span></span><span data-path-to-node="17,0,1,18">. </span><span data-path-to-node="17,0,1,20"><span class="citation-982">She previously worked in the User Data and Analytics group at Yahoo! working on reporting applications</span></span><span data-path-to-node="17,0,1,22">.</span></p>
<p><span data-path-to-node="17,0,1,24"><span class="citation-981">Rashmi has been on the ACM executive council as a Secretary/Treasurer since 2024, a Member at Large since 2022, on the ACM Practitioner Board, and is the founding member and host of the popular podcast series ACM <em>Bytecast</em></span></span><span data-path-to-node="17,0,1,26">. </span><span data-path-to-node="17,0,1,28"><span class="citation-980">Through the podcasts, she brings out the inspiring stories of eminent practitioners with the goal of highlighting their monumental work and showcasing diverse paths to success</span></span><span data-path-to-node="17,0,1,30">. </span><span data-path-to-node="17,0,1,32"><span class="citation-979">She is co-leading the Youthification charter as a part of the PTF and serving on the ACM AI Summit planning committee, specifically around AI in the workforce</span></span><span data-path-to-node="17,0,1,34">. </span><span data-path-to-node="17,0,1,36"><span class="citation-978">She has served on the ACM India council (2014-2018) as a member at large and as the ACM India council secretary (2016-2018)</span></span><span data-path-to-node="17,0,1,38">. </span><span data-path-to-node="17,0,1,40"><span class="citation-977">She actively participated in the ACM eminent speaker series, addressing ACM student chapters and ACM-W chapters across the country</span></span><span data-path-to-node="17,0,1,42">.</span></p>
<p><span data-path-to-node="17,0,1,44"><span class="citation-976">Rashmi has previously been associated with the Grace Hopper Conference in India for five years as a mentor, advisory committee member, Program Chair, and founding member of the all-women Hackathon</span></span><span data-path-to-node="17,0,1,46">. </span><span data-path-to-node="17,0,1,48"><span class="citation-975">Rashmi was also a mentor at the Oracle Startup Cloud Accelerator and Google Launchpad</span></span><span data-path-to-node="17,0,1,50">. </span><span data-path-to-node="17,0,1,52"><span class="citation-974">She is a prolific public speaker, including a widely publicized talk at TEDx Chennai</span></span><span data-path-to-node="17,0,1,54">. </span><span data-path-to-node="17,0,1,56"><span class="citation-973">Rashmi has a Bachelor&#8217;s degree in Computer Engineering from Santa Clara University</span></span><span data-path-to-node="17,0,1,58">.</span></p>
<h4>Statement</h4>
<p><span data-path-to-node="17,1,1,0"><span class="citation-972">I am truly honored and humbled to be considered for the position of Vice President of ACM</span></span><span data-path-to-node="17,1,1,2">.</span></p>
<p><span data-path-to-node="17,1,1,4"><span class="citation-971">My 12-year journey with ACM has mirrored my professional career, evolving in scope, impact, and responsibility</span></span><span data-path-to-node="17,1,1,6">. </span><span data-path-to-node="17,1,1,8"><span class="citation-970">Through leadership roles as ACM Secretary/Treasurer, serving on the ACM India Council and the ACM Practitioner Board—as well as hosting the ACM <em>Bytecast</em>—I have gained a holistic understanding of our diverse membership</span></span><span data-path-to-node="17,1,1,10">. </span><span data-path-to-node="17,1,1,12"><span class="citation-969">I am dedicated to synthesizing my industry expertise with these volunteer experiences to provide strategic perspectives that advance ACM&#8217;s global initiatives</span></span><span data-path-to-node="17,1,1,14">.</span></p>
<p><span data-path-to-node="17,1,1,16"><span class="citation-968">My goals as Vice President would be to:</span></span></p>
<ul>
<li><span data-path-to-node="17,1,1,19"><span class="citation-967">Invest in the future of computing by creating meaningful avenues for youth to collaborate, innovate ethically, and lead interdisciplinary research</span></span><span data-path-to-node="17,1,1,21">. </span><span data-path-to-node="17,1,1,23"><span class="citation-966">As a leader on the Youthification ACM presidential task force, my goal will be to understand and highlight the needs and voices of this section of our community</span></span><span data-path-to-node="17,1,1,25">.</span></li>
<li><span data-path-to-node="17,1,1,27"><span class="citation-965">Make &#8216;AI augmented Workplace&#8217; a central pillar for ACM and provide our members with a roadmap for career longevity in an automated world</span></span><span data-path-to-node="17,1,1,29">. </span><span data-path-to-node="17,1,1,31"><span class="citation-964">Create forums where students and practitioners can master these emerging paradigms, from &#8220;vibe coding&#8221; to agentic workflows; share best practices for cross-functional collaboration; and ensure that AI remains a tool for human empowerment</span></span><span data-path-to-node="17,1,1,33">.</span></li>
<li><span data-path-to-node="17,1,1,35"><span class="citation-963">Elevate ACM&#8217;s technical excellence by showcasing diverse thought leadership and pioneering research through streamlined, engaging content</span></span><span data-path-to-node="17,1,1,37">. </span><span data-path-to-node="17,1,1,39"><span class="citation-962">By translating complex breakthroughs into digestible insights, we can inspire a broader spectrum of our community and provide clear roadmaps for professional growth</span></span><span data-path-to-node="17,1,1,41">. </span><span data-path-to-node="17,1,1,43"><span class="citation-961">Use ACM&#8217;s myriad content-delivery mechanisms to bridge the gap between academic research and industry application</span></span><span data-path-to-node="17,1,1,45">.</span></li>
<li><span data-path-to-node="17,1,1,47"><span class="citation-960">Create a mentorship model that facilitates a bi-directional, easy exchange of ideas and learnings between all sections of our community</span></span><span data-path-to-node="17,1,1,49">. </span><span data-path-to-node="17,1,1,51"><span class="citation-959">Focus on two-way value, where senior members provide high-level career guidance and architectural wisdom, while our youth offer hands-on insights into emerging dev stacks, agentic AI tools, and the evolving ethics of the modern workplace</span></span><span data-path-to-node="17,1,1,53">.</span></li>
</ul>
<p><span data-path-to-node="17,1,1,55"><span class="citation-958">My experience leading organizations through change and adoption of new ideas and technologies, paired with my tenure being ACM Secretary/Treasurer, will assist me in playing this role effectively, if elected</span></span><span data-path-to-node="17,1,1,57">. </span><span data-path-to-node="17,1,1,59"><span class="citation-957">Thank you for the opportunity and I look forward to continuing to serve the ACM community</span></span><span data-path-to-node="17,1,1,61">.</span></p>
</article>
</section>
<hr />
<section>
<h2>Candidates for Secretary/Treasurer</h2>
<article style="overflow: hidden; margin-bottom: 40px; border-bottom: 1px solid #eee; padding-bottom: 5px;"><img decoding="async" class="alignleft" style="float: left; width: 275px; height: auto; margin: 0 20px 10px 0; border: 1px solid #ccc; border-radius: 4px;" src="https://cacm.acm.org/wp-content/uploads/2026/04/t_crick_secretary-treasurer.jpeg" alt="Tom Crick" /></p>
<h3 id="Crick">Tom Crick</h3>
<p><em>Professor of Digital Society and Policy<br />University of Bristol<br />Bristol, U.K.</em></p>
</article>
<article style="overflow: hidden; margin-bottom: 5px; border-bottom: 1px solid #eee; padding-bottom: 5px;">
<h4>Biography</h4>
<p><span data-path-to-node="22,0,1,0"><span class="citation-955">Professor Tom Crick is Professor of Digital Society and Policy at the University of Bristol and Chief Scientific Adviser at the U.K. Government&#8217;s Department for Culture, Media and Sport</span></span><span data-path-to-node="22,0,1,2">. </span><span data-path-to-node="22,0,1,4"><span class="citation-954">Trained as a computer scientist, his academic work spans computing, data, and artificial intelligence, with particular emphasis on the societal, cultural, economic, and institutional impacts of digital technologies</span></span><span data-path-to-node="22,0,1,6">.</span></p>
<p><span data-path-to-node="22,0,1,8"><span class="citation-953">His research portfolio is interdisciplinary and international, addressing digital governance, responsible AI, open science, and computing education</span></span><span data-path-to-node="22,0,1,10">. </span><span data-path-to-node="22,0,1,12"><span class="citation-952">His work has been supported by major public research funders and has informed policy, professional practice, and institutional reform across multiple jurisdictions</span></span><span data-path-to-node="22,0,1,14">. </span><span data-path-to-node="22,0,1,16"><span class="citation-951">He has published extensively and holds senior editorial and advisory roles within the computing research community</span></span><span data-path-to-node="22,0,1,18">. </span><span data-path-to-node="22,0,1,20"><span class="citation-950">He has also chaired major national curriculum reviews in computing and STEM education</span></span><span data-path-to-node="22,0,1,22">. </span><span data-path-to-node="22,0,1,24"><span class="citation-949">As a Chief Scientific Adviser, he provides independent scientific and technical advice to ministers and senior officials, overseeing research and evidence use, developing and implementing science strategy, and fostering responsible AI adoption within complex public systems</span></span><span data-path-to-node="22,0,1,26">. </span><span data-path-to-node="22,0,1,28"><span class="citation-948">His role includes leadership on R&amp;D prioritization, assurance, and governance, working closely with public funders, industry, and academia</span></span><span data-path-to-node="22,0,1,30">.</span></p>
<p><span data-path-to-node="22,0,1,32"><span class="citation-947">Professor Crick has extensive experience in governance and financial oversight across academic, professional, and public institutions, including board-level service in multi-billion-pound, highly regulated sectors</span></span><span data-path-to-node="22,0,1,34">. </span><span data-path-to-node="22,0,1,36"><span class="citation-946">He is a long-standing elected Member-at-Large of the ACM Council, contributing to Council deliberations on governance, finance, and institutional priorities, and has supported ACM activities spanning publications, conferences, professional practice, and financial sustainability</span></span><span data-path-to-node="22,0,1,38">. </span><span data-path-to-node="22,0,1,40"><span class="citation-945">He is a Fellow of multiple professional and learned societies and has received international recognition for contributions to computing education and digital skills</span></span><span data-path-to-node="22,0,1,42">.</span></p>
<h4>Statement</h4>
<p><span data-path-to-node="22,1,1,0"><span class="citation-944">I am standing for Secretary/Treasurer to further support the ACM in building and maintaining strong governance, financial resilience, and institutional trust during a period of sustained disruption and change</span></span><span data-path-to-node="22,1,1,2">.</span></p>
<p><span data-path-to-node="22,1,1,4"><span class="citation-943">ACM, like many global professional societies, operates in an environment shaped by rapid technological change, evolving publication and conference models, financial uncertainty, legislative/regulatory pressures, and shifting expectations from a diverse international membership</span></span><span data-path-to-node="22,1,1,6">. </span><span data-path-to-node="22,1,1,8"><span class="citation-942">Addressing these challenges requires disciplined stewardship of resources, transparent decision making, and a long-term commitment to institutional sustainability</span></span><span data-path-to-node="22,1,1,10">.</span></p>
<p><span data-path-to-node="22,1,1,12"><span class="citation-941">My professional experience spans both senior leadership and non-executive governance roles across academia, government, and highly regulated economic sectors, with direct responsibility for oversight of complex budgets, audit and risk frameworks, and assurance processes</span></span><span data-path-to-node="22,1,1,14">. </span><span data-path-to-node="22,1,1,16"><span class="citation-940">This includes stewardship of multi-million-pound national research and innovation portfolios, responsible AI adoption, and organizational decision making under public and regulatory scrutiny</span></span><span data-path-to-node="22,1,1,18">.</span></p>
<p><span data-path-to-node="22,1,1,20"><span class="citation-939">As a long-standing elected Member-at-Large of ACM Council and a member of the ACM Presidential Taskforce on Financial Models, I have seen how sound financial management and effective governance enable ACM&#8217;s core missions: supporting members worldwide, sustaining high-quality publications and conferences, and investing responsibly in future initiatives</span></span><span data-path-to-node="22,1,1,22">. </span><span data-path-to-node="22,1,1,24"><span class="citation-938">If elected, I would work constructively with Council, the Executive Committee, and ACM staff (and naturally, ACM members) to ensure that ACM remains financially robust, well governed, and able to serve the global computing community with confidence and integrity</span></span><span data-path-to-node="22,1,1,26">.</span></p>
</article>
<hr />
<article style="overflow: hidden; margin-bottom: 40px; border-bottom: 1px solid #eee; padding-bottom: 5px;"><img decoding="async" class="alignleft" style="float: left; width: 275px; height: 275px; margin: 0px 20px 10px 0px; border: 1px solid #cccccc; border-radius: 4px;" src="https://cacm.acm.org/wp-content/uploads/2026/04/J_haritsa_secretary_treasurer.jpeg" alt="Jayant Haritsa" /></p>
<h3 id="Haritsa">Jayant R Haritsa</h3>
<p><em>Senior Professor</em><br /><em>Dept. of Computational &amp; Data Sciences</em><br /><em>Indian Institute of Science</em><br /><em>Bengaluru, India</em></p>
</article>
<article style="overflow: hidden; margin-bottom: 5px; border-bottom: 1px solid #eee; padding-bottom: 5px;">
<h4>Biography</h4>
<p><span data-path-to-node="26,0,1,0"><span class="citation-936">Jayant Haritsa has served on the computer science faculty at the Indian Institute of Science (IISc), Bangalore, for more than three decades</span></span><span data-path-to-node="26,0,1,2">. </span><span data-path-to-node="26,0,1,4"><span class="citation-935 interactive-span-hovered">He has made foundational contributions to the design, implementation, and evaluation of relational database engines, which underpin modern enterprise information systems</span></span><span data-path-to-node="26,0,1,6">. </span><span data-path-to-node="26,0,1,8"><span class="citation-934">He holds a Bachelor&#8217;s degree in electrical engineering from IIT Madras, and MS and Ph.D. degrees in computer science from the University of Wisconsin-Madison</span></span><span data-path-to-node="26,0,1,10">. </span><span data-path-to-node="26,0,1,12"><span class="citation-933">During his academic career, he has spent industry sabbaticals at Lucent Bell Labs, IBM Research, and Microsoft Research</span></span><span data-path-to-node="26,0,1,14">.</span></p>
<p><span data-path-to-node="26,0,1,16"><span class="citation-932">Haritsa is a Fellow of both ACM and IEEE, and his research has received awards at premier international database conferences</span></span><span data-path-to-node="26,0,1,18">. </span><span data-path-to-node="26,0,1,20"><span class="citation-931">He has developed a suite of software tools, including the widely used Picasso query optimizer visualizer, that have received a warm reception from both academic researchers and the database industry</span></span><span data-path-to-node="26,0,1,22">. </span><span data-path-to-node="26,0,1,24"><span class="citation-930">He was the Program Co-Chair of ICDE 2010 and VLDB 2016, and is currently General Co-Chair for ACM SIGMOD 2026</span></span><span data-path-to-node="26,0,1,26">.</span></p>
<p><span data-path-to-node="26,0,1,28"><span class="citation-929">Within India, Haritsa has played a leadership role in establishing broad visibility for the country&#8217;s database research, authoring early papers in major ACM venues, including SIGMOD and CIKM, which now regularly feature an Indian presence</span></span><span data-path-to-node="26,0,1,30">. </span><span data-path-to-node="26,0,1,32"><span class="citation-928">He is a Fellow of all four Indian science and engineering academies, and a recipient of the highest national recognitions, including Swarnajayanti Fellowship (young achiever award), Shanti Swarup Bhatnagar Award (scientific excellence award), and Infosys Prize (research impact award)</span></span><span data-path-to-node="26,0,1,34">. </span><span data-path-to-node="26,0,1,36"><span class="citation-927">He is also a Distinguished Alumnus of IIT Madras</span></span><span data-path-to-node="26,0,1,38">.</span></p>
<p><span data-path-to-node="26,0,1,40"><span class="citation-926">An ACM member since 1989, Haritsa was elected to the ACM India Council in 2018 and became its President during 2020<span class="citation-1051"><span class="citation-1056">–</span></span>2022</span></span><span data-path-to-node="26,0,1,42">. </span><span data-path-to-node="26,0,1,44"><span class="citation-925">During his tenure, he led a comprehensive restructuring of the Council&#8217;s operations, strengthening engagement, effectiveness, and accountability</span></span><span data-path-to-node="26,0,1,46">. </span><span data-path-to-node="26,0,1,48"><span class="citation-924">At the global level, Haritsa has contributed to enhancing ACM&#8217;s international presence, has recently served on the editorial board of ACM Books, and is currently co-chairing the Presidential Task Force on Regional Offices</span></span><span data-path-to-node="26,0,1,50">.</span></p>
<h4>Statement</h4>
<p><span data-path-to-node="26,1,1,0"><span class="citation-923">From my vantage location in Bangalore, a global computing hub, I have witnessed profound emerging shifts in both the pedagogy and practice of computer science</span></span><span data-path-to-node="26,1,1,2">. </span><span data-path-to-node="26,1,1,4"><span class="citation-922">To successfully navigate these simultaneously exciting and challenging times, ACM must reimagine its programs, processes, and governance</span></span><span data-path-to-node="26,1,1,6">. </span><span data-path-to-node="26,1,1,8"><span class="citation-921">If elected, I would prioritize the following issues</span></span><span data-path-to-node="26,1,1,11"><span class="citation-920">:</span></span></p>
<p><span data-path-to-node="26,1,1,11"><span class="citation-920"><strong>PTF recommendations</strong>: Ten Presidential Task Forces (PTFs), spanning the spectrum of ACM activities, were constituted by recent ACM Councils, and have produced cogent proposals</span></span><span data-path-to-node="26,1,1,13">. </span><span data-path-to-node="26,1,1,15"><span class="citation-919">My close association with two PTFs—Bylaws as a member, and Regional Offices as Co-chair—has provided insight into how these recommendations can be implemented effectively</span></span><span data-path-to-node="26,1,1,17">.</span></p>
<p><span data-path-to-node="26,1,1,19"><span class="citation-918"><strong>Member participation</strong>: Despite ACM&#8217;s large and diverse membership, participation in governance and elections has historically been low</span></span><span data-path-to-node="26,1,1,21">. </span><span data-path-to-node="26,1,1,23"><span class="citation-917">This involvement gap hampers the Council&#8217;s ability to (a) accurately reflect the aspirations of the ACM diaspora and (b) amend the Constitution (which requires voting by at least 10% of the membership)</span></span><span data-path-to-node="26,1,1,25">. </span><span data-path-to-node="26,1,1,27"><span class="citation-916">I would advocate crafting strong outreach and incentive mechanisms to foster deeper member engagement in ACM&#8217;s institutional processes</span></span><span data-path-to-node="26,1,1,29">.</span></p>
<p><span data-path-to-node="26,1,1,31"><span class="citation-915"><strong>ACM conferences</strong>: A hallmark of ACM has been its large portfolio of high-quality conferences showcasing the latest research advances</span></span><span data-path-to-node="26,1,1,33">. </span><span data-path-to-node="26,1,1,35"><span class="citation-914">Sustaining this reputation, which is under pressure, requires privileging creativity over derivative work, ensuring the integrity of peer review, and emphasizing quality over quantity</span></span><span data-path-to-node="26,1,1,37">. </span><span data-path-to-node="26,1,1,39"><span class="citation-913">My efforts in this sphere would aim to ensure we remain faithful to ACM&#8217;s charter of &#8220;serving both professional and public interests.&#8221;</span></span></p>
<p><span data-path-to-node="26,1,1,42"><span class="citation-912">I believe my significant exposure to ACM operations, especially the four-year stint on the ACM India Council, would help make substantive progress toward these goals</span></span>.</p>
</article>
</section>
<article style="overflow: hidden; margin-bottom: 40px; border-bottom: 1px solid #eee; padding-bottom: 20px;"></article>
<h2><strong>Candidates for Members-at-Large</strong></h2>
<section>
<article style="overflow: hidden; margin-bottom: 40px; border-bottom: 1px solid #eee; padding-bottom: 5px;"><img decoding="async" class="alignleft" style="float: left; width: 276px; height: 276px; margin: 0px 20px 10px 0px; border: 1px solid #cccccc; border-radius: 4px;" src="https://cacm.acm.org/wp-content/uploads/2026/04/cj_barrios_hernandez_MaL.jpeg" alt="Carlos Jaime Barrios Hernández" /></p>
<h3 id="Hernandez">Carlos Jaime Barrios Hernández</h3>
<p><em>Full Professor</em><br /><em>Universidad Industrial de Santander</em><br /><em>Bucaramanga, Colombia</em></p>
<p><em>Guest Researcher</em><br /><em>LIG/INRIA-Grenoble and INSA-Lyon / CITI Laboratory, </em><em>France</em></p>
</article>
<article style="overflow: hidden; margin-bottom: 5px; border-bottom: 1px solid #eee; padding-bottom: 5px;">
<h4>Biography</h4>
<p>Carlos J. Barrios Hernández is a Full Professor at UIS Colombia and head of SC3UIS. A guest researcher at LIG/INRIA and CITI-Lyon, France, his academic background includes a Ph.D. from Université Nice-Sophia Antipolis, an M.Sc. from Grenoble-Alpes, and an HDR from INSA-Lyon (2025). As a leader in HPC and the computing continuum, he has chaired SCALAC since 2014 and co-founded the CARLA conference and the SC-CAMP School. His scientific work comprises 100+ peer-reviewed publications and several patents, focusing on scalable HPC architectures, hybrid systems, sustainability, scientific computing, and energy efficiency. He is an ACM Senior Member (2025), chair of the SIGHPC Computing Continuum Chapter, and a Senior Member of the IEEE Computer Society. He bridges global research through scientific cooperation and partnerships, fostering advanced digital and technical exchange for human wellness.</p>
<h4>Statement</h4>
<p>My research transforms traditional HPC into a <strong>sustainable computing continuum</strong>, integrating multi-scale, hybrid, and non-classical architectures. As AI demand rises, I prioritize energy efficiency and environmental responsibility, optimizing massive parallel processing to support social progress without causing ecological harm. I propose four key global actions through the ACM :</p>
<ol>
<li><strong>Global cooperation</strong>: Solidify partnerships via shared resource-scheduling protocols and joint research calls targeting global crises, shifting from competition to coordination.</li>
<li><strong>Techno-ecological responsibility</strong>: Aligning with ACM Principle 1.1, I advocate for &#8220;Sustainability Scores&#8221; in scientific activities to motivate ethical development and document social impacts.</li>
<li><strong>Open science and inclusion</strong>: Bridge the &#8220;knowledge gap&#8221; by promoting multilingual open-access repositories (Spanish, Portuguese, French) and providing remote regions with access to research platforms.</li>
<li><strong>Ethics and outreach</strong>: Establish Global Ethics Task Forces to advise on data sovereignty and carbon footprints, ensuring advanced computing protects life and fosters inclusion.</li>
</ol>
</article>
<hr />
<article class="candidate">
<article style="overflow: hidden; margin-bottom: 40px; border-bottom: 1px solid #eee; padding-bottom: 5px;"><img decoding="async" class="alignleft" style="float: left; width: 275px; height: auto; margin: 0 20px 10px 0; border: 1px solid #ccc; border-radius: 4px;" src="https://cacm.acm.org/wp-content/uploads/2026/04/y_li_MaL.jpeg" alt="Yunyao Li" /></p>
<h3 id="Li">Yunyao Li</h3>
<p><em>Director of Machine Learning<br />Adobe<br />San Jose, CA, USA</em></p>
</article>
<article style="overflow: hidden; margin-bottom: 5px; border-bottom: 1px solid #eee; padding-bottom: 5px;">
<h4>Biography</h4>
<p>Yunyao Li is a Director of Machine Learning at Adobe Experience Platform, where she leads strategic initiatives to integrate Generative AI and Knowledge Graphs into scalable enterprise systems. She previously served as Head of Machine Learning for the Apple Knowledge Platform, delivering ML solutions that power Siri and Spotlight for billions of users. Earlier, as a Distinguished Research Staff Member and Master Inventor at IBM Research-Almaden, she founded the Scalable Knowledge Intelligence department and transitioned foundational research into more than 20 commercial products.</p>
<p>Yunyao has made significant contributions to scalable natural language processing and human-computer cooperation. She has authored over 100 peer-reviewed publications, holds 36 patents, and is co-author of two books, including the recent <em>Natural Language Interfaces to Databases</em> (2024).</p>
<p>Deeply committed to bridging industry and academia, Yunyao serves on advisory boards for academic programs at the University of Michigan, UC Santa Cruz, and the University of Colorado Boulder, where she helps shape curriculum and mentors future computing professionals. Her global influence is reflected in her frequent keynote addresses and invited talks at leading institutions and conferences.</p>
<p>Her leadership is further defined by her role in architecting venues where the computing community connects, including pioneering the establishment of Industry Tracks at premier conferences such as NAACL (2018, 2021, 2022), EMNLP (2022), and ACL (2025, 2026). An ACM Distinguished Member, Yunyao has served on over 20 organizing committees for premier conferences and as guest editor for leading ACM journals. Beyond ACM, she served on the NAACL Executive Board (2022<span data-path-to-node="8,0,1,44"><span class="citation-1051"><span class="citation-1056">–</span></span></span>2024) and was selected for the American National Academies&#8217; inaugural New Voices program. She holds undergraduate degrees from Tsinghua University and a Ph.D. from the University of Michigan, Ann Arbor.</p>
<h4>Statement</h4>
<p>I am honored to be nominated for ACM Member at Large. We are at a pivotal moment where AI is fundamentally reshaping the computing landscape. In this era of rapid change, ACM has a unique and essential duty to bridge disparate communities, cultivate diverse talent, and provide the trusted leadership our field requires.</p>
<p><strong>Bridge the industry-academia gap in AI</strong>: My top priority is expanding ACM&#8217;s relevance to the industry, which today encompasses both researchers and practitioners deploying systems at a global scale. Having pioneered the inaugural Industry Track at NAACL 2018 and expanded this model across other premier venues, I have seen how intentional industry integration enriches our scientific discourse. I will champion the expansion of ACM venues to better welcome practitioner participation, fostering a tighter feedback loop between foundational discovery and real-world deployment across all ACM SIGs.</p>
<p><strong>Empower the next generation</strong>: As early-career professionals navigate a landscape of technical uncertainty, ACM must be their most reliable partner in professional development. I will lead efforts to create inclusive leadership opportunities and cross-sector mentoring programs. My goal is to equip the next generation with the networks and skills required for success across academia, industry, and engineering.</p>
<p><strong>Enhance global impact and public voice</strong>: We must broaden ACM&#8217;s reach to better support emerging research communities globally. Simultaneously, as AI raises profound societal questions, ACM must serve as the preeminent evidence-based voice for the public. I will leverage my experience with the National Academies to support efforts that translate technical expertise into accessible, actionable guidance for policymakers and the global community. I look forward to collaborating with ACM&#8217;s leadership and volunteers to ensure our organization remains the premier, inclusive home for those shaping the future of computing.</p>
</article>
<article class="candidate">
<article style="overflow: hidden; margin-bottom: 40px; border-bottom: 1px solid #eee; padding-bottom: 5px;"><img decoding="async" class="alignleft" style="float: left; width: 275px; height: auto; margin: 0 20px 10px 0; border: 1px solid #ccc; border-radius: 4px;" src="https://cacm.acm.org/wp-content/uploads/2026/04/L_tapia_MaL.jpeg" alt="Lydia Tapia" /></p>
<h3 id="Tapia">Lydia Tapia</h3>
<p><em>Professor of Computer Science<br />University of New Mexico<br />Albuquerque, NM, USA</em></p>
</article>
<article style="overflow: hidden; margin-bottom: 5px; border-bottom: 1px solid #eee; padding-bottom: 5px;">
<h4>Biography</h4>
<p>Lydia Tapia earned a Bachelor of Science degree in Computer Science from Tulane University in 1998 and a Ph.D. in Computer Science from Texas A&amp;M University in 2009. Her dissertation was Intelligent Motion Planning and Analysis with Roadmap Methods for the Study of Complex and High-Dimensional Motions. After holding a Computing Innovation Postdoctoral Fellowship at the University of Texas at Austin, she joined the University of New Mexico, where she is a Professor of Computer Science and previously served as Department Chair. In those roles, she led significant faculty growth, hiring 25% of the faculty; secured departmental accreditation; and mentored over 30 research trainees. She also served as Faculty in Residence at Google, working in robotics and engineering education.</p>
<p>Her research applies machine learning to problems in robotics, games, and computational biology, where systems must reason about complex movements and decisions. She has authored or coauthored more than 65 peer-reviewed papers, and she holds a U.S. patent. Her work has been recognized with a National Science Foundation CAREER Award, the Borg Early Career Award from the Computing Research Association (CRA), and a Best Paper Award. She helped found an ACM-W chapter as a graduate student and has since been active in ACM conferences, including SIGBIO- and SIGGRAPH-affiliated venues. She has a long record of professional service, including elected service on the CRA Board, leadership of graduate mentoring initiatives, editorial roles for major robotics journals and conferences, and organization of the international Becoming a Robot Guru undergraduate workshop series introducing students to graduate study in robotics and strategies for success.</p>
<h4>Statement</h4>
<p>My professional experience as a teacher, researcher, administrator, and national service leader has been shaped by periods of rapid change in education and technology. Educators are at the front lines, advances in artificial intelligence (AI) are currently reshaping expectations for graduates, faculty workloads, and the structure of computing programs. These shifts now directly affect employability: Students are expected to graduate with both strong foundational knowledge and practical fluency in AI-based tools and workflows. Departments and instructors are being asked to deliver this preparation while maintaining rigor, fairness, and long-term educational value.</p>
<p>As a Member at Large of the Association for Computing Machinery, my goal would be to help ACM address the education-workforce transition in a clear and practical way. ACM is well positioned as both a leader in open-access research and a hub for professional development. I believe ACM can provide leadership on how AI should be integrated into computing education, supporting both undergraduate and graduate programs as well as retraining and continuing education for current professionals, without narrowing learning to short-term tool adoption. My focus would be on advancing evidence-based curricular guidance, clarifying professional expectations in an AI-mediated workplace, and strengthening pathways from computing education to meaningful employment. I aim to help ACM remain a trusted voice that supports members at all career stages while preserving the core principles of computing as a discipline and profession.</p>
</article>
<article class="candidate">
<article style="overflow: hidden; margin-bottom: 40px; border-bottom: 1px solid #eee; padding-bottom: 5px;"><img decoding="async" class="alignleft" style="float: left; width: 275px; height: auto; margin: 0 20px 10px 0; border: 1px solid #ccc; border-radius: 4px;" src="https://cacm.acm.org/wp-content/uploads/2026/04/h_yanco_MaL.png" alt="Holly Yanco" /></p>
<h3 id="Yanco">Holly Yanco</h3>
<p><em>Distinguished Professor of Computer Science</em><br /><em>Distinguished Professor of Mechanical </em><em>and Industrial Engineering<br /></em><em>University of Massachusetts</em><br /><em>Amherst, MA, USA</em></p>
</article>
<article style="overflow: hidden; margin-bottom: 5px; border-bottom: 1px solid #eee; padding-bottom: 5px;">
<h4>Biography</h4>
<p>In September 2025, Holly Yanco joined the faculty at the University of Massachusetts Amherst after 24 years at the University of Massachusetts Lowell. Yanco is leading the expansion of robotics at UMass Amherst as a Distinguished Professor of Computer Science and of Mechanical &amp; Industrial Engineering. Yanco is the founding director of the New England Robotics Validation and Experimentation (NERVE) Center, now growing to a second location.</p>
<p>For more than 25 years, Yanco has developed and led a wide range of interdisciplinary collaborations to solve open problems in robotics and AI. Her research interests include human-robot interaction, evaluation metrics and methods for robot systems and AI, and the use of robots in K<span data-path-to-node="8,0,1,44"><span class="citation-1051"><span class="citation-1056">–</span></span></span>12 education to broaden participation in computer science.</p>
<p>Yanco is currently the Chair of the ACM Doctoral Dissertation Award Committee after joining the committee in 2023. She was General Chair of the 2012 ACM/IEEE International Conference on Human-Robot Interaction. As co-chair of its steering committee from 2013 to 2016, she co-led the creation of ACM Transactions on Human-Robot Interaction. Yanco is a member of the DARPA ISAT Study Group (2024<span data-path-to-node="8,0,1,44"><span class="citation-1051"><span class="citation-1056">–</span></span></span>2027) and has served on the Association for the Advancement of Artificial Intelligence&#8217;s (AAAI) Executive Council (2006<span data-path-to-node="8,0,1,44"><span class="citation-1051"><span class="citation-1056">–</span></span></span>2009) and the CRA Computing Community Consortium (2020<span data-path-to-node="8,0,1,44"><span class="citation-1051"><span class="citation-1056">–</span></span></span>2025).</p>
<p>Yanco is a Fellow of the AAAI and of the American Association for the Advancement of Science (AAAS). Yanco earned her BA in Computer Science and Philosophy from Wellesley College and her MS and Ph.D. from the Massachusetts Institute of Technology.</p>
<h4>Statement</h4>
<p>Computing continues to rapidly evolve, especially with recent advancements in AI. While exciting, it also brings challenges to our ACM community. In academia, AI is being taught across the university, not just in CS where we ensure that technical depth is provided. People without a CS background can produce code that once required a high level of proficiency; while broadening access is powerful, people without CS training may not learn about the ACM Code of Ethics. Conferences are experiencing tremendous pressure with skyrocketing numbers of submissions, asking more of people who volunteer to serve as organizers and program committee members.</p>
<p>Rather than react to each issue in small groups, our community needs to proactively determine how we want to shape the future together. I am honored to have been nominated as a candidate for Member at Large of the ACM Council. With more than 30 years of experience serving the community, my goal is to help ACM lead the global dialogue about the development, evaluation, and societal integration of computing technologies for the benefit of all.</p>
<p>I am particularly interested in exploring how ACM can grow by embracing interdisciplinarity, particularly as AI expands. The International Conference on Human-Robot Interaction has been jointly run by ACM and IEEE since 2006, bringing together several communities. As a Member at Large of the ACM Council, I will identify and advocate for interdisciplinary collaborations that will not only increase ACM&#8217;s impact but will also allow us to lead the discussions about how we intersect with other fields.</p>
</article>
</article>
</article>
</article>
</section>
<section></section>
</article>
]]></content:encoded>
					
					<wfw:commentRss>https://cacm.acm.org/news/acm-2026-general-election/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
		<post-id xmlns="com-wordpress:feed-additions:1">779984</post-id>	</item>
		<item>
		<title>Amazon’s Quiet Overhaul of the Trademark System</title>
		<link>https://cacm.acm.org/opinion/amazons-quiet-overhaul-of-the-trademark-system/</link>
					<comments>https://cacm.acm.org/opinion/amazons-quiet-overhaul-of-the-trademark-system/#respond</comments>
		
		<dc:creator><![CDATA[Jeanne C. Fromer and Mark P. McKenna]]></dc:creator>
		<pubDate>Thu, 23 Apr 2026 16:52:05 +0000</pubDate>
				<category><![CDATA[Data and Information]]></category>
		<category><![CDATA[Security and Privacy]]></category>
		<guid isPermaLink="false">https://cacm.acm.org/?post_type=digital-library&#038;p=780138</guid>

					<description><![CDATA[<p>Amazon's Brand Registry de-centers brands as a consumer search mechanism and will likely amplify Amazon’s own power by making its search function and algorithm more important in finding products.</p>]]></description>
										<content:encoded><![CDATA[<article>
<div class="body" lang="en">
<section id="sec1" class="sec">
<p id="p-1">More than 9.7 million third-party businesses sell goods on Amazon, and Amazon makes a lot of money charging those third parties to sell on its platform—$117.7 billion in 2022, representing 23% of Amazon’s total revenues. Those third-party sellers have long been concerned about counterfeit versions of their goods appearing on Amazon’s platform, which costs them sales and frustrates consumers who do not get what they expect from their purchase. Congress has been worried about counterfeiting too; it has held hearings about the problem and considered new legislation that would have made online platforms more readily liable for trademark infringement on their platforms.</p>
<p id="p-2">Hoping to address its sellers’ concerns and avoid regulation, Amazon launched the Brand Registry in 2015. The Brand Registry is, most basically, a private dispute resolution system that allows trademark owners to object to unauthorized uses of their marks on Amazon without having to file formal legal proceedings. Brand Registry members also get enhanced customer service and technological tools that help identify and remove potential infringing and counterfeit goods. And, importantly, Amazon gives third-party sellers higher visibility in consumer search results on Amazon, brand-analytic tools, and the ability to give their products to credible buyers for Amazon reviews.</p>
<p id="p-3">The Brand Registry has been a hit among third-party sellers: In 2021, there were more than 700,000 active trademarks enrolled in the Brand Registry worldwide. And seller and consumer satisfaction with the Brand Registry seems to have released some of the pressure for Congress to regulate. But the Registry has also become a shadow trademark system with sweeping effect on firm behavior, dramatically affecting businesses’ incentives to choose certain types of trademarks to designate their goods and to seek federal registration of those marks.</p>
<p id="p-4">Most directly, the Brand Registry has changed firms’ incentives to federally register trademarks. American trademark rights do not depend on registration. A firm develops trademark rights simply by using the mark. While registration provides certain benefits to trademark owners, unregistered marks are enforceable under federal law on largely the same terms as registered marks. For that reason, most small firms have traditionally not needed to register their marks; unregistered rights have been perfectly adequate and have the significant benefit of being free.</p>
<p id="p-5">Amazon’s policies have shifted that calculus because parties can participate in the Brand Registry in the U.S. only if their mark is registered with the U.S. Patent and Trademark Office (PTO)—or, as of very recently, if their application to register is pending. Given Amazon’s dominance as an online shopping platform, small and medium-sized businesses often feel compelled to sell on the platform, and registration in the PTO is the ticket to the meaningful enforcement and search benefits provided to participants in the Brand Registry. Businesses therefore have strong incentives to register marks when they previously would have relied on unregistered rights.</p>
<p id="p-6">The result has been a dramatic increase in the number of applications to register, which has swamped the PTO and created delays for all applicants. Our data show that annual PTO applications estimated to originate with small businesses have approximately doubled since Amazon’s Brand Registry began, increasing the proportion of filings from these entities from approximately 30% to approximately 40% annually. (We do not suggest that all of this increase is attributable to Amazon alone. Other factors like PTO encouragement of small businesses to register their marks may also have played a role.)</p>
<p id="p-7">In response to the delays that Amazon’s policies helped create, Amazon recently started qualifying parties for the Brand Registry based only on a pending application. This change enables parties to privately enforce marks that might ultimately be rejected by the PTO, which is likely to increase applications even more and to create further PTO delays.</p>
<p id="p-8">The Brand Registry also creates incentives for businesses to adopt different kinds of marks for their products. Trademark law traditionally has made it much harder to claim descriptive terms (like NATIONAL CAR RENTAL for nationwide car rental services) or generic terms (such as APPLE for apples) as trademarks. Competitors also need to be able to use those kinds of marks so they can communicate with consumers about what they’re selling, and for that reason, trademark law makes it harder to claim rights in descriptive and generic terms. But Amazon’s policies make control over those kinds of terms especially more valuable because consumers use those terms to search on Amazon. If a party can claim trademark rights in those terms, they can use them to make it much harder for others selling on Amazon.</p>
<p id="p-9">The Brand Registry’s structure also enables parties to game the limitations that the PTO uses to confine protection of descriptive and generic terms. For example, the PTO rejects parties’ claims to generic terms themselves, but it will allow a party to register a particular stylized form of that term on the assumption that a court would only enforce rights in that mark if someone else copied the stylization and not simply the word. For example, while the PTO would not register the word MUSTACHES for fake mustaches, it has registered the stylized mark shown in the accompany figure image for those goods.</p>
<figure id="attachment_780188" aria-describedby="caption-attachment-780188" class="wp-caption alignnone"><img loading="lazy" decoding="async" class=" wp-image-780188" src="https://cacm.acm.org/wp-content/uploads/2026/04/65.5.Fromer.Figure.jpg" alt="theoretical 'mustaches' brand" width="422" height="131" srcset="https://cacm.acm.org/wp-content/uploads/2026/04/65.5.Fromer.Figure.jpg 824w, https://cacm.acm.org/wp-content/uploads/2026/04/65.5.Fromer.Figure.jpg?resize=300,93 300w, https://cacm.acm.org/wp-content/uploads/2026/04/65.5.Fromer.Figure.jpg?resize=768,239 768w" sizes="auto, (max-width: 422px) 100vw, 422px" /><figcaption id="caption-attachment-780188" class="wp-caption-text">Stylized trademark example.</figcaption></figure>
<p id="p-12">The Brand Registry also creates greater incentive to claim so-called “nonsense marks,” which are strings of letters or numbers that are not comprehensible as words (“marks” like ELXXROONM, SUJIOWJNP, XUFFBV, and LXCJZY). Indeed, the PTO data suggest a tremendous increase in filings for nonsense marks in the past few years, from almost none to over twenty thousand applications annually (0.5% of annual filings to approximately 4.5%).<a class="footnote-link xref xref-fn" href="#fn1" data-jats-rid="fn1" data-jats-ref-type="fn"><sup>a</sup></a></p>
<p id="p-13">Nonsense marks are currently easy to register as trademarks because they are unlikely to be confusingly similar to existing marks and they appear not to provide any information about the goods or services with which they are used. According to trademark law, unlike competitively worrisome generic and descriptive terms, these terms are “inherently distinctive” and therefore immediately protectable upon use. But trademark law treats these terms that way because it assumes that consumers will automatically treat them as trademarks. That makes sense for words that are likely to have meaning to consumers and will likely be regarded as brands; it makes no sense for random strings of letters that will not be memorable at all. But when algorithms do the searching, businesses just need something that the algorithm can use to preference them and will get registered easily by the PTO for qualification on the Brand Registry. Nonsense will do.</p>
<p id="p-14">These changes, along with the algorithmic benefits offered to participants in the Brand Registry, have the collective effect of de-centering brands as the relevant consumer search mechanism. That might seem like a win to critics of excessive branding, who might think this portends a potential democratization of the online marketplace. But there is irony here: Amazon’s de-centering third-party branding will likely amplify Amazon’s own power by making its search function and algorithm even more important in finding products. And it certainly enhances the value of Amazon’s own branding strategies, as reflected in the massive expansion of products sold under the Amazon Basics and Amazon Essentials brands. Amazon controls its platform and can preference its own products in search results based on product descriptor keywords, making its house brand more important than product line brands. To take just one example, a search for “Hanes T-shirt” returns an Amazon Essentials T-shirt as the first result, followed by several Hanes results.</p>
<p id="p-15">In all of these ways, actors affected by Amazon’s business model and Brand Registry have overhauled central aspects of the trademark system in ways that are potentially troublesome. That effect may not have been entirely intentional—indeed, it seems very likely Amazon created the Brand Registry in part to address actual concerns about counterfeit goods and product liability and thereby avoid regulation like it would face under proposed legislation in Congress.</p>
<p id="p-16">Regardless of Amazon’s intent, its business model and Brand Registry raise profound questions for the government, business, and technology designers. Amazon is not the first commercial powerhouse, nor is it the first to create a private dispute resolution system. But Amazon’s effect on the trademark system is unprecedented. What does, or should, it mean that one private party can so significantly affect a legal system? Do we want the legal system to have to continually adapt to Amazon’s rules? If not, how can the law disable Amazon from having such a profound impact?</p>
<p id="p-17">Amazon’s model and its policies also likely increase its own power vis-à-vis third-party brands, de-centering branding more generally. It makes shopping on the platform more dependent on Amazon’s algorithm and consumer reviews and less reliant on third-party brands to give information about products. The net value of that shift may be in the eye of the beholder: It depends on how one weighs the potential benefits of search simplification and lower prices for consumers, as well as the ease of marketplace entry for third-party sellers, against the costs of Amazon’s increased power over third-party sellers. Likewise, the benefits of a reduction in the power of brands depend on whether alternative search tools, particularly algorithms that focus on product information and consumer reviews, convey relevant information to consumers as effectively as trademarks. Whether and how the law should respond to Amazon’s effects on the trademark system depends on whether we want the trademark system to demand that trademarks play their traditional role or whether instead the facts on the ground have changed so much that the premises of that system no longer hold.</p>
<p id="p-18">More broadly, the effects here raise important questions for businesses and technology designers to work through the legal implications of the systems they are putting in place for business or engineering reasons. In a world in which platform systems increasingly underpin commerce, system design decisions—even seemingly small ones—can have broad impact given the large numbers of users of these platforms.</p>
</section>
</div>
</article>
]]></content:encoded>
					
					<wfw:commentRss>https://cacm.acm.org/opinion/amazons-quiet-overhaul-of-the-trademark-system/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
		      <dc:creator><![CDATA[Mark P. McKenna]]></dc:creator>
<post-id xmlns="com-wordpress:feed-additions:1">780138</post-id>	</item>
		<item>
		<title>Is AI Intelligent?</title>
		<link>https://cacm.acm.org/opinion/is-ai-intelligent/</link>
					<comments>https://cacm.acm.org/opinion/is-ai-intelligent/#respond</comments>
		
		<dc:creator><![CDATA[Bertrand Meyer]]></dc:creator>
		<pubDate>Thu, 23 Apr 2026 16:49:29 +0000</pubDate>
				<category><![CDATA[Artificial Intelligence and Machine Learning]]></category>
		<category><![CDATA[Theory]]></category>
		<guid isPermaLink="false">https://cacm.acm.org/?post_type=digital-library&#038;p=777735</guid>

					<description><![CDATA[<p>Two fundamentally different concepts of intelligence are behind the debate over whether AI is "intelligent."</p>]]></description>
										<content:encoded><![CDATA[<article>
<div class="body" lang="en">
<section id="sec1" class="sec">
<p id="p-1"><b>Bertrand Meyer<br /></b><b>Two Concepts of Intelligence<br /></b><b>DOI:</b> 10.1145/3797898<br /><a class="ext-link" href="https://bit.ly/3MgwuHz" data-jats-ext-link-type="uri">https://bit.ly/3MgwuHz</a></p>
<p id="p-5">“But is it intelligence?”</p>
<p id="p-6">The rise of AI triggers endless discussions of what constitutes intelligence. A debate which, even sticking to computer science, goes back to Weizenbaum&#8217;s Eliza and further to Turing and von Neumann.</p>
<p id="p-7">Technical people can be wary of philosophizing (as they prefer problem-solving and action) but should not relinquish such discussions to people who may not understand the technology. This note argues that the source of many apparent disagreements is that people rely, without necessarily being aware of it, on two radically different definitions of intelligence.</p>
</section>
<section id="sec2" class="sec">
<h2 class="heading">“It Only Appears to Understand”</h2>
<p id="p-8">We find an example of the first view in a recent hearing<a class="footnote-link xref xref-fn" href="#fn1" data-jats-ref-type="fn" data-jats-rid="fn1"><sup>a</sup></a> of the French National Assembly on AI, where a philosopher of science, Olivier Rey, explained that “<i>artificial intelligence is not intelligence</i>” because the program “<i>does not understand.</i>”</p>
<p id="p-9">What does that mean? I can boast I “understand” basic linear algebra. Still, if you ask me to find the eigenvalues of a medium-size matrix, I might occasionally make a mistake, and if you ask me to prove one of the theorems in the field, I might occasionally get stuck. If you ask an LLM the same questions, it will get the answers right much of the time, but may also occasionally mess up (“<i>hallucinate</i>”). What enables me to say that a human such as me is more intelligent than Claude or Mistral? I actually suspect that the LLM will get answers right more often, but what matters is that both the LLM and I will get many answers right and some wrong.</p>
<p id="p-10">I can hear the retorts: “<i>the AI tool only </i>appears<i> to understand, I </i>really <i>understand. It doesn’t matter that I make occasional mistakes, they are superficial; the tool&#8217;s hallucinations show that it has no clue.”</i> All beautiful arguments, and worthless because they do not satisfy the basic criterion of scientific arguments: they are <em>not falsifiable</em>. Falsifiability would mean that we can construct a reliable experiment to test whether a human or tool does not just <i>apply</i> a theory, but somehow “<i>understands</i>” it. It is hard to imagine how the experiment would look. The Turing Test or Searle&#8217;s Chinese Room only measure <i>outcomes</i>. Both competent humans and today&#8217;s AI tools will pass them. Throw in enough complexity and tools often fare better than humans. Does it mean they are more intelligent? Do they “<i>understand</i>” less? Might they possibly (scary prospect) understand more?</p>
<p id="p-11">We cannot begin to address these questions until we know what it means to “<i>understand</i>” a concept.</p>
</section>
<section id="sec3" class="sec">
<h2 class="heading">American vs. European Views</h2>
<p id="p-12">I believe much of the debate is due to clashing understandings of “intelligence.”</p>
<p id="p-13">The clash reminds me of the shock I experienced when, as a student at Stanford, I first came to the legendary AI lab, at its zenith with such luminaries as John McCarthy (the founder), Arthur Samuel, Zohar Manna, Terry Winograd. “Intelligence” was on everyone&#8217;s lips; I vividly remember discovering the widely accepted working definition was “<i>the ability to adapt to new situations and learn from experience.</i>” It was scandalous to me, coming from a European intellectual perspective. These Americans, I thought, are so utilitarian, prosaic, earth-bound, pedestrian, mercenary! There has to be something deeper to intelligence than knowing how to react to circumstances: you must <i>understand</i> the situation. I had studied Latin and knew that etymology was on my side: <i>intelligo</i> means “I understand.”</p>
<p id="p-14">As I soon found out, the issue was not just with me, but reflected a difference between continental European and Anglo-Saxon views. The Larousse<a class="footnote-link xref xref-fn" href="#fn2" data-jats-ref-type="fn" data-jats-rid="fn2"><sup>b</sup></a> definition, for example, starts with “<i>the set of mental functions whose goal is conceptual and rational knowledge.</i>&#8221; Hence the schism between those who consider intelligence the ability to understand (like me back then, and Mr. Rey today) and those for whom it is the ability to cope.</p>
<p id="p-15">The European view rests on a fascinating tradition of <i>explaining</i> things (and sounding very smart). The French in particular have made a specialty of writing the definitive account of a country, explaining it to the world in general, and in particular to the country’s own gobsmacked natives, on the basis of one glorious in-and-out trip. Tocqueville is the most famous example, but there are also Barthes on Japan and Custine on Russia. Not French and not harmless, we have Marx and Freud who respectively “understood” all about (respectively) history and psychology, and explained it to us. It is petty to point out these theories have had zero success in predicting future outcomes. Or that in the first case, the main result was to destroy countries and civilizations and led to the death of millions. Who is to quibble about such details when these theories make us “understand” by “explaining” so <i>intelligently</i>!</p>
<p id="p-16">Serious scientific theories do explain, too, and make us understand complex things. The difference is that they predict correctly, and are falsifiable. Relativity made us understand the basics of time and space not just by presenting convincing ideas, but by predicting that, for a certain eclipse at a certain place, light would bend not by 0.87 arcseconds, as Newton would have it, but by twice as much. Had Eddington’s measurement been different, he would have disproved (falsified) the theory.</p>
<p id="p-17">The difference between the two concepts of intelligence—ability to understand, ability to act successfully—is also the difference between <i>deductive</i> approaches, which start from a theory and attempt to verify it through facts, and <i>inductive</i> ones, which start from facts and build up a theory. It is a deep difference, going back far in the history of thought. Among philosophers we find, on the conceptual/deductive side Descartes and Kant, and on the empirical/inductive side such English and American thinkers as Hume, John Stuart Mill, and behaviorists typified by Skinner.</p>
</section>
<section id="sec4" class="sec">
<h2 class="heading">Contrasting the Two Views</h2>
<p id="p-18">The appeal of the first view (“<i>I am intelligent because I understand</i>”) is its elegance and promise of powerful speeches. Its limitation is the difficulty of validating or falsifying it. Conspiracy theorists (including Marxists and Freudians) also make beautiful speeches. If you and I both have explanations for something, but they are incompatible, how do I convince you that mine is right and yours is wrong?</p>
<p id="p-19">The appeal of the second view (“<i>I am intelligent because I can make predictions that turn out right</i>”) is its practicality. But how do we know that what it describes is really intelligence and not just careful record-keeping?</p>
<p id="p-20">Old-AI, with its expert systems and logic-based tools, was of the first kind. The consensus is that it failed. Modern-AI is almost entirely (at least in its current, intermediate state of evolution) of the second kind. Modern-AI is <em>machine-learning</em>: it builds answers to new queries by extrapolating from a large body of validated answers to previous queries. Is it intelligence? Is the human-quality-level translation of today&#8217;s translation tools intelligent? Is a vibe-coding tool more intelligent than the programmer who uses it? Is a medical-image analysis tool which produces fewer false negatives and positives than a Stanford Hospital radiologist more or less intelligent than that doctor? For that matter, are non-AI programs such as a compiler intelligent? (No human could correctly compile a 100,000-line program in any reasonable time.)</p>
<p id="p-21">With recent advances in AI, it becomes ever harder for proponents of intelligence-as-understanding to continue asserting that those tools have no clue and “just” perform statistical next-token prediction. Borrowing Kian Katanforoosh’s examples in a Stanford lecture,<a class="footnote-link xref xref-fn" href="#fn3" data-jats-ref-type="fn" data-jats-rid="fn3"><sup>c</sup></a> today’s deep-learning systems can complete sentences such as: “<i>I poured myself a cup of</i> …” (how is that not understanding co-occurrence patterns?); “<i>The capital of France is</i> …” (how is that not understanding geographical connections?); “<i>She unlocked her phone using her</i>…” (how is that not understanding semantic connections?); “<i>The cat chased the</i> …” (multiple plausible connections, so how is that not understanding probability?); “<i>If it is raining, I should bring an</i>…” (how is that not understanding inference?).</p>
<p id="p-22">What do people mean, then, when they say “<em>A</em><i>I is not intelligence because it does not understand what it is talking about</i>”? Since they do not define what “<i>understanding</i>” is, they may just mean “AI does not understand <em>in the same way I do</em>.” That is a tenuous argument. Similar to saying “Airplanes do not fly (do not believe your own lying eyes!) because they do not fly the same way as birds.”</p>
<p id="p-23">I tend today to think that I was wrong back then, and find much to like in the empirical and inductive interpretation. But my changes of mind (a mind that could change again with new arguments and new technology) are not the subject of this discussion. What does matter, regardless of your personal preference for either of them, is the existence of two fundamentally different concepts of intelligence. Discussions of the “I” in “AI” are pointless unless they specify which one they use.</p>
</section>
</div>
</article>
]]></content:encoded>
					
					<wfw:commentRss>https://cacm.acm.org/opinion/is-ai-intelligent/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
		<post-id xmlns="com-wordpress:feed-additions:1">777735</post-id>	</item>
		<item>
		<title>Rolling in the Deep of Cognitive and AI Biases</title>
		<link>https://cacm.acm.org/research/rolling-in-the-deep-of-cognitive-and-ai-biases/</link>
					<comments>https://cacm.acm.org/research/rolling-in-the-deep-of-cognitive-and-ai-biases/#respond</comments>
		
		<dc:creator><![CDATA[Nicoleta Tantalaki and Athena Vakali]]></dc:creator>
		<pubDate>Thu, 23 Apr 2026 16:31:20 +0000</pubDate>
				<category><![CDATA[Artificial Intelligence and Machine Learning]]></category>
		<category><![CDATA[HCI]]></category>
		<category><![CDATA[Society]]></category>
		<guid isPermaLink="false">https://cacm.acm.org/?post_type=digital-library&#038;p=779047</guid>

					<description><![CDATA[<p>The HAI-ROLL framework reveals interdependencies between human heuristics and AI biases across the AI lifecycle.</p>]]></description>
										<content:encoded><![CDATA[<article>
<div class="body" lang="en">
<section id="sec1" class="sec">
<p id="p-1">Artificial intelligence (AI) has rapidly transformed decision making in numerous domains. Acting alone or as a human companion, AI offers health suggestions, approves loans, detects criminal profiles, distributes digital content, and supports a wide range of additional decision-making tasks.<a class="reference-link xref xref-bibr" href="#B30" data-jats-ref-type="bibr" data-jats-rid="B30"><sup>30</sup></a> However, the hypothesis that AI would result in decisions that are more objective and rational than those made by humans has not been verified. AI algorithms frequently produce flawed or discriminatory outcomes, driven by unexpected and unpredictable biases embedded within opaque systems.<a class="reference-link xref xref-bibr" href="#B27" data-jats-ref-type="bibr" data-jats-rid="B27"><sup>27</sup></a> Since AI reflects not only the inequitable patterns found in society but also the fair or biased perspectives of its creators,<a class="reference-link xref xref-bibr" href="#B14" data-jats-ref-type="bibr" data-jats-rid="B14"><sup>14</sup></a> it is clear that an exploration of sociotechnical bias is urgently needed.<a class="reference-link xref xref-bibr" href="#B30" data-jats-ref-type="bibr" data-jats-rid="B30"><sup>30</sup></a> In pursuit of <i>fairness in AI (FairAI)</i>, computational methods such as resampling or reweighting have been used, but these are often inadequate and costly and rely on unrealistic assumptions, such as prior knowledge of which sensitive features are most bias-prone.<a class="reference-link xref xref-bibr" href="#B18" data-jats-ref-type="bibr" data-jats-rid="B18"><sup>18</sup></a> A rich and emerging bibliography has detected bias in all phases of the AI lifecycle, including data collection (pre-processing), modeling (in-processing), and deployment (post-processing), and has identified a divergent set of computational fairness metrics for bias mitigation.<a class="reference-link xref xref-bibr" href="#B14" data-jats-ref-type="bibr" data-jats-rid="B14"><sup>14</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B26" data-jats-ref-type="bibr" data-jats-rid="B26"><sup>26</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B35" data-jats-ref-type="bibr" data-jats-rid="B35"><sup>35</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B37" data-jats-ref-type="bibr" data-jats-rid="B37"><sup>37</sup></a> These studies have shown benefits; however, (computational) AI biases remain. At the same time, the rapid and widespread adoption of large language models (LLMs) has intensified the need to identify hidden and underexplored forms of bias. This is particularly relevant in models fine-tuned through methods such as reinforcement learning from human feedback (RLHF), which aim to produce outputs that are truthful, helpful, and harmless.<a class="reference-link xref xref-bibr" href="#B8" data-jats-ref-type="bibr" data-jats-rid="B8"><sup>8</sup></a> However, because human feedback reflects subjective values and social norms, this alignment process can unintentionally reinforce cultural biases, dominant perspectives, or inconsistent expectations.<a class="reference-link xref xref-bibr" href="#B8" data-jats-ref-type="bibr" data-jats-rid="B8"><sup>8</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B20" data-jats-ref-type="bibr" data-jats-rid="B20"><sup>20</sup></a></p>
<aside class="boxed-text">
<div class="article-key-insights">
<h2>Key Insights</h2>
<ul class="list" data-jats-list-type="bullet">
<li class="list-item">
<p id="p-2">AI is never developed in isolation; it is deeply shaped by the ways humans think, decide, and interact with the world. Cognitive shortcuts and implicit heuristics embedded in human decision making often migrate into AI design, influencing system behavior and leading to unfair or unintended outcomes.</p>
</li>
<li class="list-item">
<p id="p-3">The HAI-ROLL framework systematically identifies AI lifecycle actions that act as sources of AI bias, links them to the underlying human heuristics, and reveals hidden relationships between human (cognitive) processes and algorithmic biases.</p>
</li>
<li class="list-item">
<p id="p-4">By surfacing these pathways, HAI-ROLL opens new insightful and practical AI design guidance. It promotes bias-aware decision making across the AI lifecycle and reframes fairness not as a technical afterthought but rather as a human-centered responsibility woven into the entire AI development process.</p>
</li>
</ul>
</div>
</aside>
<p id="p-5">Our work is motivated by these pervasive AI biases and the need to understand their causes and effects in our new human–AI symbiotic reality. Humans make decisions driven by their own perceptions, and such <i>human heuristics</i> (mental shortcuts, motivations, time constraints, or other limitations) often result in systematic errors, known as <i>cognitive</i> <i>biases</i>.<a class="reference-link xref xref-bibr" href="#B36" data-jats-ref-type="bibr" data-jats-rid="B36"><sup>36</sup></a> Seminal work in cognitive science has identified more than 180 cognitive biases in human judgment, and the list is continuously evolving.<a class="reference-link xref xref-bibr" href="#B2" data-jats-ref-type="bibr" data-jats-rid="B2"><sup>2</sup></a> Unfortunately, such valuable work in cognitive science has up to now been overlooked in AI. As a result, we lack a <i>deep</i> exploration of the interdependencies<a class="reference-link xref xref-bibr" href="#B37" data-jats-ref-type="bibr" data-jats-rid="B37"><sup>37</sup></a> between human heuristics (cognitive) and AI (computational) biases. We believe we must foreground the human-side factors, which cause or steer the many AI biases,<a class="reference-link xref xref-bibr" href="#B25" data-jats-ref-type="bibr" data-jats-rid="B25"><sup>25</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B30" data-jats-ref-type="bibr" data-jats-rid="B30"><sup>30</sup></a> under a new line of sociotechnical research designed to prevent harms, in particular to minority or underprivileged populations.<a class="reference-link xref xref-bibr" href="#B31" data-jats-ref-type="bibr" data-jats-rid="B31"><sup>31</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B37" data-jats-ref-type="bibr" data-jats-rid="B37"><sup>37</sup></a></p>
<p id="p-6">Unlike prior frameworks that focus solely on computational bias, our work traces the roots, forms, and manifestations of cognitive and AI biases. We introduce a systematic mapping framework that reveals interdependencies between human heuristics and AI biases across the AI lifecycle. This cognitive-to-AI bias mapping establishes a human-centric approach to FairAI that:</p>
<ul class="list" data-jats-list-type="bullet">
<li class="list-item">
<p id="p-7">Identifies particular actions as sources of AI (computational) biases</p>
</li>
<li class="list-item">
<p id="p-8">&#8220;Rolls in the deep&#8221;<a class="footnote-link xref xref-fn" href="#fn1" data-jats-ref-type="fn" data-jats-rid="fn1"><sup>a</sup></a> of these actions&#8217; causes to uncover their inherent cognitive (human-side) heuristics, which are projected into each of the AI lifecycle phases</p>
</li>
<li class="list-item">
<p id="p-9">Introduces a systematic methodology to reveal the interrelationships between cognitive heuristics and AI biases, based on the evidence of a flexible exploratory analysis.</p>
</li>
</ul>
<p id="p-10">Our work is based on the science of human heuristics introduced by Nobelist Daniel Kahneman and other pioneering researchers,<a class="reference-link xref xref-bibr" href="#B16" data-jats-ref-type="bibr" data-jats-rid="B16"><sup>16</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B22" data-jats-ref-type="bibr" data-jats-rid="B22"><sup>22</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B23" data-jats-ref-type="bibr" data-jats-rid="B23"><sup>23</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B36" data-jats-ref-type="bibr" data-jats-rid="B36"><sup>36</sup></a> and on state-of-the-art work on prominent AI biases.<a class="reference-link xref xref-bibr" href="#B14" data-jats-ref-type="bibr" data-jats-rid="B14"><sup>14</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B24" data-jats-ref-type="bibr" data-jats-rid="B24"><sup>24</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B26" data-jats-ref-type="bibr" data-jats-rid="B26"><sup>26</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B35" data-jats-ref-type="bibr" data-jats-rid="B35"><sup>35</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B37" data-jats-ref-type="bibr" data-jats-rid="B37"><sup>37</sup></a> By bridging insights from cognitive science and AI research, we open a new dialogue for systematic, cross-disciplinary exploration of bias in AI systems. We aim to equip practitioners with a clear understanding of which decision points are most susceptible to cognitive influence, offering structured support for reflection, anticipation, and bias-aware planning and deployment. Our work provides a theoretical foundation for human-centric, bias-aware decision making through a cognitive reflection framework designed to guide both practitioners and policymakers, while supporting future empirical studies.</p>
</section>
<section id="sec2" class="sec">
<h2 class="heading">Computational Biases in the AI Lifecycle</h2>
<p id="p-11">Among the many types of computational bias explored in the AI lifecycle (pre-, in-, and post-), the most prominent and widely studied<a class="reference-link xref xref-bibr" href="#B14" data-jats-ref-type="bibr" data-jats-rid="B14"><sup>14</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B24" data-jats-ref-type="bibr" data-jats-rid="B24"><sup>24</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B26" data-jats-ref-type="bibr" data-jats-rid="B26"><sup>26</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B35" data-jats-ref-type="bibr" data-jats-rid="B35"><sup>35</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B37" data-jats-ref-type="bibr" data-jats-rid="B37"><sup>37</sup></a> include:</p>
<ul class="list" data-jats-list-type="bullet">
<li class="list-item">
<p id="p-12"><i>Historical bias</i>, when a particular selective handling of certain groups or communities leads to them being advantaged or favored, while others historically remain disadvantaged or excluded. Such biases arise even if the data is sampled flawlessly, since selective preference may be inherent in the original data sources due to the majority following existing patterns or norms<a class="reference-link xref xref-bibr" href="#B30" data-jats-ref-type="bibr" data-jats-rid="B30"><sup>30</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B36" data-jats-ref-type="bibr" data-jats-rid="B36"><sup>36</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B37" data-jats-ref-type="bibr" data-jats-rid="B37"><sup>37</sup></a> For example, a mechanical engineering firm’s dataset might have inherent historical biases if it contains a significantly higher number of male employee records. Such an imbalance stems from the firm’s hiring norms favoring male candidates due to the belief that women are less capable.</p>
</li>
<li class="list-item">
<p id="p-13"><i>Representation bias,</i> when non-representative datasets are generated by sampling from a population—that is, when samples are of small scale—or wrongly selected, neglecting underrepresented groups with fewer data points, resulting in the model’s failure to generalize for these groups.<a class="reference-link xref xref-bibr" href="#B26" data-jats-ref-type="bibr" data-jats-rid="B26"><sup>26</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B35" data-jats-ref-type="bibr" data-jats-rid="B35"><sup>35</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B37" data-jats-ref-type="bibr" data-jats-rid="B37"><sup>37</sup></a> For example, when a recruiting company trains an AI model using a data sample consisting primarily of high-salary male employee profiles, this results in biased hiring recommendations, since the model cannot recognize high-salary female employees or value female candidates.</p>
</li>
<li class="list-item">
<p id="p-14"><i>Measurement bias,</i> when particular features and labels are chosen, collected, or computed, acting as “proxies” that poorly substitute for some construct (an idea or concept).<a class="reference-link xref xref-bibr" href="#B24" data-jats-ref-type="bibr" data-jats-rid="B24"><sup>24</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B35" data-jats-ref-type="bibr" data-jats-rid="B35"><sup>35</sup></a> For example, if an AI model uses Body Mass Index (BMI: estimated by calculating an individual’s weight over her/his height squared) as a proxy for obesity risk, it can mistakenly classify individuals with greater muscle mass as obese, since BMI cannot distinguish between muscle and fat and fails to take into account key factors such as age, gender, or ethnicity.</p>
</li>
<li class="list-item">
<p id="p-15"><i>Algorithmic bias</i>, when inappropriate design choices made mostly for a model’s architecture, functions, and hyperparameters, require several mathematical abstractions. Even under nondiscriminatory intent, these choices often lead to biases, since they show disproportional error rates against underrepresented groups.<a class="reference-link xref xref-bibr" href="#B37" data-jats-ref-type="bibr" data-jats-rid="B37"><sup>37</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B38" data-jats-ref-type="bibr" data-jats-rid="B38"><sup>38</sup></a> For example, an underfitting model (due to its oversimplified nature) has restricted capacity in detecting important data patterns and yields low predictive accuracy, particularly for minority groups underrepresented in the dataset.<a class="reference-link xref xref-bibr" href="#B18" data-jats-ref-type="bibr" data-jats-rid="B18"><sup>18</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B38" data-jats-ref-type="bibr" data-jats-rid="B38"><sup>38</sup></a></p>
</li>
<li class="list-item">
<p id="p-16"><i>Evaluation bias,</i> when the benchmark dataset (used for a model’s evaluation) is not representative of the target population and/or inappropriate performance metrics are used.<a class="reference-link xref xref-bibr" href="#B37" data-jats-ref-type="bibr" data-jats-rid="B37"><sup>37</sup></a> For example, if an AI smile detector model is trained on a dataset lacking adequate representation of Asian female individuals and the testing benchmark is similarly unbalanced, mirroring the training set, a bias against Asian females will go unnoticed.<a class="reference-link xref xref-bibr" href="#B24" data-jats-ref-type="bibr" data-jats-rid="B24"><sup>24</sup></a></p>
</li>
<li class="list-item">
<p id="p-17"><i>Deployment bias,</i> when an AI model is designed to solve a problem and then used to address a different problem in practice. This often occurs when an AI model is built and evaluated to provide recommendations, but is deployed as if it were fully autonomous. In this case, human reviewers overestimate its credibility, leading to a bias known as <i>automation</i> <i>bias</i>.<a class="reference-link xref xref-bibr" href="#B35" data-jats-ref-type="bibr" data-jats-rid="B35"><sup>35</sup></a></p>
</li>
</ul>
<p id="p-18">Given that all of the AI phases, but also the emerging LLMs, are largely driven by human experts, in the next sections we summarize humans&#8217; inherent heuristics, then introduce our methodology for mapping these human heuristics to AI biases.</p>
</section>
<section id="sec3" class="sec">
<h2 class="heading">Heuristics and Biases behind Human Decision Making</h2>
<p id="p-19">The detection of computational bias—recently characterized as just the tip of the bias “iceberg”<a class="reference-link xref xref-bibr" href="#B30" data-jats-ref-type="bibr" data-jats-rid="B30"><sup>30</sup></a>—has been criticized for its limited ability to account for the impact of human and systemic biases throughout the AI lifecycle.<a class="reference-link xref xref-bibr" href="#B14" data-jats-ref-type="bibr" data-jats-rid="B14"><sup>14</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B31" data-jats-ref-type="bibr" data-jats-rid="B31"><sup>31</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B37" data-jats-ref-type="bibr" data-jats-rid="B37"><sup>37</sup></a> To resolve this inefficiency, we explored the groundbreaking work on “judgment under uncertainty,” which has revealed that the broad spectrum of cognitive biases is rooted in three core heuristics: <i>representativeness</i>, <i>availability</i>, and <i>anchoring and adjustment.</i><a class="reference-link xref xref-bibr" href="#B36" data-jats-ref-type="bibr" data-jats-rid="B36"><sup>36</sup></a> We also looked at Kahneman et al.’s <i>Heuristics and Biases: The Psychology of Intuitive Judgement</i>, which highlights the prime role of the <i>affect</i> heuristic in human intuitive judgements.<a class="reference-link xref xref-bibr" href="#B16" data-jats-ref-type="bibr" data-jats-rid="B16"><sup>16</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B22" data-jats-ref-type="bibr" data-jats-rid="B22"><sup>22</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B23" data-jats-ref-type="bibr" data-jats-rid="B23"><sup>23</sup></a> Below, we briefly summarize these heuristics, which serve as a foundation for the following section, which illustrates how these human cognitive tendencies can give rise to, or interact with, AI biases.</p>
<p id="p-20"><i>Representativeness</i> refers to the spontaneous estimation of an event’s probability, based on how closely it resembles a known situation or prototype that exists in a person’s mind. People usually overweight certain information because they believe it is representative of reality, since this information is more recent, or has received more media attention, or is presented in a way that makes it seem more typical of a category. The representativeness heuristic leads to biased views based on stereotypes.<a class="reference-link xref xref-bibr" href="#B16" data-jats-ref-type="bibr" data-jats-rid="B16"><sup>16</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B22" data-jats-ref-type="bibr" data-jats-rid="B22"><sup>22</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B36" data-jats-ref-type="bibr" data-jats-rid="B36"><sup>36</sup></a> For example, a woman who is characterized as friendly and empathetic is more likely to be labeled as a nurse than as a mechanical engineer, since her traits fit the stereotype of a person in a care-giving role.</p>
<p id="p-21"><i>Availability</i> refers to the mental shortcuts humans take when assessing the frequency or plausibility of an event, based on how easily examples come to mind. This attachment to immediate thoughts often prohibits further exploration of potentially significant information, since more-frequent instances are recalled better and faster than less-frequent ones, and repeated occurrences are preferred over rare events. A prominent cognitive bias in such cases is the so-called <i>illusory correlation,</i> occurring when people conclude that variables are correlated because their pairings come to mind easily (i.e., are quick to grasp, seem likely).<a class="reference-link xref xref-bibr" href="#B16" data-jats-ref-type="bibr" data-jats-rid="B16"><sup>16</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B22" data-jats-ref-type="bibr" data-jats-rid="B22"><sup>22</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B36" data-jats-ref-type="bibr" data-jats-rid="B36"><sup>36</sup></a> For example, the availability heuristic occurs when people assume that a high number of women hold leadership positions based on the visibility of a company’s “female empowerment” campaign. This campaign is easily retrievable from memory due to its regular marketing exposure, leading to an overestimation of the actual engagement of women in leadership roles.</p>
<p id="p-22"><i>Anchoring and adjustment</i> occurs when people base their estimates or predictions on an initial piece of information (which may be completely irrelevant) that becomes ingrained in their minds. This information acts as an “anchor,” trapping people in their original opinions, even if they are unrelated to a newly presented case. Such anchor dependency may lead to a critical cognitive bias known as <i>confirmation bias</i>, which is humans&#8217; tendency to favor information that confirms their initial views, causing them to make insufficient adjustments and selectively prefer data that reinforces their initial anchor point.<a class="reference-link xref xref-bibr" href="#B29" data-jats-ref-type="bibr" data-jats-rid="B29"><sup>29</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B36" data-jats-ref-type="bibr" data-jats-rid="B36"><sup>36</sup></a> The anchoring heuristic can even influence individuals with a high level of knowledge or expertise engaged in comparative decision making, as they are swayed by multiple anchors stored in their minds.<a class="reference-link xref xref-bibr" href="#B22" data-jats-ref-type="bibr" data-jats-rid="B22"><sup>22</sup></a></p>
<p id="p-23"><i>Affect</i> refers to quick decisions based on emotions and feelings, leading to reactions that are more emotional than logical or analytical. Kahneman’s fast and intuitive so-called System 1 thinking<a class="reference-link xref xref-bibr" href="#B22" data-jats-ref-type="bibr" data-jats-rid="B22"><sup>22</sup></a> reflects people’s tendency to jump to conclusions and skip information gathering, resulting in a cognitive bias known as <i>conclusion bias</i>. Although humans may also engage in deliberate, slow, and conscious reasoning, “System 2 thinking,”<a class="reference-link xref xref-bibr" href="#B22" data-jats-ref-type="bibr" data-jats-rid="B22"><sup>22</sup></a> their emotional attachments also drive them to seek information that confirms their preexisting beliefs, giving rise once again to confirmation bias<i>.</i><a class="reference-link xref xref-bibr" href="#B22" data-jats-ref-type="bibr" data-jats-rid="B22"><sup>22</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B23" data-jats-ref-type="bibr" data-jats-rid="B23"><sup>23</sup></a> For example, the affect heuristic may manifest in hiring decisions when a recruiter who feels disappointed by previous female employees unconsciously prioritizes male candidates as more capable leaders, despite evidence that new female candidates have equal or superior qualifications.</p>
<p id="p-24">Given this radical, unprecedented mix of human and AI decision making, we must urgently reconsider the role of these human heuristics and respond to the next critical research questions. These include:</p>
<ul class="list" data-jats-list-type="bullet">
<li class="list-item">
<p id="p-25">How can we effectively define a solid methodology to reveal interdependencies between human heuristics and AI biases?</p>
</li>
<li class="list-item">
<p id="p-26">Which human heuristics affect the most critical AI biases and when?</p>
</li>
</ul>
<section id="sec4" class="inline-headings-section">
<h2 data-jats-content-type="inline-heading"><strong>HAI-ROLL: A Framework for Mapping Human Heuristics and AI Biases</strong> </h2>
<p data-jats-content-type="inline-heading">Recent surveys have roughly highlighted harmful actions that may lead to AI biases.<a class="reference-link xref xref-bibr" href="#B14" data-jats-ref-type="bibr" data-jats-rid="B14"><sup>14</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B19" data-jats-ref-type="bibr" data-jats-rid="B19"><sup>19</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B24" data-jats-ref-type="bibr" data-jats-rid="B24"><sup>24</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B26" data-jats-ref-type="bibr" data-jats-rid="B26"><sup>26</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B35" data-jats-ref-type="bibr" data-jats-rid="B35"><sup>35</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B37" data-jats-ref-type="bibr" data-jats-rid="B37"><sup>37</sup></a> However, the origins of these AI biases are poorly referenced.<a class="reference-link xref xref-bibr" href="#B19" data-jats-ref-type="bibr" data-jats-rid="B19"><sup>19</sup></a> We deviate from this restrictive line of research with our proposed framework, Human-to-AI Roll (HAI-ROLL), denoting a deep-dive methodological approach for mapping human heuristics to AI biases. HAI-ROLL positions harmful AI lifecycle actions as catalysts for mapping human heuristics and AI biases (see <a class="xref xref-fig" href="#F1" data-jats-ref-type="fig" data-jats-rid="F1">figure</a>). The HAI-ROLL framework enables a sociotechnical view in several ways. First, it identifies and encodes the particular harmful actions (henceforth noted as [HA.xx]) that lead to unfair outcomes at each phase of the AI lifecycle (pre-, in-, post-). Second, it detects the specific AI biases that have emerged due to these harmful actions. And finally, it traces the origins of these harmful actions by exploring well-established human heuristics.</p>
<p id="p-28">Our transitive reasoning responds to the above research questions since, ultimately, we <i>roll into the deep</i> evidence of human psychology to reveal the relevance, influence, and pathways linking human heuristics to AI biases. Our methodology is grounded in interdisciplinary scientific evidence, not only to detect correlations between human and AI biases but also to offer a systematic and adaptive framework for innovative, human-centric fairness-by-design strategies.</p>
<figure id="F1" class="fig">
<div class="image-container"><img decoding="async" class="graphic" title="Figure. " src="3769686_fig01.svg" alt="" data-image-id="F1" data-image-type="figure" /><figure id="attachment_779525" aria-describedby="caption-attachment-779525" class="wp-caption alignnone"><img loading="lazy" decoding="async" class="wp-image-779525 size-full" src="https://cacm.acm.org/wp-content/uploads/2026/03/Figure1.jpg" alt="" width="1024" height="520" srcset="https://cacm.acm.org/wp-content/uploads/2026/03/Figure1.jpg 6300w, https://cacm.acm.org/wp-content/uploads/2026/03/Figure1.jpg?resize=300,152 300w, https://cacm.acm.org/wp-content/uploads/2026/03/Figure1.jpg?resize=768,390 768w, https://cacm.acm.org/wp-content/uploads/2026/03/Figure1.jpg?resize=1024,520 1024w, https://cacm.acm.org/wp-content/uploads/2026/03/Figure1.jpg?resize=1536,780 1536w, https://cacm.acm.org/wp-content/uploads/2026/03/Figure1.jpg?resize=2048,1040 2048w" sizes="auto, (max-width: 1024px) 100vw, 1024px" /><figcaption id="caption-attachment-779525" class="wp-caption-text">Figure. Mapping human heuristics and AI biases via harmful actions in the AI lifecycle.</figcaption></figure></div>
</figure>
</section>
</section>
<section id="sec5" class="sec">
<h2 class="heading">HAI-ROLL: Pre-Processing Phase</h2>
<p id="p-30">Data and feature selection are critical initial actions that impact the overall AI lifecycle. Since data scales prohibit the inclusion of the entire data population, certain issues are expected to arise. For example, some groups may be excluded or underrepresented or some features cannot be acquired,  leading to inaccurate, possibly harmful results.</p>
<section id="sec6" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>Harmful action [HA.P1]: Use of inappropriate (and non-representative) data samples resulting in computational bias: Representation and historical.</strong>  Problems arise when the variability of the population is not taken into consideration, the size of the sample used is not appropriate, and the sampling method is uneven or limited.<a class="reference-link xref xref-bibr" href="#B35" data-jats-ref-type="bibr" data-jats-rid="B35"><sup>35</sup></a> Such choices often lead to AI <i>representation</i> bias, yielding results applicable to only a subset of the relevant population and not to the broader population itself (due to a non-representative sample). Even a “perfect” sampling plan might turn into an [HA.P1], since AI <i>historical</i> biases might pop up when datasets embed their own ingrained biases.<a class="reference-link xref xref-bibr" href="#B1" data-jats-ref-type="bibr" data-jats-rid="B1"><sup>1</sup></a> “Baked-in” biases, the ones due to people’s prejudices, perpetuate in the AI lifecycle and discriminate, mostly against disadvantaged groups (e.g., indigenous populations, women, people with disabilities, the elderly, etc.).<a class="reference-link xref xref-bibr" href="#B30" data-jats-ref-type="bibr" data-jats-rid="B30"><sup>30</sup></a> A notable example of historical (also known as <i>societal</i>) bias can be seen in a 2018 image search for “women CEOs,” which returned disproportionately few female results. This was not due to flaws in data sampling but rather because only about 5% of Fortune 500 CEOs were women at the time. The algorithm simply mirrored existing societal imbalances, highlighting how bias can be embedded even in “accurate” reflections of reality.<a class="reference-link xref xref-bibr" href="#B26" data-jats-ref-type="bibr" data-jats-rid="B26"><sup>26</sup></a></p>
<section id="sec7" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><em>Human heuristics mapping.</em>  The projection of the <i>representativeness</i> heuristic into AI <i>representation</i> bias is evident in [HA.P1] data-sampling cases. This critical issue has been flagged largely in AI face-recognition algorithms, which are typically trained on images of people from developed countries and consequently misidentify the faces of individuals from racial groups underrepresented in the training data.<a class="reference-link xref xref-bibr" href="#B3" data-jats-ref-type="bibr" data-jats-rid="B3"><sup>3</sup></a> Humans consistently overestimate the proportion of the world that is similar to themselves, and given that most AI agents are trained in developed countries, the selection of the sample reflects individuals who resemble the typical profile of people from those regions.<a class="reference-link xref xref-bibr" href="#B3" data-jats-ref-type="bibr" data-jats-rid="B3"><sup>3</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B25" data-jats-ref-type="bibr" data-jats-rid="B25"><sup>25</sup></a> Moreover, the <i>availability</i> heuristic often lies behind [HA.P1] sampling cases, due to the human tendency to place greater emphasis on the most readily available data, even if it does not fully represent the entire population.<a class="reference-link xref xref-bibr" href="#B17" data-jats-ref-type="bibr" data-jats-rid="B17"><sup>17</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B22" data-jats-ref-type="bibr" data-jats-rid="B22"><sup>22</sup></a> Datasets, though, may be available that seem representative but still exhibit entrenched biases. Datasets that derive from populations with baked-in <i>historical</i> biases lead to systems that reinforce patterns of discrimination. The representativeness heuristic is usually behind a lot of people’s bigotry, as this heuristic makes people believe a stereotype is true.<a class="reference-link xref xref-bibr" href="#B30" data-jats-ref-type="bibr" data-jats-rid="B30"><sup>30</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B36" data-jats-ref-type="bibr" data-jats-rid="B36"><sup>36</sup></a></p>
</section>
</section>
<section id="sec8" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>Harmful action [HA.P2]: Misuse of proxy variables resulting in computational bias: Measurement.</strong>  A proxy, that is, a concrete measurement assigned to a variable, is chosen when there is a need to approximate a “construct” (i.e, a concept or an entity) that is unobservable or immeasurable. Surely, the choice of appropriate proxies for an AI model is a challenging task. Proxies have been criticized for their role in AI <i>measurement</i> biases and their inaccurate and/or discriminatory classifier outcomes, since they either poorly reflect the target constructs or use different estimates between groups.<a class="reference-link xref xref-bibr" href="#B37" data-jats-ref-type="bibr" data-jats-rid="B37"><sup>37</sup></a> Especially when protected attributes (e.g., gender, ethnicity) are selected as proxies, challenges emerge, since these attributes partition a population into different groups that should be treated equally. A straightforward suggestion for AI developers would be to avoid using protected attributes (associated with historically discriminated-against social groups) as features in the training set, and move toward deploying AI models that do not rely on sensitive information and thus do not yield biased outcomes <i>(direct discrimination)</i>. Unfortunately, scientific evidence has shown that even removing such protected sensitive features does not prevent discrimination, as latent variables can be inferred implicitly (e.g., gender has been reported to be inferred through browsing history, and ethnicity can be inferred from a neighborhood zip code).<a class="reference-link xref xref-bibr" href="#B30" data-jats-ref-type="bibr" data-jats-rid="B30"><sup>30</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B37" data-jats-ref-type="bibr" data-jats-rid="B37"><sup>37</sup></a> Such <i>indirect discrimination</i> has occurred in scenarios [HA.P2] that set neutral or objective goals but end up treating groups or individuals less favorably, based on practices that disadvantage a protected group.</p>
<section id="sec9" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><em>Human heuristics mapping.</em>  Feature selection is made mostly by humans (e.g, AI developers), who might choose features that mostly align with their own beliefs. Thus, each AI developer’s own <i>representativeness</i> heuristic may influence their assumptions about the suitability of a proxy to reflect the actual (nonmeasurable) variable.<a class="reference-link xref xref-bibr" href="#B10" data-jats-ref-type="bibr" data-jats-rid="B10"><sup>10</sup></a> Similarly, the <i>availability</i> heuristic may cause a preference for proxies that seem related to the target variable, even though this perceived relationship does not always imply a true close correlation. Hidden variables may influence both the proxy and target variables, distorting their connection due to <i>illusory correlation</i> (identified earlier as an AI availability bias).<a class="reference-link xref xref-bibr" href="#B36" data-jats-ref-type="bibr" data-jats-rid="B36"><sup>36</sup></a> For example, a hiring manager might incorrectly assume that years of experience is an ideal proxy for predicting job performance. However, such a choice can reinforce gender bias. Men have historically had greater access to career paths and senior roles, which allowed them to accumulate more experience than women. Additionally, <i>confirmation</i> <i>bias</i> can also lead to the application of different measurement standards across groups, resulting in skewed outcomes. For instance, a fraud analyst may subject certain groups to more rigorous testing, which produces more positive cases and reinforces their initial beliefs, feelings, and prejudgements, ultimately distorting the true frequency of fraud in the overall population.<a class="reference-link xref xref-bibr" href="#B38" data-jats-ref-type="bibr" data-jats-rid="B38"><sup>38</sup></a></p>
</section>
</section>
</section>
<section id="sec10" class="sec">
<h2 class="heading">HAI-ROLL: In-Processing Phase</h2>
<p id="p-35">In this phase, the AI model is trained to optimize a specified objective (e.g., maximize a reward function in RLHF or minimize a loss function). Up to now, the AI model’s efficiency has been tested by using available (benchmark) datasets (different from the training data), and measuring, mostly, its accuracy. However, the emerging deployment of AI in sensitive and critical domains has given rise to a range of important desiderata; for example, apart from accuracy, AI models must also be tested for their capacity to preserve fairness and privacy.<a class="reference-link xref xref-bibr" href="#B18" data-jats-ref-type="bibr" data-jats-rid="B18"><sup>18</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B42" data-jats-ref-type="bibr" data-jats-rid="B42"><sup>42</sup></a> Thus, beyond the existing benchmarks used to estimate the model’s generalizability, it is now urgent to validate AI models under metrics that will assess their capacity to comply with legal and ethical rules.<a class="reference-link xref xref-bibr" href="#B28" data-jats-ref-type="bibr" data-jats-rid="B28"><sup>28</sup></a></p>
<section id="sec11" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>Harmful action [HA.I1]: Inappropriate model design choices resulting in computational bias: Algorithmic.</strong>  AI developers are responsible for critical design choices, from selecting the model’s architecture and objective function to configuring hyperparameters (e.g., external parameters such as the number of neural network layers or the training duration). [HA.I1] cases give rise to different algorithmic biases, primarily due to the following design-related pitfalls:</p>
<ul class="list" data-jats-list-type="bullet">
<li class="list-item">
<p id="p-37"><i>Choice of the model architecture and objective function</i>: Simple, underfitting models may fail to capture unique data patterns, while overfitting models, although closely fitting the training data, may adapt to noisy or unrepresentative training data.<a class="reference-link xref xref-bibr" href="#B18" data-jats-ref-type="bibr" data-jats-rid="B18"><sup>18</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B38" data-jats-ref-type="bibr" data-jats-rid="B38"><sup>38</sup></a> Model design choices become even more complex in the context of LLMs, due to their intricate architectures and multi-phase training pipelines, which can embed subtle forms of bias and amplify cognitive distortions.<a class="reference-link xref xref-bibr" href="#B20" data-jats-ref-type="bibr" data-jats-rid="B20"><sup>20</sup></a> In particular, fine-tuning techniques, originally developed to align these models (LLMs) with human values, have garnered substantial attention as methods for adapting generative systems by modifying their objective function. A key approach in this process is reward modeling (RM), in which a separate reward model is trained to approximate human preferences.<a class="reference-link xref xref-bibr" href="#B8" data-jats-ref-type="bibr" data-jats-rid="B8"><sup>8</sup></a> During reinforcement learning, this reward model effectively shapes the objective function that the main model is trained to optimize. However, because human judgments are influenced by cognitive heuristics and contextual biases, the reward model may end up favoring responses that appear more familiar, confident, or emotionally appealing, rather than those that are most accurate or helpful. As a result, the final model is optimized to reproduce these imperfect human preferences.<a class="reference-link xref xref-bibr" href="#B8" data-jats-ref-type="bibr" data-jats-rid="B8"><sup>8</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B9" data-jats-ref-type="bibr" data-jats-rid="B9"><sup>9</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B20" data-jats-ref-type="bibr" data-jats-rid="B20"><sup>20</sup></a></p>
</li>
<li class="list-item">
<p id="p-38"><i>Balancing multiple model requirements</i>: Achieving high <i>accuracy</i> while also ensuring <i>fairness</i> and <i>privacy</i> involves inherent trade-offs.<a class="reference-link xref xref-bibr" href="#B18" data-jats-ref-type="bibr" data-jats-rid="B18"><sup>18</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B42" data-jats-ref-type="bibr" data-jats-rid="B42"><sup>42</sup></a> For example, in cases where a loss function—typically minimized to reduce the difference between the algorithm’s output and the ground truth—is biased toward a specific demographic group, the resulting model may perform better for that group in terms of accuracy. However, this improvement does not extend to other critical parameters, such as fairness.<a class="reference-link xref xref-bibr" href="#B18" data-jats-ref-type="bibr" data-jats-rid="B18"><sup>18</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B42" data-jats-ref-type="bibr" data-jats-rid="B42"><sup>42</sup></a> Furthermore, in privacy-preserving analysis, a cutting-edge solution known as differential privacy enhances privacy by using techniques such as noise injection, which adds controlled noise to the data or output, masking the contribution of any individual in the dataset. However, this approach can disproportionately reduce the model’s accuracy for underrepresented subgroups, thereby affecting the model’s fairness<i>.</i><a class="reference-link xref xref-bibr" href="#B42" data-jats-ref-type="bibr" data-jats-rid="B42"><sup>42</sup></a></p>
</li>
<li class="list-item">
<p id="p-39"><i>Selection of more subtle parameters</i>: Fine-grained design choices, such as the duration of the training process or the learning rate, can significantly affect error rates, especially for underrepresented features, which may be disproportionately impacted by such adjustments. For example, early stopping (a technique used to prevent overfitting while preserving model accuracy) may inadvertently exacerbate biases. Since underrepresented features are typically learned later in the training process, selecting an inappropriate early stopping point can negatively impact specific subsets of the data population.<a class="reference-link xref xref-bibr" href="#B18" data-jats-ref-type="bibr" data-jats-rid="B18"><sup>18</sup></a></p>
</li>
</ul>
<section id="sec12" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><em>Human heuristics mapping.</em>  AI developers make subtle choices, seek patterns, and often substitute complex problems with simpler ones guided by the <i>representativeness</i> and <i>availability</i> heuristics.<a class="reference-link xref xref-bibr" href="#B21" data-jats-ref-type="bibr" data-jats-rid="B21"><sup>21</sup></a> For instance, the human preference for simpler models with fewer, more accessible parameters is linked to the availability heuristic.<a class="reference-link xref xref-bibr" href="#B29" data-jats-ref-type="bibr" data-jats-rid="B29"><sup>29</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B30" data-jats-ref-type="bibr" data-jats-rid="B30"><sup>30</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B36" data-jats-ref-type="bibr" data-jats-rid="B36"><sup>36</sup></a> Moreover, the <i>affect</i> heuristic can contribute to [HA.I1] cases and act as a source of <i>algorithmic</i> bias. For example, a developer may prioritize <i>privacy</i> in a healthcare system (based on prior experience or personal values), emotionally emphasizing data protection while neglecting <i>accuracy</i> concerns.<a class="reference-link xref xref-bibr" href="#B18" data-jats-ref-type="bibr" data-jats-rid="B18"><sup>18</sup></a> In the case of LLMs, empirical work has shown that cognitive heuristics are already instilled in them through their training data and in-context learning patterns.<a class="reference-link xref xref-bibr" href="#B32" data-jats-ref-type="bibr" data-jats-rid="B32"><sup>32</sup></a> For example, frontier LLMs exhibit the representativeness heuristic by mimicking frequent training patterns, even when these conflict with sound reasoning. Their strong in-context learning capabilities can also amplify the <i>anchoring</i> effect, whereby early prompt tokens disproportionately influence the model’s output.<a class="reference-link xref xref-bibr" href="#B32" data-jats-ref-type="bibr" data-jats-rid="B32"><sup>32</sup></a> RLHF-aligned models have been shown to exhibit even greater bias than their base counterparts.<a class="reference-link xref xref-bibr" href="#B20" data-jats-ref-type="bibr" data-jats-rid="B20"><sup>20</sup></a> Human annotators who tend to up-score responses that align with their own beliefs or feel more agreeable may introduce <i>affective</i> and <i>confirmation</i> bias into the reward model, thereby fostering sycophantic behavior in the aligned model (i.e., producing overly agreeable responses even when they are less factual).<a class="reference-link xref xref-bibr" href="#B33" data-jats-ref-type="bibr" data-jats-rid="B33"><sup>33</sup></a> Similarly, human raters tend to favor longer outputs (length anchoring), prompting the model to equate verbosity with quality.<a class="reference-link xref xref-bibr" href="#B34" data-jats-ref-type="bibr" data-jats-rid="B34"><sup>34</sup></a> Ultimately, human heuristics cause models to favor agreeable or verbose outputs over factual or fair ones. Once embedded in the reward model, these biases are amplified during fine-tuning as the model learns to optimize for flawed human preferences.<a class="reference-link xref xref-bibr" href="#B8" data-jats-ref-type="bibr" data-jats-rid="B8"><sup>8</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B20" data-jats-ref-type="bibr" data-jats-rid="B20"><sup>20</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B33" data-jats-ref-type="bibr" data-jats-rid="B33"><sup>33</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B34" data-jats-ref-type="bibr" data-jats-rid="B34"><sup>34</sup></a> Furthermore, as models grow in size and capability, they become more susceptible to internalizing human-like cognitive biases, due to their increased capacity to capture and reproduce subtle patterns (including flawed reasoning) from training data and human feedback. This enhanced ability to mimic both rational and irrational human behavior creates a trade-off: greater fluency and alignment with human preferences, but also a deeper entrenchment of cognitive distortions.<a class="reference-link xref xref-bibr" href="#B20" data-jats-ref-type="bibr" data-jats-rid="B20"><sup>20</sup></a></p>
</section>
</section>
<section id="sec13" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>Harmful action [HA.I2]: Use of inappropriate (and non-representative) benchmarks and performance metrics resulting in computational bias: Evaluation.</strong>  In this AI phase, performance is assessed by using unseen data; biased results might occur when such testing datasets are non-representative of the actual population. To prevent [HA.I2] cases, the overall population must be represented and the AI model should perform well not only for a subset of the population. For example, a Named Entity Recognition (NER) model trained on data with more male names will perform well on a male-dominant benchmark dataset, but once a more balanced gender representative benchmark dataset is used, female misclassification will be revealed. Moreover, AI model evaluation based on metrics such as accuracy or precision has been heavily criticized, since these metrics do not address fairness across groups.<a class="reference-link xref xref-bibr" href="#B28" data-jats-ref-type="bibr" data-jats-rid="B28"><sup>28</sup></a> Subgroup validity approaches aim to ensure fairness by examining performance metrics across different subgroups. They use protected attributes to provide analysts with essential group information, indicating whether an individual belongs to a protected or unprotected group.<a class="reference-link xref xref-bibr" href="#B28" data-jats-ref-type="bibr" data-jats-rid="B28"><sup>28</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B35" data-jats-ref-type="bibr" data-jats-rid="B35"><sup>35</sup></a> However, generalizing a fairness definition and then mathematically representing it remains too complex of a task, since the numerous fairness metrics proposed up to now cannot be satisfied simultaneously and do not explore human-centric sources, which might result in emerging AI biases.<a class="reference-link xref xref-bibr" href="#B28" data-jats-ref-type="bibr" data-jats-rid="B28"><sup>28</sup></a></p>
<section id="sec14" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><em>Human heuristics mapping.</em>  Since humans (AI developers) make the decision on how to split the overall dataset into training and testing benchmark sets, their choices might cause [HA.I2] faults, due to their <i>representativeness</i> and <i>availability</i> heuristics (as in [HA.P1] above). Prominent studies have also reported that <i>confirmation bias</i> in AI software testers leads them to prioritize evaluation schemes with testing datasets and metrics that confirm, rather than question, the correct function of an AI model.<a class="reference-link xref xref-bibr" href="#B5" data-jats-ref-type="bibr" data-jats-rid="B5"><sup>5</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B29" data-jats-ref-type="bibr" data-jats-rid="B29"><sup>29</sup></a> Guided by their <i>anchoring and adjustment</i> heuristic, human AI testers often get stuck in their belief that the algorithm is accurate and focus on scenarios or inputs that are consistent with their initial assumptions, rather than exploring a broader range of possibilities that could challenge their initial beliefs.<a class="reference-link xref xref-bibr" href="#B5" data-jats-ref-type="bibr" data-jats-rid="B5"><sup>5</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B25" data-jats-ref-type="bibr" data-jats-rid="B25"><sup>25</sup></a></p>
</section>
</section>
</section>
<section id="sec15" class="sec">
<h2 class="heading">HAI-ROLL: Post-Processing Phase</h2>
<p id="p-43">In this AI phase, model outcomes are used and interpreted, and the model is refined on the basis of user reactions. This iterative process triggers feedback loops, where a model’s output serves as new input for further refinements (e.g., by retraining) and reevaluation of the AI lifecycle. These loops aim to make AI models adaptive and dynamic under varying conditions.<a class="reference-link xref xref-bibr" href="#B25" data-jats-ref-type="bibr" data-jats-rid="B25"><sup>25</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B37" data-jats-ref-type="bibr" data-jats-rid="B37"><sup>37</sup></a> However, reliance on previous AI outputs poses the risk of perpetuating biases, leading to the “butterfly effect,”<a class="reference-link xref xref-bibr" href="#B15" data-jats-ref-type="bibr" data-jats-rid="B15"><sup>15</sup></a> where the origins of bias become unclear and untraceable, complicating AI fairness resilience.</p>
<section id="sec16" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>Harmful action [HA.D1]: Use of the AI model in scenarios that deviate from its initial scope resulting in computational bias: Deployment.</strong>  In the validation phase, the original intended scope of the AI model must be carefully considered against its actual use.<a class="reference-link xref xref-bibr" href="#B30" data-jats-ref-type="bibr" data-jats-rid="B30"><sup>30</sup></a> The <i>deployment</i> bias in this phase occurs due to [HA.D1] cases that highlight discrepancies in the AI model’s use when compared to the original scope for which the model was designed. For example, the wide range of AI models, from strictly automated decision making without human involvement to AI that supports human decision makers,<a class="reference-link xref xref-bibr" href="#B19" data-jats-ref-type="bibr" data-jats-rid="B19"><sup>19</sup></a> embeds risky [HA.D1] cases. Such risks emerge when humans blindly put their faith in automated AI decision making, while the AI model was built only for a human-supportive scope.<a class="reference-link xref xref-bibr" href="#B19" data-jats-ref-type="bibr" data-jats-rid="B19"><sup>19</sup></a> This deployment bias in AI-assisted settings, often rooted in anchoring bias, is commonly referred to as <i>automation</i> bias.<a class="reference-link xref xref-bibr" href="#B29" data-jats-ref-type="bibr" data-jats-rid="B29"><sup>29</sup></a></p>
<section id="sec17" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><em>Human heuristics mapping.</em>  The <i>anchoring and adjustment</i> heuristic lies behind several [HA.D1] cases that cause <i>deployment</i> bias.<a class="reference-link xref xref-bibr" href="#B29" data-jats-ref-type="bibr" data-jats-rid="B29"><sup>29</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B30" data-jats-ref-type="bibr" data-jats-rid="B30"><sup>30</sup></a> <i>Confirmation</i> bias occurs when decision makers favor AI-generated results that support their preexisting beliefs (anchors), forcing them to disregard other potentially more relevant or accurate results that may contradict their assumptions.<a class="reference-link xref xref-bibr" href="#B29" data-jats-ref-type="bibr" data-jats-rid="B29"><sup>29</sup></a> Decision makers often fall into [HA.D1] faults by relying on AI decision-support systems without fully understanding their inner workings. Several studies have highlighted that domain experts who make decisions tend to overestimate their understanding of a domain, while not being aware of the underlying (opaque) AI and data science tasks that impact their decisions.<a class="reference-link xref xref-bibr" href="#B22" data-jats-ref-type="bibr" data-jats-rid="B22"><sup>22</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B30" data-jats-ref-type="bibr" data-jats-rid="B30"><sup>30</sup></a> The <i>affect</i> heuristic has also been identified as a driver of [HA.D1] cases, since the more human-like a system appears, the more likely it is that users attribute more human traits and capabilities to it. Automated systems often foster a sense of security, easing anxiety and increasing trust; such emotional comfort overshadows critical evaluation.<a class="reference-link xref xref-bibr" href="#B29" data-jats-ref-type="bibr" data-jats-rid="B29"><sup>29</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B30" data-jats-ref-type="bibr" data-jats-rid="B30"><sup>30</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B39" data-jats-ref-type="bibr" data-jats-rid="B39"><sup>39</sup></a> A common example is the so-called AI anthropomorphic conversational agents—LLM applications that engage users through dialogue. When people interact with LLMs, they may start to perceive them as human-like, share more information than they normally would, and rely on them more than they should.<a class="reference-link xref xref-bibr" href="#B39" data-jats-ref-type="bibr" data-jats-rid="B39"><sup>39</sup></a></p>
</section>
</section>
</section>
<section id="sec18" class="sec">
<h2 class="heading">HAI-ROLL in Practice: Implementation Guidelines and Limitations</h2>
<p id="p-46">Adopting the HAI-ROLL framework in real-world AI systems demands cautious configurations and deep understanding of its benefits and limitations. Given that it is the first framework (to our knowledge) that systematically explores the continuum from human heuristics to AI biases, HAI-ROLL is able to balance societal (human) and technical (AI) constraints and requirements. AI professionals (system designers and developers) are more aware of biases embedded in data or algorithms, basing their choices on existing fairness metrics and mitigation techniques.<a class="reference-link xref xref-bibr" href="#B2" data-jats-ref-type="bibr" data-jats-rid="B2"><sup>2</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B6" data-jats-ref-type="bibr" data-jats-rid="B6"><sup>6</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B14" data-jats-ref-type="bibr" data-jats-rid="B14"><sup>14</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B28" data-jats-ref-type="bibr" data-jats-rid="B28"><sup>28</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B37" data-jats-ref-type="bibr" data-jats-rid="B37"><sup>37</sup></a> As a result, they are less conscious of their own cognitive biases’ impact in shaping their decisions during the AI pre-, in-, and post- processing phases.<a class="reference-link xref xref-bibr" href="#B6" data-jats-ref-type="bibr" data-jats-rid="B6"><sup>6</sup></a> HAI-ROLL offers AI developers an actionable guideline service to help them critically reflect on how cognitive (often unconscious) heuristics shape AI-lifecycle technical choices. At the same time, HAI-ROLL offers actionable guidance to policymakers, who will recognize the importance of a complete sociotechnical AI framework. HAI-ROLL supports AI developers in anticipating where biases may emerge and in taking preemptive action, while reinforcing their bias-aware critical thinking skills, encouraging them to shift to a reflective engineering mindset grounded in both technical evidence and psychological insights. In summary, HAI-ROLL offers a paradigm shift, revisiting human and AI synergies at each phase of the real-world AI lifecycle. In the following, we discuss the next steps forward.</p>
<section id="sec19" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>HAI-ROLL as a pre-processing driver.</strong>  Developers may identify cognitive distortions and blind spots (overlooked gaps where certain subpopulations, contextual nuances, or relevant attributes are misrepresented or excluded), which lead to representational and historical biases in data-collection choices.<a class="reference-link xref xref-bibr" href="#B4" data-jats-ref-type="bibr" data-jats-rid="B4"><sup>4</sup></a> By encouraging cognitively driven dataset audits (e.g., through checklists<a class="reference-link xref xref-bibr" href="#B7" data-jats-ref-type="bibr" data-jats-rid="B7"><sup>7</sup></a>) and structured documentation (e.g., with datasheets<a class="reference-link xref xref-bibr" href="#B14" data-jats-ref-type="bibr" data-jats-rid="B14"><sup>14</sup></a>), HAI-ROLL supports more inclusive, context-sensitive, and cognitive-aware design and enhances bias anticipation.<a class="reference-link xref xref-bibr" href="#B2" data-jats-ref-type="bibr" data-jats-rid="B2"><sup>2</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B4" data-jats-ref-type="bibr" data-jats-rid="B4"><sup>4</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B14" data-jats-ref-type="bibr" data-jats-rid="B14"><sup>14</sup></a> It further supports the proper selection of proxy variables (during feature engineering) by overcoming the current explainable AI (XAI) methods (e.g., SHAP and LIME), which indicate feature importance but fail to explain why particular proxies were chosen or assess their social and ethical validity.<a class="reference-link xref xref-bibr" href="#B12" data-jats-ref-type="bibr" data-jats-rid="B12"><sup>12</sup></a> HAI-ROLL highlights that explanations must align with human reasoning strategies and heuristics,<a class="reference-link xref xref-bibr" href="#B29" data-jats-ref-type="bibr" data-jats-rid="B29"><sup>29</sup></a> and reveals how cognitive biases (such as illusory correlation and confirmation) can distort feature selection, prompting developers to critically examine the rationale behind their design decisions.</p>
</section>
<section id="sec20" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>HAI-ROLL as an in-processing monitor.</strong>  By safeguarding socially fair or ethically sound AI methods, which currently focus only on technical optimizations (e.g., hyperparameter tuning, adversarial training) mostly used to boost accuracy and robustness,<a class="reference-link xref xref-bibr" href="#B14" data-jats-ref-type="bibr" data-jats-rid="B14"><sup>14</sup></a> HAI-ROLL encourages AI professional teams to reflect on whether their AI model design defaults are driven by fairness considerations or merely by AI in-processing simplicity. The emerging alignment techniques (such as RLHF) in LLMs and the new forms of bias underscore the role of HAI-ROLL in supporting bias-mitigation strategies.<a class="reference-link xref xref-bibr" href="#B20" data-jats-ref-type="bibr" data-jats-rid="B20"><sup>20</sup></a> By addressing concerns around both human and AI biases, HAI-ROLL offers insights to rethink ways of collecting human feedback, which must always include diversifying annotator perspectives.<a class="reference-link xref xref-bibr" href="#B9" data-jats-ref-type="bibr" data-jats-rid="B9"><sup>9</sup></a> Given that RLHF reward models may easily reinforce responses that are emotionally appealing or agreeable rather than logically sound,<a class="reference-link xref xref-bibr" href="#B8" data-jats-ref-type="bibr" data-jats-rid="B8"><sup>8</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B33" data-jats-ref-type="bibr" data-jats-rid="B33"><sup>33</sup></a> HAI-ROLL introduces a complementary practical layer for new sociotechnical insights, to better train annotators and guide reward-model audits. By explicitly identifying the heuristics that often shape human preferences (such as affect and confirmation), HAI-ROLL strongly supports developers in anticipating where feedback may introduce cognitive distortions into the reward model. During the AI model evaluation phase, cognitive biases can distort both the selection and interpretation of benchmarks and metrics.<a class="reference-link xref xref-bibr" href="#B11" data-jats-ref-type="bibr" data-jats-rid="B11"><sup>11</sup></a> Empirical audits have revealed frequent inconsistencies in standards and overly narrow definitions of fairness.<a class="reference-link xref xref-bibr" href="#B4" data-jats-ref-type="bibr" data-jats-rid="B4"><sup>4</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B11" data-jats-ref-type="bibr" data-jats-rid="B11"><sup>11</sup></a> Evaluation methods in real-world scenarios are often constrained by questionable assumptions (such as the availability of reliable, sensitive attributes or clear group boundaries), which can lead to evaluation bias and undermine fairness guarantees.<a class="reference-link xref xref-bibr" href="#B4" data-jats-ref-type="bibr" data-jats-rid="B4"><sup>4</sup></a> HAI-ROLL acknowledges how human reasoning and contextual AI settings impact fairness judgments. It offers a novel sociotechnical approach to in-processing monitoring by enabling bias-aware tools (e.g., mind maps and checklists), which will facilitate more-reflective evaluation practices and inspire new synergies between AI professionals and policymakers.</p>
</section>
<section id="sec21" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>HAI-ROLL as a post-processing auditor.</strong>  The framework addressess AI fairness mitigation as a shared responsibility, and as a collective and collaborative synergy between AI developers, ethicists, domain experts, and, crucially, policymakers. In practice, deployment bias occurs when AI systems (despite careful design) are misused and over-relied on, or are misunderstood by real-world adopters, since cognitive biases persist even when explainability tools exist.<a class="reference-link xref xref-bibr" href="#B4" data-jats-ref-type="bibr" data-jats-rid="B4"><sup>4</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B6" data-jats-ref-type="bibr" data-jats-rid="B6"><sup>6</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B29" data-jats-ref-type="bibr" data-jats-rid="B29"><sup>29</sup></a> For example, displaying confidence scores may lead to over-trust and blind acceptance of flawed outputs,<a class="reference-link xref xref-bibr" href="#B29" data-jats-ref-type="bibr" data-jats-rid="B29"><sup>29</sup></a> since AI accuracy and transparency alone do not ensure fairness, and do not anticipate human heuristics. In real-world AI implementations, humans rely on cognitive shortcuts, and they may rubber-stamp low-quality outputs or gradually stop questioning the system as their trust increases. Such unexamined decisions may feed back into AI training data (pre-processing phases), thus perpetuating potential discrimination through AI.<a class="reference-link xref xref-bibr" href="#B29" data-jats-ref-type="bibr" data-jats-rid="B29"><sup>29</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B37" data-jats-ref-type="bibr" data-jats-rid="B37"><sup>37</sup></a> HAI-ROLL can assist in reflecting such biases to specific metrics, such as a “blind accept” rate, that will enable interdisciplinary teams (AI developers along with policymakers) to build dashboards, trigger justification prompts, and set alert thresholds to detect drift early. HAI-ROLL may also be adopted into certification processes, to set the requirements that will promote fair and accountable AI development. For example, under Article 9 of the EU AI Act, providers must implement continuous risk management.<a class="reference-link xref xref-bibr" href="#B13" data-jats-ref-type="bibr" data-jats-rid="B13"><sup>13</sup></a> Based on the HAI-ROLL framework, a Cognitive Bias Impact Assessment (CBIA)—similar to Data Protection Impact Assessments (DPIAs) under the GDPR<a class="reference-link xref xref-bibr" href="#B41" data-jats-ref-type="bibr" data-jats-rid="B41"><sup>41</sup></a>—could be appended to this process, requiring teams to identify which cognitive heuristics are likely to surface at each stage and to document appropriate countermeasures. In addition, HAI-ROLL lends itself well to policy integration, as its human-AI bias mapping can be translated into auditable performance indicators. For example, it may leverage indicators such as a potential <i>blind-spot score</i> to measure the share of out-of-distribution records detected during sampling tests, or a <i>blind-accept rate</i> to track how often humans rubber-stamp automated outputs into post-deployment monitoring and audit protocols.<a class="reference-link xref xref-bibr" href="#B4" data-jats-ref-type="bibr" data-jats-rid="B4"><sup>4</sup></a></p>
</section>
<section id="sec22" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>Limitations and potential uptake constraints.</strong>  Certain limitations should be recognized prior to institutionalizing HAI-ROLL in practice. Building fairness-aware AI systems requires more than technical soundness; it calls for practices that are sustainable and grounded in continuous reflection on the social and ethical impacts of design choices. Cultural and organizational change, along with synchronized activities (educational programs, awareness campaigns, and cognitive-bias-informed design protocols), are needed to enable AI professionals and policymakers to agree on the roots of AI unfairness. Equally important is the promotion of inclusivity and diversity within AI development teams to counteract groupthink and foster perspectives that better detect and address bias.<a class="reference-link xref xref-bibr" href="#B40" data-jats-ref-type="bibr" data-jats-rid="B40"><sup>40</sup></a> HAI-ROLL offers a conceptual framework and opens research dialogues to move away from current computationally focused fairness solutions, by recognizing that even AI models that are carefully designed to be fair might still remain biased due to human harms creeping into the overall AI lifecycle. The proposed framework is based on extensive study of  state-of-the-art evidence from cognitive psychology, data science, and AI, and rolls in the deeper understanding of the causes and effects of this human-AI confluence. HAI-ROLL and the proposed work are not exhaustive; rather, they offer a forward-looking and scientifically grounded foundation. The practical readiness of such bias-aware AI systems should be further validated through thorough empirical studies. Adoption of HAI-ROLL would require domain-specific case studies and an interdisciplinary research consensus to practically verify its scientific evidence and advance its systematic mapping in real-world settings. As discussed earlier, we claim to offer a stepping stone to widen the study of human biases in FairAI by surfacing overlooked connections between cognitive heuristics and computational biases, and by highlighting meaningful patterns of influence and interdependencies. Our ongoing work on FairAI focuses on extending HAI-ROLL mapping with a holistic human-inclusive taxonomy, which will be theoretically grounded and experimentally tested under varying bias-sensitive constraints posed in several critical FairAI domains. Future uptake steps should include empirical validation of each mapping proposition by designing experiments that compare decision making under the influence of a cognitive heuristic to a condition in which no such biasing influence is present (i.e., a neutral baseline), similar to control-versus-treatment setups.<a class="reference-link xref xref-bibr" href="#B20" data-jats-ref-type="bibr" data-jats-rid="B20"><sup>20</sup></a> As a starting point, we propose eliciting the heuristic through realistic design vignettes, logging the resulting design choices, and auditing the downstream model for the predicted computational bias—thereby providing converging evidence for “heuristic → harmful action → AI bias” chains in real-world scenarios and applications. By identifying the potential sources of bias at an early stage of the AI lifecycle, HAI-ROLL inspires the design of effective bias-mitigation approaches. Our proposition does not suggest that the hypotheses or assumptions of other researchers are wrong. Rather, we leverage existing cognitive biases and FairAI scientific outcomes to bridge gaps and initiate a deeper exploration of the complex human-to-AI bias “iceberg.”</p>
</section>
</section>
</div>
<footer class="back"></footer>
</article>
]]></content:encoded>
					
					<wfw:commentRss>https://cacm.acm.org/research/rolling-in-the-deep-of-cognitive-and-ai-biases/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
		      <dc:creator><![CDATA[Athena Vakali]]></dc:creator>
<post-id xmlns="com-wordpress:feed-additions:1">779047</post-id>	</item>
		<item>
		<title>Are We Actually There? Assessing RPKI Maturity</title>
		<link>https://cacm.acm.org/research/are-we-actually-there-assessing-rpki-maturity/</link>
					<comments>https://cacm.acm.org/research/are-we-actually-there-assessing-rpki-maturity/#respond</comments>
		
		<dc:creator><![CDATA[Haya Schulmann, Niklas Vogel, and Michael Waidner]]></dc:creator>
		<pubDate>Tue, 21 Apr 2026 19:59:07 +0000</pubDate>
				<category><![CDATA[Security and Privacy]]></category>
		<category><![CDATA[Systems and Networking]]></category>
		<guid isPermaLink="false">https://cacm.acm.org/?post_type=digital-library&#038;p=779184</guid>

					<description><![CDATA[<p>RPKI lacks maturity in its specification and in resilience against attacks, but delivers measurable benefits and is an essential part of ongoing efforts to improve routing security.</p>]]></description>
										<content:encoded><![CDATA[<article>
<div class="body" lang="en">
<section id="sec1" class="sec">
<p id="p-1">The Border Gateway Protocol (BGP) underlies all modern Internet communication by enabling the exchange of IP routing information between autonomous systems (ASes). Despite its criticality, BGP is insecure by design, and to this day attacks on BGP are frequent and harmful.<a class="reference-link xref xref-bibr" href="#B8" data-jats-ref-type="bibr" data-jats-rid="B8"><sup>8</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B16" data-jats-ref-type="bibr" data-jats-rid="B16"><sup>16</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B38" data-jats-ref-type="bibr" data-jats-rid="B38"><sup>38</sup></a> The Resource Public Key Infrastructure (RPKI) was standardized to add cryptographic security to BGP, allowing ASes to validate BGP announcements, for example, through route origin authorizations (ROAs).<a class="reference-link xref xref-bibr" href="#B23" data-jats-ref-type="bibr" data-jats-rid="B23"><sup>23</sup></a> The publication of RFC6811 introduced route origin validation (ROV), enabling ASes to use ROAs to verify whether a given BGP announcement was originated by the legitimate holder of a prefix. RPKI also supports other validations, such as BGP path validation through BGPsec and route leak detection through AS provider authorizations (ASPAs). This ability to validate legitimacy guides routers in their decisions and prevents attacks on BGP.</p>
<aside class="boxed-text">
<div class="article-key-insights">
<h2>Key Insights</h2>
<ul class="list" data-jats-list-type="bullet">
<li class="list-item">
<p id="p-2">RPKI is essential to protect global routing and is reaching widespread deployment. However, RPKI still has many practical issues, putting its readiness for adoption into question.</p>
</li>
<li class="list-item">
<p id="p-3">The most pressing issues in RPKI are vagueness and conflicting requirements in specification, bugs and vulnerabilities in implementations, and errors and scalability issues in deployment.</p>
</li>
<li class="list-item">
<p id="p-4">These issues need to be addressed to achieve the desired maturity of a widely deployed security architecture. However, deploying non-perfect RPKI is better than not deploying RPKI at all. </p>
</li>
</ul>
</div>
</aside>
<section id="sec2" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>Chronicles of RPKI.</strong>  Upon its introduction,<a class="reference-link xref xref-bibr" href="#B1" data-jats-ref-type="bibr" data-jats-rid="B1"><sup>1</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B5" data-jats-ref-type="bibr" data-jats-rid="B5"><sup>5</sup></a> RPKI was considered an experimental technology and its adoption was sparse. After 2010, RPKI began transitioning from a theoretical concept to practical implementations. Early software development was often led by research institutions and individual engineers working on open source RPKI tools. In 2011, the regional Internet registry for Europe, RIPE NCC, launched one of the first tools for validating ROAs, the RPKI Validator.<a class="reference-link xref xref-bibr" href="#B40" data-jats-ref-type="bibr" data-jats-rid="B40"><sup>40</sup></a> During this early phase in 2013, some institutions, such as the Center for Applied Internet Data Analysis (CAIDA) and RIPE NCC, as well as a number of ISPs and carriers, experimented with RPKI. While interest was growing, the technology had been deployed only in experimental or partial configurations by about a dozen ISPs globally, mainly in testbeds and pilot environments to gauge the impact on routing without affecting real traffic.<a class="reference-link xref xref-bibr" href="#B32" data-jats-ref-type="bibr" data-jats-rid="B32"><sup>32</sup></a> RIPE NCC and the American Registry for Internet Numbers (ARIN) were among the first to provide RPKI services for operators to create and publish ROAs. These efforts focused on building the infrastructure for certificate issuance and ROA validation. After 2016, more open source relying party (RP) software packages were created, offering operators more variety in adopting RPKI. Nevertheless, RPKI implementations were still mostly research prototypes rather than stable production software. In 2018, major Internet stakeholders, such as the regional Internet registries (RIRs), began promoting RPKI adoption and large networks, service providers, and content delivery networks (CDNs) began testing it. In a surprising step, Cloudflare announced full support for RPKI validation in 2018, marking a significant milestone as one of the first large public networks to enable ROV in a production environment.<a class="reference-link xref xref-bibr" href="#B11" data-jats-ref-type="bibr" data-jats-rid="B11"><sup>11</sup></a> In 2020, the Mutually Agreed Norms for Routing Security (MANRS) initiative pushed for widespread ROV adoption, emphasizing routing security and promoting best practices. RIPE NCC also advocated for RPKI deployment and ROV enforcement. The following year, more large providers, including Google and Amazon, began implementing ROV, helping to drive global adoption and encouraging other networks to follow suit. A notable increase in ROAs publication was observed globally in 2022. This allowed more networks to validate the authenticity of BGP announcements, even though full ROV enforcement across all networks was not yet in place. These milestones highlight the significance RPKI has gained over the past decade. Starting off as an experimental technology, it has become a central component of the Internet, affecting a large percentage of its networks. Today, more than 50% of announced prefixes are covered with ROAs, and about 25% of networks enforce ROV.<a class="reference-link xref xref-bibr" href="#B21" data-jats-ref-type="bibr" data-jats-rid="B21"><sup>21</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B28" data-jats-ref-type="bibr" data-jats-rid="B28"><sup>28</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B33" data-jats-ref-type="bibr" data-jats-rid="B33"><sup>33</sup></a></p>
</section>
<section id="sec3" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>Growing political significance.</strong>  The U.S. was one of the first countries to recognize the threat that vulnerable Internet routing poses to national security and the necessity of using RPKI to secure its inter-domain BGP routing. As such, it listed routing security as one of the main items on its cybersecurity agenda in March 2023.<a class="reference-link xref xref-bibr" href="#B39" data-jats-ref-type="bibr" data-jats-rid="B39"><sup>39</sup></a> Shortly after, in September 2024, the White House issued a strategic roadmap. The roadmap establishes a plan of action and promotes the use of RPKI in all U.S. networks, with the goal of improving routing security against attacks.<a class="reference-link xref xref-bibr" href="#B35" data-jats-ref-type="bibr" data-jats-rid="B35"><sup>35</sup></a> It complements a Notice of Proposed Rulemaking (NPRM) by the U.S. Federal Communications Commission (FCC), published a few months earlier in May 2024. Both documents provide strategic and policy frameworks as well as technical steps, with a focus on compliance and operation. The White House&#8217;s roadmap identifies RPKI as a mature, ready-to-implement technology to mitigate vulnerabilities in BGP, and recommends deploying it on all networks: </p>
<p id="p-7"><i>The roadmap released today advocates for the adoption of Resource Public Key Infrastructure (RPKI) as a mature, ready-to-implement approach to mitigate vulnerabilities in BGP.</i></p>
<p id="p-8">The roadmap considers RPKI technology to be readily available and traces the challenges to RPKI adoption to three factors. First, decisionmakers lack a thorough understanding of Internet routing security risks. Second, network operators do not prioritize and do not have resources to deploy new BGP security mechanisms. And third, organizations encounter administrative barriers with RIRs during adoption. The roadmap concludes that these challenges contribute to a reluctance to prioritize routing security.</p>
</section>
<section id="sec4" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>Recommendations for extending the roadmap.</strong>  In this article, we analyze additional gaps that need to be bridged to facilitate broad RPKI adoption. Our analysis covers standard specifications, software packages, operation, and deployment. We propose enhancing the roadmap to also include these concrete items. We use the insights and observations from our research to derive recommendations for the different stakeholders in RPKI and routing security to provide a path forward for securing the technology. We recommend first collecting experience by operating RPKI in strict validation mode to refine specification considerations, then transitioning networks to production-mode RPKI validation. We also recommend refining the RPKI standards to remove conflicting or underspecified requirements, thus making clear recommendations for developers and operators of RPKI. We further recommend investing in developing RPKI-specific automated tools for software developers and operators, as well as prioritizing and automating patch management. We suggest updating the guidelines and strategies to reflect the full threat landscape of RPKI, as the current roadmap guidelines focus exclusively on RPKI operation under benign conditions. We further suggest considering the resilience of RPKI under malicious attacks. Finally, we recommend looking at the increasing attack surface associated with networks deploying RPKI that arises from exploitation of not only benign bugs but also intentional backdoors. Improving routing security is a global and complex effort. We encourage more countries to contribute to the goal of securing Internet routing with RPKI, and hope our recommendations will inform and guide their efforts.</p>
</section>
</section>
<section id="sec5" class="sec">
<h2 class="heading">Overview of BGP Security</h2>
<p id="p-10">BGP lacks cryptographic authentication of announcements, which enables a range of routing attacks. RPKI was designed to mitigate attacks on BGP by adding cryptographic security to the insecure BGP protocol.<a class="reference-link xref xref-bibr" href="#B25" data-jats-ref-type="bibr" data-jats-rid="B25"><sup>25</sup></a> RPKI provides an architecture to distribute verifiable BGP information globally. The <a href="#F1">figure</a> provides an overview of RPKI from the perspective of a system operator. The red color illustrates untrusted RPKI components controlled by third parties, while the green color marks components trusted by the system and running within the local network.</p>
<figure id="F1" class="fig" data-jats-position="float">
<div class="image-container"><img loading="lazy" decoding="async" class="graphic alignnone" title="Figure. " src="https://cacm.acm.org/wp-content/uploads/2026/03/3769687_fig01.jpg" alt="" width="1227" height="653" data-image-id="F1" data-image-type="figure" /></div><figcaption><span class="caption-label">Figure. </span><span class="p">Overview of RPKI.</span></p>
<div class="figcaption-footer"> </div>
</figcaption></figure>
<section id="sec6" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>RPKI repository.</strong>  All BGP information in RPKI is stored inside distributed repository servers. RPKI offers two hosting models: In hosted mode, participating systems create their RPKI objects within a repository operated by one of the five RIRs, while delegated mode allows systems to host their own servers, allowing full control over created objects. Systems accessing RPKI repositories should treat them as untrusted entities, and the validity of contained objects should be assumed only through cryptographic validation (RFC6481).</p>
</section>
<section id="sec7" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>RPKI objects.</strong>  RPKI offers a range of objects that systems can upload to RPKI repositories to provide data for BGP routing. Most prominently, ROAs contain the valid BGP origin of a set of IP prefixes; systems can use ROAs to verify if the origin of a given BGP announcement has been authorized by the owner of the announced prefix, protecting against origin hijacks.<a class="reference-link xref xref-bibr" href="#B26" data-jats-ref-type="bibr" data-jats-rid="B26"><sup>26</sup></a> Additional objects include BGPsec certificates<a class="reference-link xref xref-bibr" href="#B27" data-jats-ref-type="bibr" data-jats-rid="B27"><sup>27</sup></a> that contain a public key to verify a signed BGP path, preventing path hijacks, and AS provider authorizations (ASPAs) that verify the valid provider set of a system to prevent route leaks.<a class="reference-link xref xref-bibr" href="#B2" data-jats-ref-type="bibr" data-jats-rid="B2"><sup>2</sup></a> Further, RPKI defines objects that ensure the integrity of other objects and provide attestation of resource ownership, for example, to allow verification if a system is authorized to issue ROAs for a set of IP prefixes. This resource ownership is verified over X.509 certificates that bind a cryptographic key to IP resources and are signed by the parent that issued the IP resources.<a class="reference-link xref xref-bibr" href="#B22" data-jats-ref-type="bibr" data-jats-rid="B22"><sup>22</sup></a></p>
</section>
<section id="sec8" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>Relying party.</strong>  To reduce computational load and implementation complexity, BGP routers do not directly interact with RPKI repositories to download and validate the RPKI objects. Instead, systems install a relying party (RP) client as middleware between the RPKI repository system and their routers. The RP handles all interactions with the repositories; it downloads all objects from the available RPKI repositories, validates their integrity and cryptographic signatures, and compiles a list of all RPKI data for the routers. RPs recursively iterate all RPKI repositories on the Internet, using hardcoded URIs of the five hosted repositories stored in trust anchor locators (TALs) for bootstrapping before iteratively querying the objects from all delegated repositories. Thus, each live RPKI repository will be regularly contacted by all globally running RP clients to download the contained RPKI objects. Since the RP is responsible for validation and inherently trusted, it should be installed within a trusted environment such as the local network.</p>
</section>
<section id="sec9" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>BGP router.</strong>  All BGP routers in an AS that want to use RPKI information regularly poll the RP to download the validated RPKI data over the RPKI to Router (RTR) protocol. Routers thus heavily depend on the availability of the RP; <i>if the RP is unavailable, routers will eventually downgrade to regular BGP.</i><a class="reference-link xref xref-bibr" href="#B3" data-jats-ref-type="bibr" data-jats-rid="B3"><sup>3</sup></a> Thus, RPs need to ensure the availability of routers. Further, routers do not conduct any cryptographic validations of the data provided to them by their RP; RPs are fully trusted. Due to the availability and trust requirements of RPs, it is generally recommended to place them close to routers, preventing disruptions of the connection and reducing the threat of communications being manipulated.<a class="reference-link xref xref-bibr" href="#B4" data-jats-ref-type="bibr" data-jats-rid="B4"><sup>4</sup></a> After downloading the RPKI information, routers may use the data in their routing decisions. For example, they should use the ROA data to validate the origin in received BGP announcements, and discard any announcements that conflict with an ROA. Similarly, the routers may use ASPAs to reject route-leaking announcements and BGPsec keys to validate received BGPsec paths to protect against path hijacks.</p>
</section>
<section id="sec10" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>Maturity of RPKI technology.</strong>  The White House roadmap uses the term <i>mature</i> with respect to RPKI in a rather informal way. In general, the maturity of a complex technology like RPKI is challenging to assess, as maturity goes beyond the number of vulnerabilities or the deployment status. To assess the maturity of RPKI, we will review its resilience and stability through an analysis of RPKI specifications, software packages and patch management, RPKI operation, and deployments. Based on our analysis, we show that RPKI lacks maturity in its specification, in resilience against attacks and misconfigurations, and in its current operational mode, but is mature with regard to its operational feasibility.</p>
</section>
</section>
<section id="sec11" class="sec">
<h2 class="heading">Maturity of RPKI Specification</h2>
<p id="p-17">The Internet Engineering Taskforce (IETF) defined about 40 RPKI-related RFCs, which are generally complementary, each addressing a specific aspect of the RPKI ecosystem. Although the multiple standards aim to address missing or vague details and guide developers in implementing RPKI, their large number and often conflicting requirements increase complexity and hence the risk of bugs and vulnerabilities. Next, we explain potential problems arising from conflicting requirements or under-specification and provide concrete examples from our own evaluations and Internet measurements.</p>
<section id="sec12" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>Conflicting requirements.</strong>  Some networks adopt updated RFCs while others stick to older versions, resulting in inconsistent operational choices among these networks. These discrepancies in operational choices can lead to routing instabilities, rejected routes, and even security vulnerabilities. In our analysis, we found many conflicting requirements in the validation rules, particularly those around ROA validity, certificate expiration, and error handling. The conflicting validation rules create small windows where a network could unintentionally accept hijacked prefixes. By adhering to older or more permissive requirements or being too flexible in handling invalid or expired objects, networks might allow unauthorized route announcements to propagate, exposing them to prefix hijacks.</p>
<section id="sec13" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><em>Discrepancies in filtering logic. </em> As an example, RFC6811 introduces strict origin validation with ROAs. RFC8416 introduces an additional mechanism, Simple Linux Utility for Resource Management (SLURM), to overwrite RPKI validity for certain routes. Using SLURM, a network operator may fix known issues with invalid routes, for example, manually overwriting the invalidity of multi-homed announcements only covered by a single ROA. Such cases are still prevalent today.<a class="reference-link xref xref-bibr" href="#B33" data-jats-ref-type="bibr" data-jats-rid="B33"><sup>33</sup></a> Systems not using SLURM to fix these problems will conclude different validity statuses of routes compared to systems with manually configured rules, leading to inconsistencies in the global routing table and potential route-hijacking risks for networks that do not manually configure the required SLURM rules to fix such issues with missing ROAs. Vice-versa, the ability to configure arbitrary SLURM rules circumventing RPKI data also increases the potential for misconfiguration, thus opening systems up to hijacks.</p>
</section>
<section id="sec14" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><em>Discrepancies over which routes to filter.</em>  A further example of discrepancies in validation is inherent in RFC6811, which is clarified by update RFC8481. The phrasing on which received routes to validate sets a strong focus on BGP Update messages, leaving open whether other related messages, such as iBGP, should be validated. It also leaves ambiguity over how to validate other messages, such as which AS number to use for validation. RFC8481 clarifies potential misunderstandings, stating <i>all</i> routes should be validated and which ASN to use for validation in which cases. While operators who read the clarification will apply filtering as intended by the initial RFC, the large amount of published RPKI RFCs makes it likely that many operators do not keep track of newly published RFCs that update existing recommendations; the diverse interpretations of the initial RFC will thus lead to inconsistent routing behavior between systems.</p>
</section>
</section>
<section id="sec15" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>Vague and underspecified requirements.</strong>  Certain vague or underspecified requirements within RPKI RFCs can lead to inconsistent implementations and operational uncertainty. The lack of clear guidance in areas such as certificate handling, ROA validation, error handling, and manifest management creates potential windows of vulnerability, where different networks might adopt conflicting behaviors. This can increase the risk of security issues such as prefix hijacks or operational disruptions.</p>
<section id="sec16" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><em>Discrepancies in processing certificates.</em>  One example of vague specification is the handling of optional unknown field types within certificates. While the RPKI certificate template mandates the existence (or nonexistence) of certain fields—for example, mandating that the <code class="monospace">signer identifier</code> is present—it leaves open how validation should handle unknown optional fields. The X.509 template, however, allows for a range of additional optional fields not specified in the RPKI standard, and systems may decide to use these fields for specific use cases. This problem already manifests in real-world RPKI operation. In previous research,<a class="reference-link xref xref-bibr" href="#B29" data-jats-ref-type="bibr" data-jats-rid="B29"><sup>29</sup></a> we identified that one of the four major relying parties, Fort, rejects objects with unknown optional fields in certificates, while all others ignore the field and still validate the object. We found that Amazon used an optional name field within their certificates, not specified within the RPKI standard. While all other relying parties validated the object, Fort rejected all such Amazon objects, leaving more than 6,000 prefixes by Amazon vulnerable to hijack in any system using Fort.<a class="footnote-link xref xref-fn" href="#FN1" data-jats-rid="FN1" data-jats-ref-type="fn"><sup>a</sup></a> After raising the issue with developers, there was disagreement between different implementations over how to interpret the lack of specification in the standard, and the RPKI client even adapted their implementation to follow the interpretation by Fort.</p>
</section>
<section id="sec17" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><em>Discrepancies in key rollover.</em>  A further example is guidelines for key rollover in RFC6489, which leave operational details open to interpretation. Specifically, they do not fully address how frequently key rollovers should occur or how long old keys should remain valid. In practice, this leads to implementations using different key rollover timelines, which can result in operational inconsistencies. A lack of detailed guidance on timing creates the risk that some network operators may perform key rollovers too quickly, leading to temporary route-validation issues, while others may retain old keys for too long, increasing the vulnerability window for compromised keys. Another example is RFC6486, which covers manifests (used to ensure repository consistency) but does not clearly define how often manifests should be updated or what happens if a manifest becomes temporarily unavailable. It also lacks precise guidance on how to handle missing objects in the manifest. If a manifest is unavailable for a short time or if certain objects listed in the manifest are missing from the repository, RFC6486 leaves it unclear whether operators should reject all repository data, trust older manifest versions, or temporarily proceed without validation. This ambiguity could lead to inconsistent behaviors across networks, with some rejecting valid routes due to temporary manifest issues while others continue to operate without validation.</p>
</section>
<section id="sec18" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><em>Discrepancies in thresholds.</em>  Another example is the requirements to impose thresholds on RP functionalities. While standards need to offer some flexibility of implementation to allow for competition among products, it is also important to avoid standards that allow implementations that defeat the purpose of the standard. The issue of thresholds and limits on computation of RPs was not clearly specified. This vagueness was shown to expose RPs to stalling attacks by malicious repositories.<a class="reference-link xref xref-bibr" href="#B19" data-jats-ref-type="bibr" data-jats-rid="B19"><sup>19</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B20" data-jats-ref-type="bibr" data-jats-rid="B20"><sup>20</sup></a> In the course of an attack, an adversary can hang RPs for significant amounts of time, thereby stalling validation long enough for RPKI protection to be downgraded in BGP routers. While these problems have been addressed by adding thresholds, research shows that thresholds<a class="reference-link xref xref-bibr" href="#B20" data-jats-ref-type="bibr" data-jats-rid="B20"><sup>20</sup></a> not only do not eliminate the attacks against RPKI validation, but worse: They introduce failures under benign network conditions. Current thresholds are too permissive to prevent attacks and much stricter thresholds are required to enhance resilience against attacks. However, even permissive thresholds may cause the RPs to fail when fetching RPKI objects. A measurement we carried out in 2023 indicates that about 11.78% of attempts to fetch RPKI objects fail due to thresholds kicking in and dropping the connection.<a class="reference-link xref xref-bibr" href="#B20" data-jats-ref-type="bibr" data-jats-rid="B20"><sup>20</sup></a></p>
<p id="p-25">Source-code analysis reveals that the implemented threshold values differ significantly across RP implementations. For example, the connection timeout parameter for different implementations can range from 60s to 2.9h. These large discrepancies indicate that developers had to rely on their intuition, due to a lack of official guidance for selection of threshold values. Indeed, we have not found an analysis of the effect of different threshold values, nor have we found guidance on their selection. This lack of specific requirements in the standards for thresholds and lack of analysis on the ramifications of thresholds on RPKI validation under diverse network conditions cause developers to select arbitrary values in their implementations.</p>
</section>
</section>
</section>
<section id="sec19" class="sec">
<h2 class="heading">Maturity of RPKI Software Packages</h2>
<p id="p-26">Since the introduction of RPKI, implementations of RPKI software have emerged as open source projects, developed and managed by small groups of developers. Looking at the contribution history of the respective projects, the amount of active contributors (those who made contributions in the past 12 months) ranges from one to five people. All projects are open source and accept pull requests, with one implementation specifically encouraging contributions. While open source software is beneficial to improve the security, public participation, and trustworthiness of software, the XZ-Utils attack<a class="reference-link xref xref-bibr" href="#B34" data-jats-ref-type="bibr" data-jats-rid="B34"><sup>34</sup></a> has illustrated that especially projects managed by small developer teams may be subject to intentional introduction of malicious code from a sophisticated attacker. Three implementations are developed by known organizations located in North America, South America, and Europe, indicating more available resources for security management and the vetting of code contributions.</p>
<section id="sec20" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>Software bugs.</strong>  While the quality of implementations has continuously improved over recent years, implementations repeatedly exhibit large amounts of vulnerabilities, with some problems persisting to this day. The RPKI specification, software packages, and repository implementations are still not sufficiently stable and contain critical vulnerabilities, and RPKI developers lack tools to test their software to identify bugs and vulnerabilities.<a class="reference-link xref xref-bibr" href="#B6" data-jats-ref-type="bibr" data-jats-rid="B6"><sup>6</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B19" data-jats-ref-type="bibr" data-jats-rid="B19"><sup>19</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B20" data-jats-ref-type="bibr" data-jats-rid="B20"><sup>20</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B24" data-jats-ref-type="bibr" data-jats-rid="B24"><sup>24</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B29" data-jats-ref-type="bibr" data-jats-rid="B29"><sup>29</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B31" data-jats-ref-type="bibr" data-jats-rid="B31"><sup>31</sup></a> Although vulnerabilities were discovered in all RPKI components, we found the largest amount in RP validator implementations, the middleware responsible for downloading, parsing, and validating RPKI objects. Overall, at least 53 vulnerabilities were disclosed, including persistent DoS, authentication bypass, cache poisoning, and remote code execution. While most vulnerabilities were swiftly fixed, they still raise questions about the resilience of implementations and potential existence of other zero-days.</p>
<p id="p-28">Previous work has indicated multiple reasons for the persistence of vulnerabilities and problems in RPKI implementations, namely a lack of test-tooling, complexity of the cryptographic architecture, and vagueness of RFC requirements. Further contributing to the problem, two of the three largest implementations of RPKI relying parties are written in C, a non-memory-safe language that is more difficult to protect against memory-related exploitation.</p>
<p id="p-29">While we expect that software security will increase in the future with improved secure coding patterns and the increasing availability of tooling, the current state of software security in RPKI makes it attractive for attackers, with a relative abundance of vulnerabilities that have potentially devastating consequences for RPKI validation, and might even open a backdoor into the local network running the vulnerable software component.</p>
</section>
<section id="sec21" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>Intentional backdoors.</strong>  As RPKI gains traction, the risk of intentional backdoors will also grow. Following the exposure of the XZ-Utils attack,<a class="reference-link xref xref-bibr" href="#B34" data-jats-ref-type="bibr" data-jats-rid="B34"><sup>34</sup></a> the security community has paid increasing attention to the issue of planted backdoors in open source software. Since all popular RPKI software implementations are open source and accept code contributions by the community, the threat of intentional backdoors is substantial. Such backdoors might, for example, include compromises to validation integrity to get malicious data validated, eavesdropping on local networks, or even manipulated program flow to enable remote code execution (RCE).</p>
<p id="p-31">Contributions to open source projects must be vetted by the developers, and the community should invest resources in tracking changes to repositories and investigating suspicious changes to the software.</p>
</section>
</section>
<section id="sec22" class="sec">
<h2 class="heading">Maturity of RPKI Deployments</h2>
<p id="p-32">The challenges in deploying RPKI revolve around errors with registering network resources in RPKI and lack of experience with strict RPKI validation in production environments.</p>
<section id="sec23" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>Errors in ROAs.</strong>  The deployment of RPKI involves creating ROAs for authenticating the associated prefixes. The objects should be stored in public RPKI repositories. Managing ROAs and maintaining cryptographic keys introduces complexity. Research has shown that complexity often leads to errors and misconfigurations in RPKI objects, which result in route filtering and loss of connectivity for legitimate routes.<a class="reference-link xref xref-bibr" href="#B15" data-jats-ref-type="bibr" data-jats-rid="B15"><sup>15</sup></a> Despite awareness of the issue of erroneous ROAs, and the fact that ROV-enforcing networks filter traffic from prefixes with conflicts in ROAs, erroneous ROAs still exist.<a class="footnote-link xref xref-fn" href="#FN2" data-jats-rid="FN2" data-jats-ref-type="fn"><sup>b</sup></a> Although there are proposals to automate prefix registration,<a class="reference-link xref xref-bibr" href="#B17" data-jats-ref-type="bibr" data-jats-rid="B17"><sup>17</sup></a> no mechanisms are used in practice.</p>
</section>
<section id="sec24" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>Fail-open ROV deployments.</strong>  Networks set up and use RPs to fetch and validate ROAs and other RPKI material. The validated objects inform the routers in their routing decisions, in a process called route origin validation (ROV).</p>
<section id="sec25" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><em>Fail-open mode.</em>  Enforcement of RPKI validity with ROV is currently performed in a non-strict “fail open” mode, as is recommended in RFC7115. According to these permissive policies, the operators should use RPKI to prefer valid announcements, assign a lower preference to &#8220;Not Found&#8221; announcements, and either discard &#8220;Invalid&#8221; announcements or give them a very low preference. If ROAs cannot be fetched—for example, because they do not exist, are invalid, or their repository is unreachable—RPKI validation for those resources gets the status “Not Found” and announcements are hence accepted. Operating RPKI validation in fail-open mode, not invalidating unavailable resources, is critical for Internet stability, as not all address space is covered by ROAs yet, and adversaries may be able to prevent RPs from fetching RPKI objects.<a class="reference-link xref xref-bibr" href="#B19" data-jats-ref-type="bibr" data-jats-rid="B19"><sup>19</sup></a></p>
<p id="p-36">Failures in accessing RPKI repositories may occur even under benign network conditions.<a class="reference-link xref xref-bibr" href="#B20" data-jats-ref-type="bibr" data-jats-rid="B20"><sup>20</sup></a> Our measurements show that about 6,000 validated ROA payloads (VRPs) in RPKI are unavailable due to operational issues, for example, in their repository, or due to configuration issues, such as being expired. If strict validation is applied, prefixes lacking covering RPKI objects will be filtered, impairing reachability to those ASes even if the unavailability is caused by benign errors, not attacks. The risk of losing legitimate traffic is a major concern for network operators and one of the main obstacles hindering wide adoption of RPKI validation. Fail-open mode is meant to facilitate incremental RPKI deployment, reducing failures and traffic loss. The downside of fail-open is that networks deploying RPKI filtering with ROV may still be vulnerable to routing hijacks, even of ROA-covered prefixes, if adversaries can disable RPKI validation, for example, by blocking access to RPKI repositories or preventing RPs from fetching RPKI objects.<a class="reference-link xref xref-bibr" href="#B19" data-jats-ref-type="bibr" data-jats-rid="B19"><sup>19</sup></a> If strict validation were performed, unavailable ROAs would invalidate BGP announcements, preventing hijacks, but causing traffic loss. In case of attack, this is desirable, as hijacks can have more detrimental impacts than DoS. The transition from fail-open to strict is a trade-off between resilience against (benign) RPKI unavailability causing traffic loss, and protection against BGP hijacks.</p>
<p id="p-37">Prerequisites and considerations for strict validation thus go beyond the discussion in RFC7715, which only covers full ROA deployment. Strict validation is beneficial only if RPKI data availability is reliable and misconfigurations causing data unavailability are rare. Otherwise, systems using RPKI experience regular DoS, even in the absence of attacks. Achieving this high level of availability is feasible, as such availability requirements already exist for other technologies: Internet users rely on full availability of their DNS recursive resolver, usually hosted by their ISP, to use Internet services, and TLS implementations will not establish a server connection if its certificate is unavailable. Reaching comparable availability would allow RPKI to shift toward strict validations to increase RPKI security guarantees. Improving availability must thus be a core concern in RPKI deployments and standardization efforts. In addition, there should be a plan for transition to operating RPKI in production environments in strict validation mode, to identify any possible challenges and problems. Only after the community collects sufficient experience operating full-fledged RPKI in strict validation mode can RPKI undergo the shift to strict validation.</p>
</section>
<section id="sec26" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><em>Experimental setup.</em>  In addition to using fail-open mode for &#8220;Not Found,&#8221; recommendations generally suggest starting with an experimental setup for all RPKI filtering. The FCC recommends that networks deploy RPKI filtering in a staged process, starting with an experimental setup that does not impact their routing by using fail-open even for “Invalid.” MANRS also recommends that  networks first install RPKI with logging, and not actually filter routes.</p>
<p id="p-39">This is a sensible recommendation, considering enforcing ROV can compromise BGP availability if ROAs or RPs are faulty, and activating ROV requires reconfiguration of all BGP routers. The challenges of moving from an experimental to a production-enforcing state for ROV are also evident in the discrepancy between the number of systems issuing ROAs and the number enforcing ROV: Though more than 50% of BGP prefixes are covered with ROAs, only about 25% of systems enforce ROV.<a class="reference-link xref xref-bibr" href="#B28" data-jats-ref-type="bibr" data-jats-rid="B28"><sup>28</sup></a> Many adopters remain in an experimental full fail-open stage without enforcement and do not gain experience with full-fledged production RPKI management.</p>
<p id="p-40">It is thus vital that more experience in RPKI operation, including the impact of strict validation, is published, and more research into the operational considerations of RPKI is conducted to provide a more solid understanding of RPKI deployments and guide operators. Such research will inform the standardization, as well as developers and operators, toward a mature and ready-to-implement RPKI that can be deployed faster and more securely.</p>
</section>
</section>
<section id="sec27" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>Challenges to full deployment.</strong>  Despite the continuous growth of RPKI, the scenario of full global RPKI deployment is not yet well understood, and simulations for RPKI scalability in different deployment scenarios have not yet been sufficiently explored. This is important, since research shows multiple indications for potential problems in full deployment.<a class="reference-link xref xref-bibr" href="#B20" data-jats-ref-type="bibr" data-jats-rid="B20"><sup>20</sup></a> The growth in the number of RPKI objects and the standardization of additional objects will significantly increase download and validation times for RPs, increasing update intervals and extending the delay between object changes and updates to the routing behavior of systems. This will also decrease the agility of systems to quickly react to problems and routing attacks. Further, with increasing deployment in smaller systems, the number of RPs will increase, adding substantial additional load to RPKI repositories. With more operational experience in RPKI, and under pressure from attacks on the architecture, more systems might also decide to operate their own RPKI repository, significantly increasing the number of repositories.</p>
</section>
<section id="sec28" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>Challenges with patch management.</strong>  Many operators struggle with keeping RPKI software updated due to the lack of automated patching mechanisms. Getting the most up-to-date RP software requires administrators to manually get and install the new version. We found that a significant percentage of RPKI installations run outdated versions with known vulnerabilities, increasing the risk of attacks. To measure how many RPs were not patched, we set up our own RPKI publication point and logged version headers of global RPs that contacted our repository to fetch RPKI objects. We checked how many clients use software with at least one publicly known severe vulnerability (CVSS &gt;= 7.0), only including vulnerabilities disclosed more than one year before our measurements. We found that 41.2% of global RPs are vulnerable to at least one long-disclosed attack. This number is alarming, as it shows that even when vulnerabilities are discovered and patched, many RPs can still be attacked long after patch issuance. This lack of timely patching also illustrates that management of RPs is not prioritized.</p>
</section>
<section id="sec29" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>Challenges with new technologies.</strong>  The most widely deployed use case of RPKI today is ROAs/ROV, with 50% of announced prefixes ROA-covered and roughly 25% of ASes validating with ROV. RPKI was created to support many more use cases, but no other is widely deployed to date. BGPsec was never deployed due to computational load and missing protections in partial deployment, and only two BGPsec certificates are available in RPKI today. Route-leak protection through ASPAs was proposed in 2018, but the RFC remains in a draft stage, currently in iteration 22, and only 83 ASPAs are available through RPKI. Other uses of RPKI are in early phases of standardization. While RPKI was designed for broad protections of BGP, the slow standardization process has led to only ROAs and ROV reaching wide-scale deployment. It is vital that specification development advances to allow RPKI protection to cover all BGP insecurities, including path hijacks and route leaks.</p>
</section>
<section id="sec30" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>RPKI operation.</strong>  Automation tools for RPKI configuration and management are still being developed. Many operators rely on manual processes, which can lead to misconfigurations. But operational challenges also arise after RPKI components have been configured and set up. A recent study<a class="reference-link xref xref-bibr" href="#B30" data-jats-ref-type="bibr" data-jats-rid="B30"><sup>30</sup></a> discovered a large and persistent amount of networking and configuration errors in RPKI repositories, including unreachability, slow connections, and failing DNSSEC validation, frequently leading to unavailable objects and slowing the fetching of RPKI objects for all globally running RPs. Another study showed that inconsistency of RPKI validation poses challenges to RPKI deployments.<a class="reference-link xref xref-bibr" href="#B29" data-jats-ref-type="bibr" data-jats-rid="B29"><sup>29</sup></a> With RPKI using established cryptographic algorithms, it would be expected that different implementations reach consistent validation results on identical objects. This is not the case in practice: In follow-up research, we identified 25 inconsistencies that cause discrepancies in RPKI validation results across different RPs. Multiple inconsistencies that we discovered lead to differing validation results of real-world RPKI objects, therefore impacting production routing.<a class="reference-link xref xref-bibr" href="#B12" data-jats-ref-type="bibr" data-jats-rid="B12"><sup>12</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B13" data-jats-ref-type="bibr" data-jats-rid="B13"><sup>13</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B29" data-jats-ref-type="bibr" data-jats-rid="B29"><sup>29</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B37" data-jats-ref-type="bibr" data-jats-rid="B37"><sup>37</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B42" data-jats-ref-type="bibr" data-jats-rid="B42"><sup>42</sup></a> For instance, some implementations do not accept repositories with unknown objects, such as ASPAs, or do not accept snapshots where the same object was added twice. Some implementations do not accept objects with identical SKIs, which can occur in cases of improper key management. This inconsistency can cause legitimate routes to be incorrectly marked as invalid, leading to instability in routing across different networks.</p>
</section>
</section>
<section id="sec31" class="sec">
<h2 class="heading">Emerging RPKI Attack Surface</h2>
<p id="p-45">The expanding adoption of RPKI in global networks also increases its attractiveness as a target for attacks. Therefore, it is important to extend the White House roadmap to also consider attacks against RPKI, and attacks through RPKI against its adopters. Such RPKI attacks may be closely directed toward RPKI functionality, for example, downgrading or circumventing hijack protections in victim systems, but can also target RPKI components as entry points into the network where vulnerable RPKI software is located. Enumeration of potential targets and development of directed attacks is straightforward, as the IP addresses of RPs and RPKI repositories are known, and they continuously communicate with other hosts in the Internet. Using enumerated targets, adversaries can attack the RP instances directly by setting up their own RPKI repository, or indirectly by attacking other RPKI repositories. Attacks often use specially crafted RPKI objects that disable or manipulate RPKI validation or, in the worst case, allow the attacker to obtain access to the RPKI-adopting network. What sets RPKI apart from other technologies is that all global RPs regularly contact all globally available repositories, allowing one server to attack all running instances directly. Widespread attacks are thus straightforward to execute, even for low-resource attackers. Further, due to the fail-open validation mode of RPKI, even DoS attacks—often considered low-impact vulnerabilities—have high impact in RPKI since they downgrade protection, leaving prefixes open for hijack. Despite the high impact of DoS in RPKI and the straightforward potential for widespread exploitation, previous work found an abundance of DoS and other vulnerabilities in RPKI, compromising its resilience and security. In this section, we discuss the RPKI-specific and general attack surface of networks.</p>
<section id="sec32" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>Downgrade of RPKI validation.</strong>  RPKI validation of BGP messages requires availability of RPKI data. We found that attacks on availability are the most prominent threat in RPKI, with 53 such vulnerabilities published in the scientific literature within the past three years.<a class="reference-link xref xref-bibr" href="#B18" data-jats-ref-type="bibr" data-jats-rid="B18"><sup>18</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B19" data-jats-ref-type="bibr" data-jats-rid="B19"><sup>19</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B20" data-jats-ref-type="bibr" data-jats-rid="B20"><sup>20</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B24" data-jats-ref-type="bibr" data-jats-rid="B24"><sup>24</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B29" data-jats-ref-type="bibr" data-jats-rid="B29"><sup>29</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B31" data-jats-ref-type="bibr" data-jats-rid="B31"><sup>31</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B41" data-jats-ref-type="bibr" data-jats-rid="B41"><sup>41</sup></a> Attacks on availability may target all parts of the RPKI architecture, with attacks ranging from rate-limiting RPKI repositories or creating complex repository structures to stall RPs, to crashing RPs through malformed objects or protocol headers. We summarize the known vulnerabilities against RPKI components in the <a href="#T1">table</a>. Forty-two vulnerabilities have already been disclosed that could be exploited to crash RPs during processing, one crashing through RRDP and three that compromised availability through stalling. To prevent RPs from accessing RPKI data, adversaries can launch DoS attacks against repositories, exploiting vulnerabilities in repositories&#8217; software<a class="reference-link xref xref-bibr" href="#B6" data-jats-ref-type="bibr" data-jats-rid="B6"><sup>6</sup></a> and causing them to crash, or flooding them with requests.<a class="reference-link xref xref-bibr" href="#B18" data-jats-ref-type="bibr" data-jats-rid="B18"><sup>18</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B19" data-jats-ref-type="bibr" data-jats-rid="B19"><sup>19</sup></a></p>
<figure id="T1" class="table-wrap" data-jats-position="float">
<div class="caption"><span class="caption-label">Table. </span><span class="p">Published vulnerabilities in RPKI components.</span></div>
<div class="table-container">
<table class="table table-bordered table-condensed table-hover" data-jats-frame="hsides" data-jats-rules="rows">
<colgroup>
<col align="left" valign="top" />
<col align="left" valign="top" />
<col align="left" valign="top" />
<col align="left" valign="top" />
<col align="left" valign="top" />
<col align="left" valign="top" />
<col align="left" valign="top" /> </colgroup>
<thead style="vertical-align: bottom;">
<tr>
<th style="text-align: left;">Vulnerability</th>
<th style="text-align: left;">Amount</th>
<th style="text-align: left;">Target</th>
<th style="text-align: left;">Availability</th>
<th style="text-align: left;">RCE</th>
<th style="text-align: left;" data-jats-scope="col">Validation Integrity</th>
<th style="text-align: left;">CVSS</th>
</tr>
</thead>
<tbody>
<tr>
<td style="text-align: left;">Crash</td>
<td style="text-align: left;">42</td>
<td style="text-align: left;">RP</td>
<td style="text-align: left;">✓</td>
<td style="text-align: left;">—</td>
<td style="text-align: left;">—</td>
<td style="text-align: left;">7.5</td>
</tr>
<tr>
<td style="text-align: left;">Stalling</td>
<td style="text-align: left;">3</td>
<td style="text-align: left;">PP</td>
<td style="text-align: left;">✓</td>
<td style="text-align: left;">—</td>
<td style="text-align: left;">—</td>
<td style="text-align: left;">N.A.</td>
</tr>
<tr>
<td style="text-align: left;">Kill-switch</td>
<td style="text-align: left;">1</td>
<td style="text-align: left;">RRDP</td>
<td style="text-align: left;">✓</td>
<td style="text-align: left;">—</td>
<td style="text-align: left;">—</td>
<td style="text-align: left;">7.5</td>
</tr>
<tr>
<td style="text-align: left;">PDU</td>
<td style="text-align: left;">1</td>
<td style="text-align: left;">RTR</td>
<td style="text-align: left;">✓</td>
<td style="text-align: left;">—</td>
<td style="text-align: left;">—</td>
<td style="text-align: left;">N.A.</td>
</tr>
<tr>
<td style="text-align: left;">Info-Leak</td>
<td style="text-align: left;">1</td>
<td style="text-align: left;">PP</td>
<td style="text-align: left;">—</td>
<td style="text-align: left;">—</td>
<td style="text-align: left;">—</td>
<td style="text-align: left;">N.A.</td>
</tr>
<tr>
<td style="text-align: left;">Delta-Snapshot</td>
<td style="text-align: left;">1</td>
<td style="text-align: left;">PP</td>
<td style="text-align: left;">—</td>
<td style="text-align: left;">—</td>
<td style="text-align: left;">—</td>
<td style="text-align: left;">N.A.</td>
</tr>
<tr>
<td style="text-align: left;">Cache-Poisoning</td>
<td style="text-align: left;">1</td>
<td style="text-align: left;">RP</td>
<td style="text-align: left;">✓</td>
<td style="text-align: left;">—</td>
<td style="text-align: left;">✓</td>
<td style="text-align: left;">N.A.</td>
</tr>
<tr>
<td style="text-align: left;">Path-Traversal</td>
<td style="text-align: left;">2</td>
<td style="text-align: left;">RP</td>
<td style="text-align: left;">✓</td>
<td style="text-align: left;">✓</td>
<td style="text-align: left;">✓</td>
<td style="text-align: left;">9.3</td>
</tr>
<tr>
<td style="text-align: left;">Buffer-Overflow</td>
<td style="text-align: left;">1</td>
<td style="text-align: left;">RP</td>
<td style="text-align: left;">✓</td>
<td style="text-align: left;">✓</td>
<td style="text-align: left;">✓</td>
<td style="text-align: left;">9.8</td>
</tr>
</tbody>
</table>
</div>
</figure>
<section id="sec33" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><em>Effect. </em> If a border router cannot fetch fresh RPKI data from an RP, it will not use RPKI for making routing decisions in BGP; that is, its RPKI protection will be downgraded to insecure BGP. Default routers cache-flush times range from 360s (Cisco<a class="reference-link xref xref-bibr" href="#B9" data-jats-ref-type="bibr" data-jats-rid="B9"><sup>9</sup></a>) to 7,200s (FRR<a class="reference-link xref xref-bibr" href="#B14" data-jats-ref-type="bibr" data-jats-rid="B14"><sup>14</sup></a>) if an RP is unavailable. Further, if an RP cannot fetch objects due to stalling, cached objects will eventually expire, generally within 24 hours. While the handling of stale objects depends on local policy, most implementations will eventually discard stale objects, and protection will thus be downgraded. As a result, when the networks cannot apply RPKI validation, they are exposed to routing attacks on supposedly RPKI-protected systems. With increasing deployment of RPKI, including networks that have been historically attractive for cyberattacks like DNS servers or ASes hosting cryptocurrency services, attacks on availability become more attractive.</p>
</section>
</section>
<section id="sec34" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>Manipulating RPKI validation integrity.</strong>  While attacks on the availability of RPs or repositories downgrade RPKI protection, attacks against integrity aim to manipulate RPKI validation results to circumvent protection. Multiple vulnerabilities have been disclosed in recent years that subvert validation integrity.</p>
<section id="sec35" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><em>Attacks against relying parties.</em>  Known attacks allow malicious objects to get validated, or existing objects invalidated (see <a href="#T1">table</a>). The former attack can be done by adding a malicious TAL via a path-traversal attack against RPs, allowing an attacker to introduce malicious RPKI objects.<a class="reference-link xref xref-bibr" href="#B29" data-jats-ref-type="bibr" data-jats-rid="B29"><sup>29</sup></a> The attacker can then attest ownership of arbitrary IP prefixes and thus conduct BGP attacks that seem like RPKI-valid announcements. The latter targets specific objects within RPKI, and was first demonstrated by Mirdita et al.<a class="reference-link xref xref-bibr" href="#B29" data-jats-ref-type="bibr" data-jats-rid="B29"><sup>29</sup></a></p>
</section>
<section id="sec36" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><em>Attacks against RPKI repositories.</em>  Repositories can be attacked, for example, over a management interface exposed to the Internet, allowing attackers to gain access to the management tooling when weak or default credentials are used.<a class="reference-link xref xref-bibr" href="#B10" data-jats-ref-type="bibr" data-jats-rid="B10"><sup>10</sup></a></p>
</section>
<section id="sec37" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><em>Effect.</em>  These attacks are very difficult to detect and are thus much more attractive for attackers than attacks on availability, though zero-days are likely much harder to find, with only two such vulnerabilities published to date.</p>
</section>
</section>
<section id="sec38" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>RPKI as a foothold into the network.</strong>  The attack surface of networks deploying RPKI is not limited to attacks on RPKI functionality. RPKI components, particularly RPs, can expose networks to a wider range of attacks. It is recommended to set up RPs in proximity to border routers, ideally on the same network (BCP-185, RFC7115). Consequently, any intentional backdoors or erroneous bugs in RPs can be exploited by adversaries to attack the border routers and even penetrate the network.</p>
<section id="sec39" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><em>RCE in Fort.</em>  We explain this risk with an example of a remote code execution (RCE) against RPKI, which we discovered in our research.<a class="reference-link xref xref-bibr" href="#B24" data-jats-ref-type="bibr" data-jats-rid="B24"><sup>24</sup></a> Concretely, we discovered a buffer-overflow vulnerability in the validation pipeline of the RP Fort. The vulnerability stems from a bug in the processing of the key-usage extension of an X.509 certificate, contained in most RPKI objects. The implementation allocates a 2 byte buffer, but lacks a length check for the attacker-controlled value, allowing an attacker to overflow the stack and write arbitrary bytes into memory behind the allocated buffer. The attack can be exploited by setting up a production RPKI repository, uploading an object with a manipulated key-usage extension, and serving it to any requesting RP. Through the vulnerability, the attacker can manipulate Fort&#8217;s control flow, thereby achieving RCE on all clients running it. While protections of modern compilers and operating systems, such as address space layer randomization (ASLR) and stack canaries, can mitigate direct exploitation, existence of additional bugs that, for example, leak the canary, can enable exploitation, even with protections.<a class="reference-link xref xref-bibr" href="#B24" data-jats-ref-type="bibr" data-jats-rid="B24"><sup>24</sup></a> The severity of the bug raises questions about its origin, and we found potential indications for both intentional introduction and a benign coding error. The code not implementing any functionality, its convenient remote accessibility for stealthy exploitation, and the easy activation of the vulnerability through compiler flags could indicate malicious introduction of the bug. However, we did not find any operational indications of malicious planting. The code was committed by the long-term main developer of Fort, and similar code sections exist in other parts of the RP that implement actual functionality, making a copy-paste error likely. While the bug thus likely constitutes a benign coding error, it illustrates the ease with which such a backdoor could be inserted, and how the community needs to be aware of the increasing attack surface—in particular, the risk of backdoors that RPKI introduces.</p>
</section>
<section id="sec40" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><em>Effect.</em>  RCE attacks<a class="reference-link xref xref-bibr" href="#B24" data-jats-ref-type="bibr" data-jats-rid="B24"><sup>24</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B29" data-jats-ref-type="bibr" data-jats-rid="B29"><sup>29</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B41" data-jats-ref-type="bibr" data-jats-rid="B41"><sup>41</sup></a> have a very severe impact, as they open the victim up to a wide range of follow-up exploitation. For example, an attacker can establish a reverse shell to the victim RP or repository, gain initial access, and both infiltrate the system itself and use it as a backdoor into the local network to attack other servers, such as BGP routers. Further, the attacker might find sensitive information, such as router credentials, stored on the server. Gaining access to the RPKI component also compromises validation integrity, as the attacker can add arbitrary data to the unauthenticated RPKI payload list, sent to the routers without any additional validation.</p>
</section>
</section>
</section>
<section id="sec41" class="sec">
<h2 class="heading">Perfect Is the Enemy of the Good</h2>
<p id="p-56">Until recently, few people outside the Internet operational, engineering, and research communities were aware of RPKI. That changed in September 2024, when the White House identified RPKI as the key component for securing Internet routing, pushing RPKI from niche to mainstream. It might be expected that mainstream technologies are fully mature—in particular, fully stable and secure. As a niche technology, RPKI developed organically in many small steps, each inching a bit closer to maturity. But, as the previous sections have shown, RPKI is far from full maturity. What does this actually mean in practice? Did the White House push for the adoption of an immature technology, potentially doing more harm than good? Or did they promote the best available, &#8220;good enough&#8221; technology, motivating research and industry to speed up and put more resources behind improving RPKI?</p>
<section id="sec42" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>Academic analysis shows RPKI is not mature.</strong>  Academic research has repeatedly shown issues in RPKI that indicate it has not yet reached full maturity. RPKI implementations are not sufficiently stable and lack resilience to cyber attacks. RPKI validation exhibits inconsistent results. The RPKI standard specifications have not yet been finalized. And the developers and operators lack documentation and automated tools for development and configuration of RPKI technology. All of these indicate that RPKI is not mature.</p>
</section>
<section id="sec43" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>But, so what? Systems in the real world are never fully mature.</strong>  Arguably, demanding full maturity before large-scale deployment is a very academic expectation; in reality, there is nothing like full maturity and perfection, only more or less good enough. For example, the Internet, like most information and communication technologies, is not mature, from applications to Internet protocols and popular security mechanisms such as IPsec and SSL/TLS—all are vulnerable and contain bugs. Many Internet systems started from collaborative efforts between researchers and operators and grew organically. Over time, these efforts mature from experimental research prototypes and individual initiatives into deployments by large networks. The software is improved “on-the-fly” with periodic patches that close bugs or add new features. This organic systems maturity is not aligned with academic definitions and frameworks, but in practice, the systems are never 100% perfect and mature, rather evolving gradually over time. It is thus essential to include the practical operational benefits of technology into discussions of its maturity.</p>
</section>
<section id="sec44" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>Mature or not, BGP connects the Internet.</strong>  Examples of immature, but nevertheless heavily used, technologies exist in abundance; Internet routing with BGP is among the most prominent. Nowadays, BGP enables all Internet activities. As BGP took on a central role in all online activity, its complexity also grew. BGP was designed on three napkins to connect different Internet domains, and was not designed to be a robust and secure protocol that one would rely on for critical applications such as the Internet. Since then, BGP has evolved, in the number of steps of its decision process, the number of attributes, and the number of networks it supports—all of these got more complicated, more difficult to configure, and more vulnerable. Indeed, software bugs and issues in protocol specification are common in inter-domain routing with BGP, and may lead to outages, failures, and attacks. For instance, FRR routers crashed and disconnected large networks from the Internet because they could not parse standard-compliant BGP attributes in routing announcements.<a class="reference-link xref xref-bibr" href="#B7" data-jats-ref-type="bibr" data-jats-rid="B7"><sup>7</sup></a> In addition, the complexity of BGP may create a chain of side effects, such that a small failure or misconfiguration in one part of the Internet can have devastating global consequences. But despite all the problems, outages, and attacks, the &#8220;triple napkin&#8221; protocol continues to connect the Internet. Moreover, BGP applications evolved far beyond BGP’s original purpose, including many new and emerging applications, such as internal routing in datacenters, MPLS VPN across organizational sites, and load balancing.</p>
</section>
<section id="sec45" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>Our analysis should be used as a to-do list.</strong>  Our study found multiple issues in RPKI that show immaturity—an abundance of vulnerabilities that, considering RPKI’s design, could be exploited with high impact. Still, RPKI shows that a list of problems does not necessarily reflect the operational readiness of a system, and even deploying an immature technology can provide significant operational benefits. To date, no malicious attacks on RPKI infrastructure are known, and RPKI effectively protects the Internet from hijacks.<a class="reference-link xref xref-bibr" href="#B21" data-jats-ref-type="bibr" data-jats-rid="B21"><sup>21</sup></a> Academic analysis is still important, as it allows us to identify important directions toward improving the security and stability of systems, but <i>the implementation of academic analysis needs to be adapted to how the systems evolve and mature in the real world</i>. An academic analysis provides a <i>to-do </i>list to guide adopters, operators, and developers in prioritizing their actions, addressing problems one at a time to improve the maturity of the operational system.</p>
</section>
<section id="sec46" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>The roadmap is a huge leap forward.</strong>  The White House’s 2024 roadmap’s recognition of RPKI as a critical security measure is an important step forward. Until recently, RPKI was mostly experimental, but the cybersecurity strategy of the White House, the FCC NPRM, and the recent roadmap, make a huge leap forward toward securing the routing infrastructure with RPKI. This roadmap will push the adoption of RPKI forward, thereby increasing global routing security. Now it is important to identify the hurdles that need to be resolved toward this goal. We outlined a number of challenges that need to be addressed in this article, and initial work on improving the core RPKI design for better security and efficiency has already been published.<a class="reference-link xref xref-bibr" href="#B36" data-jats-ref-type="bibr" data-jats-rid="B36"><sup>36</sup></a></p>
</section>
</section>
<section id="sec47" class="sec">
<h2 class="heading">Conclusion</h2>
<p id="p-62">RPKI implementations started as collaborative efforts between researchers, operators, and the broader IETF community. Over time, these efforts matured from experimental research projects and individual operator initiatives into deployments by some of the largest networks in the world. Our research shows that RPKI still suffers from problems and is not sufficiently stable, motivating more efforts by the community to increase its resilience, improve the specification, and mitigate vulnerabilities. Nevertheless, RPKI already delivers measurable benefits and is an essential part of ongoing efforts to improve routing security. Research shows that RPKI can substantially limit the propagation of invalid BGP announcements, hence mitigating traffic hijacks.<a class="reference-link xref xref-bibr" href="#B21" data-jats-ref-type="bibr" data-jats-rid="B21"><sup>21</sup></a> RPKI also provides an important prerequisite for prospective routing-security solutions, including path-validation and route-leak prevention. The White-House roadmap is a huge leap for RPKI, and therefore also for Internet routing, to truly mature and meet the expectations of security, reliability, and scalability for production-level deployments across the global Internet.</p>
</section>
<section id="sec48" class="sec">
<h2 class="heading">Acknowledgments</h2>
<p id="p-63">This work has been co-funded by the German Federal Ministry of Education and Research and the Hessen State Ministry for Higher Education, Research and Arts within their joint support of the National Research Center for Applied Cybersecurity ATHENE and by the Deutsche Forschungsgemeinschaft (DFG, German Research Foundation) SFB 1119.</p>
</section>
<section id="sec49" class="sec"></section>
</div>
</article>
]]></content:encoded>
					
					<wfw:commentRss>https://cacm.acm.org/research/are-we-actually-there-assessing-rpki-maturity/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
		      <dc:creator><![CDATA[Niklas Vogel]]></dc:creator>
      <dc:creator><![CDATA[Michael Waidner]]></dc:creator>
<post-id xmlns="com-wordpress:feed-additions:1">779184</post-id>	</item>
		<item>
		<title>The New Era of Sensory Augmentation</title>
		<link>https://cacm.acm.org/news/the-new-era-of-sensory-augmentation/</link>
					<comments>https://cacm.acm.org/news/the-new-era-of-sensory-augmentation/#respond</comments>
		
		<dc:creator><![CDATA[Logan Kugler]]></dc:creator>
		<pubDate>Fri, 17 Apr 2026 21:05:15 +0000</pubDate>
				<category><![CDATA[Architecture and Hardware]]></category>
		<category><![CDATA[Artificial Intelligence and Machine Learning]]></category>
		<category><![CDATA[Computing Applications]]></category>
		<category><![CDATA[HCI]]></category>
		<guid isPermaLink="false">https://cacm.acm.org/?post_type=digital-library&#038;p=776832</guid>

					<description><![CDATA[<p>The field of sensory augmentation is moving beyond simple repair and increasingly is focused on enhancing and extending human capabilities.</p>]]></description>
										<content:encoded><![CDATA[<article>
<div class="body" lang="en">
<section id="sec1" class="sec">
<p id="p-1">The boundaries between human biology and intelligent technology are beginning to blur. Nowhere is this convergence more evident than in the emerging field of sensory augmentation, where artificial intelligence (AI), neuroscience, and wearable computing are coming together to redefine our fundamental senses.</p>
<p id="p-2">For decades, the primary goal of this field was restoration, or developing technologies to help people regain lost sight or hearing. However, experts say that balance is now shifting. The field is moving beyond simple repair and increasingly is focused on augmentation: not just restoring, but actively enhancing and extending human capabilities.</p>
<p id="p-3">This shift from restoration to enhancement recently was highlighted in a remarkable study. Researchers at Fudan University in Shanghai created an artificial retina that not only restored sight to blind animals but also gave them “super vision,” or the extraordinary ability to see infrared light, a spectrum invisible to them naturally.</p>
<p id="p-4">This advance builds on decades of foundational work in sensory restoration. The most commercially successful and widespread example, the cochlear implant, already has restored functional hearing to hundreds of thousands of individuals by translating sound into electrical signals that are sent directly to the auditory nerve.</p>
<p id="p-5">In the visual domain, devices like the pioneering Argus II retinal prosthesis (or “bionic eye”) proved that surgically implanted electrodes could restore a basic, low-resolution sense of sight to people with specific forms of blindness, establishing the viability of the “invasive” path of sensory augmentation that researchers are still pursuing. (The Argus II was discontinued after it failed to generate enough revenue and the company behind it, Second Sight, ran into financial issues.)</p>
<p id="p-6">This acceleration is being driven by more than just implant technology. The true catalyst for the augmentation era is the processing power of artificial intelligence (AI).</p>
<p id="p-7">AI is becoming essential for interpreting complex, real-world data before it is delivered to a human user. This is already common in advanced hearing aids, which use deep neural networks to isolate a single voice in a crowded room. This same computational power also is making non-invasive “sensory substitution” devices, which might translate a visual scene into complex soundscapes or tactile patterns, more sophisticated and learnable than ever before.</p>
<figure id="attachment_780129" aria-describedby="caption-attachment-780129" class="wp-caption alignnone"><img loading="lazy" decoding="async" class="size-full wp-image-780129" src="https://cacm.acm.org/wp-content/uploads/2026/04/cacm69.5-p29.photo-cap.jpg" alt="torsion transducer and cellphone interface" width="1024" height="605" srcset="https://cacm.acm.org/wp-content/uploads/2026/04/cacm69.5-p29.photo-cap.jpg 1700w, https://cacm.acm.org/wp-content/uploads/2026/04/cacm69.5-p29.photo-cap.jpg?resize=300,177 300w, https://cacm.acm.org/wp-content/uploads/2026/04/cacm69.5-p29.photo-cap.jpg?resize=768,454 768w, https://cacm.acm.org/wp-content/uploads/2026/04/cacm69.5-p29.photo-cap.jpg?resize=1024,605 1024w, https://cacm.acm.org/wp-content/uploads/2026/04/cacm69.5-p29.photo-cap.jpg?resize=1536,908 1536w" sizes="auto, (max-width: 1024px) 100vw, 1024px" /><figcaption id="caption-attachment-780129" class="wp-caption-text">A blindfolded test subject navigates a room and locates a chair using only a haptic device,<br />in an experiment at Northwestern University.</figcaption></figure>
<p id="p-8">“While the motivation for restorative technologies remains high, we are entering an era increasingly focused on the augmentation (such as enhancement and extension) of human capabilities,” said Amber Maimon, a researcher specializing in sensory augmentation at Haifa University in Israel.</p>
</section>
<section id="sec2" class="sec">
<h2 class="heading">The Dual Frontiers of Sensory Augmentation</h2>
<p id="p-9">Sensory augmentation work is advancing along two major, parallel paths, according to Maimon.</p>
<p id="p-10">The first is the invasive approach, which involves neural prosthetics and other interventions that directly interface with or modify sensory pathways. This is the category that includes the bionic eye, which aims to restore vision by electrically stimulating the remaining neural pathways.</p>
<p id="p-11">The second is the non-invasive approach, such as sensory substitution systems that translate information from one sense to another. A key example is the EyeCane, a device that translates distance information into sound or vibration for the user.</p>
<p id="p-12">“Both paths have been explored for decades, yet recent progress in neuroscience, materials, and computation is transforming what is possible,” said Maimon. Invasive neurotechnologies, for instance, are becoming increasingly sophisticated.</p>
<p id="p-13">“Early visual prostheses relied on direct stimulation of retinal or cortical regions, producing only basic light or shape percepts,” she said. “Current approaches move toward higher-resolution, wireless, and biocompatible implants that combine advanced image processing with materials designed to adapt to neural tissue over time.”</p>
<p id="p-14">Maimon said dense, flexible electrode arrays now enable more stable and long-term integration with the brain, as well as “improvements in wireless data and power transmission” that enhance both safety and reliability.</p>
<p id="p-15">“At the same time,” she added, “emerging bidirectional brain-computer interfaces are forming closed-loop systems in which artificial sensors and neural circuits continuously interact,” bringing prosthetic vision closer to natural perception, or even extending its capabilities.</p>
<p id="p-16">Non-invasive systems are evolving in parallel. Instead of surgically interfacing with the nervous system, these methods rely on training and active exploration to enable users to build new perceptual mappings. Research has shown, Maimon said, that the brain can reorganize itself to interpret these inputs as perceptual rather than symbolic. As a result, this path can yield impressive results while avoiding surgical risks and allowing rapid iteration, all in a way that can be tailored to individual users.</p>
<p id="p-17">In fact, while high-tech implants capture the imagination, James Negen, a researcher at Liverpool John Moores University in the U.K., said the most exciting and effective technology in this space requires no electronics at all.</p>
<p id="p-18">“Human echolocation is the objectively correct answer,” Negen said. “It is the only sensory augmentation approach to have a sustained community of real-world everyday users.” Negen pointed out that echolocation, a form of sensory substitution where a person uses clicks and their echoes to perceive their environment, has been “refined for decades” by scientists like Lore Thaler and practitioners like Daniel Kish.</p>
<p id="p-19">Yet the most promising direction moving forward, said Maimon, is not pursuing one path over the other, but in developing hybrid systems that combine invasive and non-invasive approaches. By bringing together the precision and fidelity of invasive interfaces with the adaptability and accessibility of non-invasive ones, she said, new systems could “enable richer, more natural, and more adaptive forms of perception.”</p>
<p id="p-20">“For example, a visual implant could convey the overall spatial structure of a scene directly to the visual cortex, while a non-invasive sensory substitution device provides complementary information about texture, contrast, or fine details,” she explained.</p>
</section>
<section id="sec3" class="sec">
<h2 class="heading">The Perception Problem</h2>
<p id="p-21">Despite the rapid progress on both frontiers, the dream of a truly functional artificial eye remains distant. According to Negen, when it comes to that specific goal, “We are so far away that we don’t yet know what we don’t know.”</p>
<p id="p-22">“It is like asking Aristotle to lay out the path to the atom bomb,” he said. “We have only the most shallow understanding of basic questions like what physical properties give rise to conscious experience, what makes an experience perceptual, and how we come to feel ownership over a sensory stream.”</p>
<p id="p-23">This philosophical hurdle points to a core technical issue.</p>
<p id="p-24">“The greatest challenge is that perception is not simply electrical,” said Maimon. “It’s not enough to activate the right neurons or reproduce the correct signal patterns. The information must be organized in a way the brain can interpret and integrate, matching the timing, dynamics, and structure of natural sensory input.”</p>
<p id="p-25">In other words, stimulation alone does not create perception; the brain has to make sense of what it receives.</p>
<p id="p-26">Compounding this problem is the brain’s constant plasticity. Neural pathways reorganize as we learn and adapt, meaning an artificial implant “must be dynamic rather than static,” Maimon said. It must be “capable of adjusting not only to long-term neural reorganization, but also to the brain’s moment-to-moment variability across states of attention, context, and experience.”</p>
<p id="p-27">Achin Bhowmik, CTO at Eden Prairie, MN-based Starkey Hearing, agreed, noting the technical challenges remain “formidable.” He cited key hurdles such as developing energy-efficient, real-time AI algorithms that can operate at the edge, creating biocompatible interfaces that communicate seamlessly with the nervous system and, perhaps most importantly, understanding how to harmonize artificial inputs with the brain’s perceptual and cognitive processes.</p>
<p id="p-28">This challenge of creating a feeling of perception is also a challenge for non-invasive systems.</p>
<p id="p-29">“It is not enough for users to decode patterns. The feedback must evoke a sensory quality that feels immediate and grounded in perception,” Maimon said, adding that it must do so without requiring extensive training or cognitive effort.</p>
</section>
<section id="sec4" class="sec">
<h2 class="heading">From Restoration to Enhancement</h2>
<p id="p-30">For much of its history, sensory augmentation was driven by a restorative goal. But as the technology matures, that distinction is fading.</p>
<p id="p-31">“Looking ahead, I believe the distinction between restoration and enhancement will fade,” said Bhowmik. “Once we learn to emulate and extend our natural senses, the same technologies that restore hearing or vision will also enable people to perceive the world in entirely new ways.”</p>
<p id="p-32">This new focus on augmentation generally falls into two categories. The first is enhancement, which aims to refine existing senses, such as sharpening vision, or improving touch or hearing. The second is extension, which moves further by granting access to sensory information humans do not naturally perceive, such as thermal, ultrasonic, or magnetic information.</p>
<p id="p-33">This shift opens the door to a much larger potential user base.</p>
<p id="p-34">“We have become much better at preventing sensory loss over the last century,” Negen said, explaining that “the potential user base for sensory substitution is much smaller than the potential user base for sensory augmentation.”</p>
<p id="p-35">As new sensory skills combine with existing perception, they can enhance precision, making them a “true augmentation, not a fallback, even for someone without any sensory loss,” according to Negen. This creates an opportunity for niche but useful applications in areas like aging, working in hazardous environments, and perhaps even sports.</p>
<p id="p-36">The final frontier, however, is the creation of truly novel senses that translate symbolic or informational content into sensory form.</p>
<p id="p-37">“Imagine feeling air quality, perceiving electromagnetic fields, or intuitively sensing data streams,” said Maimon. “Such possibilities challenge our assumptions about perception, cognition, and the boundaries of human experience. They suggest that sensing is not merely biological, but something that can be designed, extended, and reimagined.”</p>
<p id="p-38">This future is arriving in subtle ways. Bhowmik pointed to modern hearing aids, which have evolved far beyond simple sound amplification. “Modern hearing aids, for instance, use deep neural networks to separate speech from background noise, monitor physiological signals, and even detect falls,” he said. “They effectively function as miniature AI computers for the ear.”</p>
</section>
</div>
<h2 id="FurtherReading" class="heading">Further Reading</h2>
<ul id="reflist1" class="ref-list">
<li class="ref">
<div id="B1" class="citation"><span class="mixed-citation" data-jats-publication-type="other"><em>Bhowmik, A.</em> <br /><strong>Virtual and augmented reality: Human sensory-perceptual requirements and trends for immersive spatial computing experiences, <em>Journal of the Society for Information Display</em>, Aug. 11, 2024, </strong><a class="ext-link" href="https://sid.onlinelibrary.wiley.com/doi/abs/10.1002/jsid.2001" data-jats-ext-link-type="uri"><strong>https://sid.onlinelibrary.wiley.com/doi/abs/10.1002/jsid.2001</strong></a></span></div>
</li>
<li class="ref">
<div id="B2" class="citation"><span class="mixed-citation" data-jats-publication-type="other"><em>Huang, S.</em> <br /><strong>China’s rare mineral tellurium gives blind animals ‘super vision’, offers hope for humans, <em>South China Morning Post</em>, Jun. 8, 2025, </strong><a class="ext-link" href="https://www.scmp.com/news/china/science/article/3313402/chinas-rare-mineral-tellurium-gives-blind-animals-super-vision-offers-hope-humans" data-jats-ext-link-type="uri"><strong>https://www.scmp.com/news/china/science/article/3313402/chinas-rare-mineral-tellurium-gives-blind-animals-super-vision-offers-hope-humans</strong></a></span></div>
</li>
<li class="ref">
<div id="B3" class="citation"><span class="mixed-citation" data-jats-publication-type="other"><em>Maimon, A.</em> <br /><strong>Perceiving depth beyond sight: Evaluating intrinsic and learned cues via a proof of concept sensory substitution method in the visually impaired and sighted, <em>PLOS ONE</em>, Sep. 25, 2024, <a class="ext-link" href="https://journals.plos.org/plosone/article?id=10.1371/journal.pone.0310033" data-jats-ext-link-type="uri">https://journals.plos.org/plosone/article?id=10.1371/journal.pone.0310033</a></strong></span></div>
</li>
</ul>
</article>
]]></content:encoded>
					
					<wfw:commentRss>https://cacm.acm.org/news/the-new-era-of-sensory-augmentation/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
		<post-id xmlns="com-wordpress:feed-additions:1">776832</post-id>	</item>
		<item>
		<title>The Golden Rule of Big Memory: Persistence Is Not Harmful</title>
		<link>https://cacm.acm.org/opinion/the-golden-rule-of-big-memory-persistence-is-not-harmful/</link>
					<comments>https://cacm.acm.org/opinion/the-golden-rule-of-big-memory-persistence-is-not-harmful/#respond</comments>
		
		<dc:creator><![CDATA[Yu Hua, Xue Liu, and Ion Stoica]]></dc:creator>
		<pubDate>Thu, 16 Apr 2026 16:33:57 +0000</pubDate>
				<category><![CDATA[Architecture and Hardware]]></category>
		<category><![CDATA[Data and Information]]></category>
		<guid isPermaLink="false">https://cacm.acm.org/?post_type=digital-library&#038;p=779178</guid>

					<description><![CDATA[<p>Merging the performance of volatile memory with the reliability of persistent storage reshapes the foundation of memory-centric computing.</p>]]></description>
										<content:encoded><![CDATA[<article>
<div class="body" lang="en">
<section id="sec1" class="sec">
<p id="p-1">Conventional memory schemes follow the Pareto Principle, in which approximately maintaining 20% hot data can meet 80% of requests. Large-scale applications, such as generative AI, recommendation systems, big data, and HPC systems, require large-capacity and high-speed memory and are changing the power-law locality, which necessitate the support of <i>Big Memory</i>. Big Memory is a transformative memory-centric system that consolidates massive and heterogeneous memory resources into a unified and shared address space, delivering near-DRAM latency and persistence at a terabyte-to-petabyte scale. This paradigm mitigates traditional I/O bottlenecks, simplifies data placement and programming models, and unlocks unprecedented scalability. Memory is one of the most contended hardware resources and often becomes the performance bottleneck in the I/O critical path that connects computing with storage. Due to limited memory capacity, more applications have to explore and exploit the memory beyond traditional local main memory, including empty memory on remote servers, and disaggregated memory in a memory pool.<a class="reference-link xref xref-bibr" href="#B7" data-jats-ref-type="bibr" data-jats-rid="B7"><sup>7</sup></a> This non-local far memory can expand memory size and avoid memory stranding. Since far memory is much slower than local memory, existing systems leverage local memory as a cache by transparently swapping memory pages between local and far memory, which unfortunately causes read or write amplification.</p>
<p id="p-2">In order to deliver high performance and achieve cost-efficiency, we need to carefully consider memory overhead that comes from the differential requirements of high-level applications and low-level devices. Specifically, high-level applications generally include peer-to-peer, blockchain, artificial intelligence, computability storage, and so forth. Different storage applications exhibit different needs and access patterns. On the other hand, the low-level devices become heterogeneous and highly hierarchical, which exhibit different physical properties.<a class="reference-link xref xref-bibr" href="#B3" data-jats-ref-type="bibr" data-jats-rid="B3"><sup>3</sup></a> For example, hard disks, SSD, and PCM are non-volatile, and DRAM, cache, and registers are volatile. Hard disks and SSD are block-addressable, while others are byte-addressable. SSD and PCM suffer from the endurance with limited writes. These two-level different requirements, like a hamburger, introduce the high memory overhead, which means that to support program execution in memory, we need to consume extra costs for data movements and energy consumption, as well as high-security risks.<a class="reference-link xref xref-bibr" href="#B1" data-jats-ref-type="bibr" data-jats-rid="B1"><sup>1</sup></a> To alleviate high memory overhead, we build big memory that not only provides large capacity but also supports fast persistence. The persistence guarantees a complete lifetime for data to be finally written into non-volatile devices via protocol buffers or checkpoints.<a class="reference-link xref xref-bibr" href="#B2" data-jats-ref-type="bibr" data-jats-rid="B2"><sup>2</sup></a> Since most operations can be executed and completed within the big memory, we significantly reduce data movements through the memory and storage hierarchy.</p>
<p id="p-3">In this hierarchical architecture, modern networking technologies like RDMA and CXL are erasing the boundaries between memory and storage, turning distributed resources into a seamless, low-latency memory fabric. CXL enables elastic, high-speed memory pooling across servers at near-DRAM speed, whereas RDMA facilitates efficient data transfers by bypassing CPU and software overheads. Instead of slow block storage, systems now operate on byte-addressable memory, where persistence is a property. Traditional cache, DRAM, and disk hierarchies become inefficient, thus replacing them with a scalable network-spanning memory system that grows in performance and capacity as the infrastructure expands.</p>
<p id="p-4">The transformation of conventional memory into big memory systems leverages both vertical and horizontal extensions to existing memory hierarchies. Vertically, this expansion enables each node to host significantly larger datasets by increasing memory capacity and delivering high performance within a flattened hierarchy that coalesces traditional memory and storage tiers into a unified, high-speed access layer. Horizontally, the architecture scales out across distributed systems, integrating multiple nodes into a pool that collectively provides massive, scalable memory resources while maintaining efficient data access and coherence across the entire system. More data is aggregated together via atomic operation guarantee, which actually coalesces the central and distributed designs. Hence, by leveraging the vertical and horizontal ways, we build a big computer, as shown in Figure <a class="xref xref-fig" href="#F1" data-jats-ref-type="fig" data-jats-rid="F1">1</a>.</p>
<figure id="F1" class="fig" data-jats-position="float">
<div class="image-container"><img decoding="async" class="graphic" title="Figure 1. " src="https://cacm.acm.org/wp-content/uploads/2026/03/3769003_fig01.jpg" alt="" data-image-id="F1" data-image-type="figure" /></div><figcaption><span class="caption-label">Figure 1. </span> <span class="p">Big computer coalesces the memory and storage, and the central and distributed schemes.</span></p>
<div class="figcaption-footer"> </div>
</figcaption></figure>
<p id="p-6">The foundational rationale for this architecture stems from the influential methodology of “the computer as a network.” This perspective draws a structural analogy between traditional computer hierarchy, which comprises low-level storage, intermediate memory, and high-level CPUs, and networked systems, where analogous roles are fulfilled by remote cloud resources, adjacent edge nodes, and end-user clients, respectively. This conceptual mapping allows for the application of well-understood network principles to the design of integrated computing systems, thereby enabling more scalable, efficient, and flexible resource orchestration across localized and distributed environments. In the context of a big computer, these items share some similarities. The long-latency storage seems like a remote cloud, the high-speed memory is like an adjacent edge, and the computing units are like clients that generate data. After being generated from clients/CPUs, data will be instantaneously transmitted to the close-by edges/memory, and consume a long latency to arrive at the remote clouds/storage.</p>
<p id="p-7">This methodology can be implemented via existing devices. For example, non-volatile devices offer about TB-scale memory capacity, which can be considered a vertical extension. Moreover, from the horizontal view, RDMA and CXL protocols support memory and cache coherence.<a class="reference-link xref xref-bibr" href="#B7" data-jats-ref-type="bibr" data-jats-rid="B7"><sup>7</sup></a> The big computer hence achieves persistence in both vertical and horizontal extensions. A program in a big computer can be atomically executed in an end-to-end way, whether from CPU to non-volatile devices or from clients to clouds.</p>
</section>
<section id="sec2" class="sec">
<h2 class="heading">Understanding the &#8216;Big&#8217;</h2>
<p id="p-8">In the context of big memory, we need to clarify the <i>BIG</i>, which is also the essential difference between big memory and traditional one. To fully understand the mentioned BIG, we first discuss other related BIG terms in the research community. Specifically, the difference between data and big data is the well-known Vs, including volume, velocity, variety, veracity, and value. Moreover, the difference between a model and a big model comes from algorithms, data, and computation. Compared with a chip, the big chip needs to consider the programmable, scalable, highly available, and adaptive properties. The salient feature behind BIGs is their Golden Rule.</p>
<p id="p-9">The transition from conventional memory to “big memory” necessitates a clear understanding of their fundamental differences and the establishment of a guiding principle, that is, the “Golden Rule of Big Memory.” Vertically, this evolution signifies not only an expansion in capacity but also an extension in effective access distance, moving beyond local DRAM to incorporate broader hierarchies. Horizontally, it introduces immense scaling complexity in distributed systems deploying numerous GPUs for large model training. In such environments, horizontal scaling often relies on distributed shared memory to temporarily maintain intermediate results as replicas in remote volatile memory, offering probabilistic persistence but posing high risks of data loss. Without robust persistence guarantee, these volatile checkpoints consume excessive memory space and incur substantial broadcast overhead for synchronization and recovery.<a class="reference-link xref xref-bibr" href="#B5" data-jats-ref-type="bibr" data-jats-rid="B5"><sup>5</sup></a> This problem is exemplified by large language models, whose checkpoint states reach terabyte scales, which far surpasses the capacity of individual GPU memories. This leads to significant performance degradation when remote GPUs cannot host entire checkpoints.</p>
</section>
<section id="sec3" class="sec">
<h2 class="heading">The Golden Rule</h2>
<p id="p-10">We believe that <i>persistence</i> is the Golden Rule of Big Memory systems. Contemporary big memory architectures have fundamentally reimagined data persistence, transforming it from a system constraint into a powerful performance enabler. This shift is driven by hardware-optimized approaches. Specifically, speculative persistence employs parallel RDMA replication to maintain multiple synchronized copies across nodes, enabling non-blocking durable operations. Deterministic persistence leverages CXL-connected NVM with efficient cache management to deliver memory-tier durability without traditional storage penalties. The result is a paradigm where persistence accelerates rather than hinders operations. In-memory databases achieve instant commits, machine learning systems maintain seamless checkpoints and distributed applications preserve consistency at memory speeds. By merging the performance of volatile memory with the reliability of persistent storage, this advancement is reshaping the foundation of memory-centric computing.</p>
<p id="p-11">A case in point comes from an insight that both Von Neumann and Non-Von Neumann architectures need to guarantee the persistence. In practice, the memory/storage hierarchy in the Von Neumann architecture is a multi-level subset model, which includes disks, SSD, PCM, DRAM, L1/L2/L3 caches, and registers. In this subset model, the devices in the lower levels contain the data in the higher ones. All data needs to be written into persistent devices. On the other hand, for Non-Von Neumann architecture, persistence still plays an important role. A typical workflow in the memristors needs to compute the input vector in the matrix and the result is a new vector that is volatile. By using extra operations, the new vector can be finally persisted in the non-volatile devices. Hence, both Von Neumann and Non-Von Neumann architectures follow the persistence principle.</p>
<p id="p-12">There exist some non-persistence operations in conventional systems, for example, replicating data into the DRAMs across servers or writing data in a batch through a cache/memory/storage hierarchy. These operations often cause misleading the understanding that persistence is harmful to high performance, which is actually not. To avoid the high overheads of rereading or recomputing the lost data, systems persist data but suffer from the huge gap between high-speed memory and low-speed disks. Many efforts have to be put into maintaining data in a volatile manner, which is vulnerable to just one system crash or failure. This awkward situation can be efficiently addressed using the big memory that offers large-size non-volatility with a shortened critical path. For example, eADR (extended Asynchronous DRAM Refresh) offers persistence for in-cache data and guarantees to flush them into non-volatile devices in case of system crash or failure. Hence, the number of flushing data is significantly reduced even during normal system execution. Systems thus deliver high performance with the aid of persistence.</p>
<p id="p-13">Conventional computing systems operate on a heterogeneous hierarchy: fast, byte-addressable, volatile main memory, and slow, block-addressable, non-volatile storage. This separation necessitates data to migrate frequently and expensively between the two tiers to achieve persistence, which not only introduces significant performance overheads due to constant data swapping but also amplifies the risk of consistency bugs. Persistence needs to be guaranteed by storing states and data in non-volatile storage devices. A program hence transfers states and data to and from storage devices via an I/O bus, which is supported by leveraging specific instructions in a program. Persistence is important to deal with system failure and power down without data loss.</p>
<p id="p-14">Persistent memory<a class="reference-link xref xref-bibr" href="#B9" data-jats-ref-type="bibr" data-jats-rid="B9"><sup>9</sup></a> fundamentally rearchitects data persistence by merging memory and storage into a single tier, offering DRAM-like performance with disk-like non-volatility through a unified, flat address space. This design allows applications to maintain consistent states directly in memory and flush them to non-volatile media in batches,<a class="reference-link xref xref-bibr" href="#B6" data-jats-ref-type="bibr" data-jats-rid="B6"><sup>6</sup></a> reducing the costly data movements while simultaneously avoiding security vulnerabilities associated with bus snooping. The flat addressing model simplifies error-prone persistence logic, enables near-instantaneous recovery, and permits direct manipulation of persistent data. However, while technologies like eADR ensure the persistence of data residing in multi-level CPU caches, data within registers remain vulnerable to power loss, as onboard power capacitors cannot efficiently flush register contents to non-volatile media. To address this limitation, modern systems employ checkpointing techniques to capture and guarantee the persistence of full-stack architectural states, ensuring comprehensive recovery integrity across the entire hierarchy.</p>
</section>
<section id="sec4" class="sec">
<h2 class="heading">Full-Stack and Moving Persistence</h2>
<p id="p-15">As shown in Figure <a class="xref xref-fig" href="#F2" data-jats-ref-type="fig" data-jats-rid="F2">2</a>, achieving full-stack persistence across the entire memory hierarchy represents a paradigm shift in system design, fundamentally moving the persistence boundary from the traditional interface between SSDs and DRAM directly into CPU. This transformation is enabled through a combination of hardware and software mechanisms. Persistent memory provides foundational data persistence at scale due to its inherent non-volatile physical properties. ADR ensures that data within the memory controller are flushed to non-volatile media during power events. eADR further guarantees that data residing in multi-level CPU caches are persisted. To address the volatility of registers, which remain beyond the reach of capacitor-based flushing mechanisms, systems employ frequent and expensive checkpointing techniques to capture computational states.<a class="reference-link xref xref-bibr" href="#B8" data-jats-ref-type="bibr" data-jats-rid="B8"><sup>8</sup></a> Consequently, the entire hierarchy, from computational registers to large-scale persistent memory, now supports end-to-end persistence, enabling not only near-instantaneous recovery and simplified programming models but also a substantial expansion of effective memory capacity available for persistent operations.</p>
<figure id="F2" class="fig" data-jats-position="float">
<div class="image-container"><img decoding="async" class="graphic" title="Figure 2. " src="https://cacm.acm.org/wp-content/uploads/2026/03/3769003_fig02.jpg" alt="" data-image-id="F2" data-image-type="figure" /></div><figcaption><span class="caption-label">Figure 2. </span> <span class="p">The full-stack and moving persistence.</span></p>
<div class="figcaption-footer"> </div>
</figcaption></figure>
<p id="p-17">Full-stack persistence requires the support from non-volatile devices and periodic checkpointing to facilitate fast recovery from system crashes or power failures. Although this hierarchy supports full-stack persistence, the implementations are different. For persistent memory, persistence can be naturally obtained due to its physical features. However, the memory controller and multi-level caches, which are volatile, require extra power to flush data into non-volatile devices, called moving persistence. Since data move via the I/O bus, there exist security issues due to bus snooping. These problems can be addressed by efficiently coordinating the hierarchy, capacity, and operations for high performance and strong security, thus achieving a suitable trade-off between performance and security.</p>
<p id="p-18">For large-scale distributed systems, we conventionally leverage IP protocol to connect the hard disks in multiple nodes. RDMA is further used to aggregate memory resources together. CXL can achieve the cache coherence.<a class="reference-link xref xref-bibr" href="#B4" data-jats-ref-type="bibr" data-jats-rid="B4"><sup>4</sup></a> These designs allow the processing boundary to move up and more operations can be completed in the higher levels. Since data need to be finally written into non-volatile devices, the end-to-end critical path can be significantly shortened. We hence achieve the savings of data movements and decrease energy consumption, as well as alleviating bus conflicts and side-channel attacks.</p>
</section>
<section id="sec5" class="sec">
<h2 class="heading">Quantitative Validation</h2>
<p id="p-19">To demonstrate the real-world impact of persistence, we select our state-of-the-art designs and real implementations for persistent memory, which provide detailed comparisons with standard non-persistent memory systems (that is, conventional DRAM-based systems). As shown in the accompanying table, the widely used evaluation metrics include throughput, memory overhead, and typical operation latency. The comparisons leverage various workloads from real-world applications and large-scale benchmarks. These workloads exhibit the typical patterns of existing and emerging applications in the context of big memory. We mainly show the average values due to space limitation. More details about configurations can be found in the experiments of our original papers. Moreover, applications consist of transactions, indexing, and security, which are widely used and well recognized in the memory community. Due to the heterogeneity of these applications, we present the evaluation results normalized to non-persistent memory to demonstrate the strength of persistence.</p>
<p id="p-20">Persistent memory systems achieve significant performance improvements in all evaluation metrics. Specifically, in terms of throughput, the persistent systems achieve almost two times improvements compared with the non-persistent ones. The memory space can be significantly saved in most applications. Furthermore, the typical operations decrease the latency in the persistent memory systems. The main reason comes from the fact that the persistence guarantees consistency, integrity and reliability, which essentially decrease system overheads and avoid extra costs, such as frequent data movements and CPU/memory instructions.</p>
<figure id="UT1" class="table-wrap">
<div class="caption"><span class="caption-label">Table. </span></p>
<div class="title">Quantitative Performance Comparisons in terms of multiple metrics.</div>
</div>
<div class="table-container">
<table class="table table-bordered table-condensed table-hover" data-jats-frame="hsides" data-jats-rules="rows">
<colgroup>
<col align="center" valign="top" />
<col align="center" valign="top" />
<col align="center" valign="top" />
<col align="center" valign="top" />
<col align="center" valign="top" /> </colgroup>
<thead style="vertical-align: bottom;">
<tr>
<th style="text-align: center;" rowspan="2">Applications</th>
<th style="text-align: center;">Metrics</th>
<th style="text-align: center;" rowspan="2">Throughput</th>
<th style="text-align: center;" rowspan="2">Memory Overhead</th>
<th style="text-align: center;" rowspan="2">Operation Latency</th>
</tr>
<tr>
<th style="text-align: center;">Examples</th>
</tr>
</thead>
<tbody>
<tr>
<td style="text-align: center;"><b>Non-Persistent Memory</b></td>
<td style="text-align: center;">Standard</td>
<td style="text-align: center;">1</td>
<td style="text-align: center;">1</td>
<td style="text-align: center;">1</td>
</tr>
<tr>
<td style="text-align: center;" rowspan="2"><b>Transactions</b></td>
<td style="text-align: center;">Motor[OSDI 24]</td>
<td style="text-align: center;">2.15</td>
<td style="text-align: center;">0.57</td>
<td style="text-align: center;">N/A</td>
</tr>
<tr>
<td style="text-align: center;">SiLo[HPCA 23]</td>
<td style="text-align: center;">3.63</td>
<td style="text-align: center;">0.24</td>
<td style="text-align: center;">N/A</td>
</tr>
<tr>
<td style="text-align: center;" rowspan="4"><b>Indexing</b></td>
<td style="text-align: center;">DALdex[ICS 25]</td>
<td style="text-align: center;">2.66</td>
<td style="text-align: center;">0.61</td>
<td style="text-align: center;">N/A</td>
</tr>
<tr>
<td style="text-align: center;">Level[OSDI 18]</td>
<td style="text-align: center;">N/A</td>
<td style="text-align: center;">N/A</td>
<td style="text-align: center;">0.42(Insert)/0.51(Update)</td>
</tr>
<tr>
<td style="text-align: center;">LightCheck[ICCD 23]</td>
<td style="text-align: center;">N/A</td>
<td style="text-align: center;">0.62</td>
<td style="text-align: center;">0.07(Recovery)</td>
</tr>
<tr>
<td style="text-align: center;">FINEdex[VLDB 22]</td>
<td style="text-align: center;">2.26</td>
<td style="text-align: center;">0.05</td>
<td style="text-align: center;">N/A</td>
</tr>
<tr>
<td style="text-align: center;" rowspan="2"><b>Security</b></td>
<td style="text-align: center;">STAR[HPCA 21]</td>
<td style="text-align: center;">N/A</td>
<td style="text-align: center;">N/A</td>
<td style="text-align: center;">0.35 (Write)</td>
</tr>
<tr>
<td style="text-align: center;">Secon[DAC 22]</td>
<td style="text-align: center;">1.95</td>
<td style="text-align: center;">N/A</td>
<td style="text-align: center;">N/A</td>
</tr>
</tbody>
</table>
</div>
</figure>
</section>
<section id="sec6" class="sec">
<h2 class="heading">Shaping the Future</h2>
<p id="p-21">Big memory systems, leveraging technologies like persistent memory, large-scale DRAM, and disaggregated memory architectures, have transformed data-intensive computing by enabling faster access to vast datasets. However, significant challenges remain in ensuring efficient and reliable operations. Crash consistency and durability in big memory systems pose a key challenge, since balancing performance with correctness is non-trivial. Memory disaggregation introduces latency and contention issues, while security concerns such as persistent memory vulnerabilities and secure memory sharing in distributed setups require further exploration. Moreover, programmability is still difficult, since developers lack universal models for managing big memory systems, and debugging persistent applications is still complex. Addressing these challenges is crucial for advancing big memory adoption.</p>
<p id="p-22">To overcome these challenges, hardware and software innovations are essential. Hardware advancements, such as CXL 3.0+ for improved memory pooling and near-memory processing to minimize data movement, could significantly boost performance. On the software side, persistent memory-aware file systems, lightweight checkpointing mechanisms, and machine learning-driven memory tiering could optimize data placement across DRAM and NVM. Distributed coherence protocols and hardware-accelerated RDMA persistence mechanisms are also needed to ensure efficient memory sharing in disaggregated environments. Multi-tier memory techniques are narrowing performance gaps and enabling scalable, cost-effective memory systems. CXL, in particular, seamlessly integrates persistent memory, GPUs, and accelerators, allowing direct memory access and improving system scalability.</p>
<p id="p-23">Beyond technical improvements, big memory systems unlock novel applications across various domains. In AI and real-time analytics, they enable in-memory machine learning training and large-scale graph processing. Databases benefit from NVM-optimized designs that support instant recovery and high-speed transactions, while scientific computing leverages big memory for in-situ visualization and simulations. Edge and cloud computing can adopt memory-as-a-service (MaaS) models for dynamic resource allocation. Modern networking technologies are flattening memory and storage hierarchy, creating a seamless, low-latency memory fabric. This evolution enables byte-addressable persistence, replacing traditional storage hierarchies with scalable, network-spanning memory systems. As research progresses, the focus remains on hardware-software co-design, intelligent memory management, and expanding use cases in emerging fields, thus paving the way for transformative paradigms.</p>
</section>
</div>
<footer class="back"></footer>
</article>
]]></content:encoded>
					
					<wfw:commentRss>https://cacm.acm.org/opinion/the-golden-rule-of-big-memory-persistence-is-not-harmful/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
		      <dc:creator><![CDATA[Xue Liu]]></dc:creator>
      <dc:creator><![CDATA[Ion Stoica]]></dc:creator>
<post-id xmlns="com-wordpress:feed-additions:1">779178</post-id>	</item>
		<item>
		<title>Subscription Bombing: Email under Attack</title>
		<link>https://cacm.acm.org/practice/subscription-bombing-email-under-attack/</link>
					<comments>https://cacm.acm.org/practice/subscription-bombing-email-under-attack/#respond</comments>
		
		<dc:creator><![CDATA[Mathias Payer and Candid Wüest]]></dc:creator>
		<pubDate>Wed, 15 Apr 2026 14:16:09 +0000</pubDate>
				<category><![CDATA[Computing Applications]]></category>
		<category><![CDATA[Data and Information]]></category>
		<category><![CDATA[Security and Privacy]]></category>
		<guid isPermaLink="false">https://cacm.acm.org/?post_type=digital-library&#038;p=779431</guid>

					<description><![CDATA[<p>An analysis of 24 subscription bombing attacks.</p>]]></description>
										<content:encoded><![CDATA[<article>
<div class="body" lang="en">
<section id="sec1" class="sec">
<p id="p-1">Email subscription bombing (also known as subscription flooding or email spam bombing) is an attack technique that overwhelms a target’s inbox with a high volume of <i>benign but unwanted</i> emails. The emails themselves are benign and are sent by legitimate third parties (such as mailing lists), and therefore typically pass spam filters. The origins of this attack vector trace back to the beginning of the public Internet but have recently re-emerged at scale. Unlike classic email bombing, where attackers send large volumes directly, subscription bombing introduces a layer of indirection. The attacker registers the victim&#8217;s email address with thousands of third-party online services, which act as amplifiers that flood the target&#8217;s inbox with legitimate subscription confirmations and notification messages. Such emails are much harder to block automatically by the receiving email server.</p>
<p id="p-2">Modern spam filtering relies on multiple signals, including sender reputation, email headers, and message content. For subscription attacks, the sending server generally has a good reputation and is not on a public blocklist; header checks (such as SPF, DMARC, or DKIM) are usually passed, and the subject and body text are different for each email and generally do not match typical phishing phrases that could be detected. While the flood of messages renders the inbox unusable, the true objective is often not just denial-of-service but to hide specific legitimate emails that arrive during the attack, thereby obscuring unauthorized financial transactions, account compromises, or ransomware deployment attempts. Early attacks scanned the Internet for sign-up forms and then scripted the sign-up process. In 1996, a stockbroker was bombarded with 25,000 emails, gathering some media attention at the time.<a class="reference-link xref xref-bibr" href="#B1" data-jats-ref-type="bibr" data-jats-rid="B1"><sup>1</sup></a> Since then, attacks have come in waves and were modernized. In 2016, hundreds of email addresses of the U.S. government were targeted, and some were subscribed to more than 10,000 newsletters.<a class="reference-link xref xref-bibr" href="#B6" data-jats-ref-type="bibr" data-jats-rid="B6"><sup>6</sup></a> As organizations increasingly rely on digital communication channels, understanding the mechanics of these orchestrated distractions and their role in broader attack chains has become essential for maintaining operational continuity and financial security.</p>
<p id="p-3">Beyond newsletters, attackers exploit other services capable of generating emails to a chosen target. This includes password reset requests, account registration confirmations, customer support forms, promotional emails, and social media notifications. These third-party services act as unwitting relays, with attackers hijacking their good reputation and bandwidth to execute the subscription bombing attack. We have monitored email trends since 2015 and have observed a recent increase in such attacks. In this study, we analyze 24 concrete subscription bombing attack campaigns using a large dataset from an email security provider, encompassing 46,970 involved unwanted emails. We define key metrics, outline attack timelines, and provide insight into the operational patterns of these campaigns. Furthermore, we examine subscription bombing services offered on the Dark Web and categorize their capabilities. Mitigating these attacks is challenging due to the large set of email senders, but automated unsubscription may reduce the impact. Given the simplicity of carrying out these asymmetric subscription bombing attacks, we expect the volume to grow in the future.</p>
</section>
<section id="sec2" class="sec">
<h2 class="heading">Subscription Bombing</h2>
<p id="p-4">Attackers typically abuse services that are well known, reputable, and easy to automate using sign-up scripts. Newsletters are the most common targets, followed by account registration confirmation messages. However, attackers have also been seen posting product listings, job postings, support inquiries, hotel bookings, or even apartment listings on online platforms, generating genuine inquiries from human users. The third-party services involved are not compromised or necessarily misconfigured, and no vulnerability in the email server is exploited. The attackers are simply abusing benign (but sometimes overly permissive) features.</p>
<p id="p-5">The underlying security issue for all services is that, essentially, an attacker can register the victim, without their permission, who will then receive unwanted messages in their inbox. Key to eradicating this attack vector is that each service will have to improve its sign-up process with a two-step confirmation procedure. There are three key challenges here: a lot of distinct services will have to be updated; a service provider may not be interested in improving its security if it considers the number of registered users as a success metric; and even with proper validation the first verification email can still be used to flood victims. As a result, email service providers must mitigate these floods of unwanted emails. Existing academic research on subscription bombing<a class="reference-link xref xref-bibr" href="#B13" data-jats-ref-type="bibr" data-jats-rid="B13"><sup>13</sup></a> so far targets bypassing filters<a class="reference-link xref xref-bibr" href="#B9" data-jats-ref-type="bibr" data-jats-rid="B9"><sup>9</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B14" data-jats-ref-type="bibr" data-jats-rid="B14"><sup>14</sup></a> or detects unwanted emails through ML-based mechanisms.<a class="reference-link xref xref-bibr" href="#B8" data-jats-ref-type="bibr" data-jats-rid="B8"><sup>8</sup></a></p>
<p id="p-6">Some Dark Web services claim to have access to more than 100,000 mailing lists and services. Advances in artificial intelligence (AI) agents may further simplify the process of adding new services to the database of newsletters that can be abused, complete with all the necessary details about which HTTP POST or API request needs to be performed to subscribe new victims. Discussions with supporters of an underground subscription bombing service confirmed that they frequently update their lists with new services.</p>
<p id="p-7">Subscription bombing is offered as a service but may be run by individual attackers themselves. All an attacker needs is a large list of newsletter providers and an automation script to subscribe the victim&#8217;s email. An attack consists of a list of victims&#8217; email addresses that will receive the unwanted emails and a list of sender services which are sending the emails to the victims. A single attack wave consists of the attacker adding the victims address to a preselected number of services within a given timeframe. There can be multiple attack waves in a larger campaign against the same victim. The process of creating a subscription bombing service includes three broad steps:</p>
<ol class="list" data-jats-list-type="order">
<li class="list-item">
<p id="p-8">Build a database of services that send emails by, for example, crawling the Web for newsletters.</p>
</li>
<li class="list-item">
<p id="p-9">Create a script that automatically subscribes new emails to a subset of the database.</p>
</li>
<li class="list-item">
<p id="p-10">Periodically verify the list and replace faulty entries with new ones.</p>
</li>
</ol>
<p id="p-11">Underground forum discussions indicate that some groups simply record the POST requests required to sign up for newsletters and then replay it with the new target address. This is not always trivial, as many websites use CAPTCHAs or <i>cross site request forgery</i> (CSRF) tokens to prevent single-call subscriptions. In the services abused in our attack case study, only 4% used a simple HTTP GET request to subscribe an email address. Around 70% used some form of nonce or CSRF token, requiring an initial HTTP GET request to obtain the token, followed by an HTTP POST request to complete the subscription. By using browser developer tools, an attacker can fill out the required form once, log all requests, and then export the necessary calls directly as a CURL command. This semi-automated process allows attackers to manually build a database. It is also plausible that some attackers targeted WordPress newsletter plug-ins and similar tools, as these allow them to analyze the subscription process once and replicate it across multiple domains. To test this hypothesis, we analyzed our dataset to identify the content management system (CMS) used by the websites, utilizing the service WhatCMS.org for identification. While WordPress was the most commonly used CMS among the sending domains (with a total of 16%, followed by Presta Shop with 6% and Laravel with 5%), we could not identify a single CMS plug-in or online shop environment that would indicate a massive monoculture.</p>
<p id="p-12">With the rise of agentic AI and browser automation tools, attackers can easily commandeer a Web browser and have the agent automatically fill in the target email address in subscription forms. Services such as airtop.ai, fellou.ai, or axiom.ai can automate browser interactions and even assist with scaling and CAPTCHA solving. We successfully tested such a setup for newsletter subscriptions. However, to achieve performance and scalability, a paid service would be required. We have not seen any indication that attackers are already using these paid automation services for their attacks. This may change in the future with the evolution of AI tools. Since each newsletter service is hosted by a different Web entity, there is no central mechanism to detect or block multiple requests from the same IP address. Nevertheless, most attack services we analyzed claim to utilize proxies to rotate IP addresses and implement rate limitations to avoid getting blocked.</p>
</section>
<section id="sec3" class="sec">
<h2 class="heading">Large-Scale Dataset</h2>
<p id="p-13">Our dataset comprises 46,970 unwanted emails collected across 24 subscription bombing attack waves. The dataset includes <i>email header</i>, <i>To</i> and <i>From</i> <i>address</i>, and <i>subject line</i>. After receiving approval from the involved parties, we gathered the data in the period from Oct. 9, 2024 until Dec. 4, 2024 from midsize and large European customers. Note that collecting emails to form such a study is challenging, as the email service provider, the involved company, and individuals have to agree to the data collection. The dataset was collected by the email security solution provider xorlab AG and contains timestamp, sender email address, subject line, anonymized target email address, and extracted tags from email header.</p>
<p id="p-14">The largest attack wave delivered 4,847 emails in under two hours or 3,387 emails per hour, while the smallest wave delivered 81 emails in one hour. The average number of emails per attack was 1,957 emails or 1,516 emails per hour. The average number of emails received by these accounts on normal days was less than 10 per hour. The attackers targeted various industries, including retail, technology, and manufacturing, with no clear pattern. The victims were typically individuals in management positions—director and above—with roles involving external visibility. Their email addresses were publicly accessible on open websites or easy to guess. Victims experienced varying levels of disruption: 71% of them were targeted once, 18% were hit twice, and 12% faced three separate attacks.</p>
<p id="p-15">Half the attacks occurred on Friday (<a class="xref xref-fig" href="#F1" data-jats-ref-type="fig" data-jats-rid="F1">Figure 1</a>), marking the last day of the work week for these targets. Friday is a common day for cyberattacks, as people may be inclined to leave early, allowing the attacks to go undetected over the weekend. Furthermore, IT security teams might operate with limited staff numbers on the weekend. The attacks most frequently began between 8:00 am and 10:00 am (<a class="xref xref-fig" href="#F2" data-jats-ref-type="fig" data-jats-rid="F2">Figure 2</a>), likely to overwhelm victims at the start of the day when they are most likely to check their inbox. Eighty percent of the attacks began during the morning hours of the targets. No attacks were recorded from 10:00 pm to 7:00 am, suggesting the attackers researched the working hours of their targets for targeted attacks. This timing also aligns with the potential objectives of the flooding attacks, as attackers often need to interact with either the victim or a target service—actions that are not feasible during nighttime hours.</p>
<figure id="F1" class="fig" data-jats-position="float">
<div class="image-container"><img decoding="async" class="graphic" title="Figure 1. " src="https://cacm.acm.org/wp-content/uploads/2026/04/3797487_fig01.jpg" alt="" data-image-id="F1" data-image-type="figure" /></div><figcaption><span class="caption-label">Figure 1. </span> <span class="p">Subscription bombing attack distribution per weekday.</span></p>
<div class="figcaption-footer"> </div>
</figcaption></figure>
<figure id="F2" class="fig" data-jats-position="float">
<div class="image-container"><img decoding="async" class="graphic" title="Figure 2. " src="https://cacm.acm.org/wp-content/uploads/2026/04/3797487_fig02.jpg" alt="" data-image-id="F2" data-image-type="figure" /></div><figcaption><span class="caption-label">Figure 2. </span> <span class="p">Subscription bomb attack distribution by hour.</span></p>
<div class="figcaption-footer"> </div>
</figcaption></figure>
</section>
<section id="sec4" class="sec">
<h2 class="heading">Case Study: Company Omega</h2>
<p id="p-18">The following analysis is focused on the data from two individual targets of subscription bombing within the same large international organization. Both persons work in upper management and were targeted with one day difference.</p>
<ul class="list" data-jats-list-type="bullet">
<li class="list-item">
<p id="p-19">Victim Alpha received 999 emails between Dec. 3, 2024 @ 10:19:45 and Dec. 3, 2024 @ 11:24:14</p>
</li>
<li class="list-item">
<p id="p-20">Victim Beta received 719 emails between Dec. 4, 2024 @ 12:50:29 and Dec. 4, 2024 @ 13:55:08</p>
</li>
</ul>
<p id="p-21">There was no overlap between the attack waves in our case study, meaning each victim received emails from different domains. However, we observed minor overlaps across other attacks, suggesting that the pool of available senders is large, with only marginal overlap between randomly selected subsets. The top-level domains (TLDs) of the sending domains vary, with a slight bias toward European countries. This was also reflected in the diverse languages of the newsletter and registration messages. As these emails are automatically generated, 95% of sending email addresses were non-individual accounts (non-human identities). With 26.5% the info@ was the most frequently observed sender account, followed by noreply@ with 7%. <a class="xref xref-table" href="#T1" data-jats-ref-type="table" data-jats-rid="T1">Table 1</a> shows details.</p>
<figure id="T1" class="table-wrap" data-jats-position="float">
<div class="caption"><span class="caption-label">Table 1. </span> <span class="p">Distributions of sender TLDs, account prefixes, and email subject keywords.</span></div>
<div class="table-container">
<table class="table table-bordered table-condensed table-hover">
<colgroup>
<col />
<col /> </colgroup>
<thead>
<tr>
<th><b>TLD</b></th>
<th><b>Percentage</b></th>
</tr>
</thead>
<tbody>
<tr>
<td colspan="2">(a) Top10 TLDs of senders.</td>
</tr>
<tr>
<td>com</td>
<td>52.79%</td>
</tr>
<tr>
<td>de</td>
<td>6.87%</td>
</tr>
<tr>
<td>pl</td>
<td>3.14%</td>
</tr>
<tr>
<td>org</td>
<td>3.14%</td>
</tr>
<tr>
<td>ru</td>
<td>2.68%</td>
</tr>
<tr>
<td>net</td>
<td>2.21%</td>
</tr>
<tr>
<td>it</td>
<td>2.10%</td>
</tr>
<tr>
<td>fr</td>
<td>1.63%</td>
</tr>
<tr>
<td>nl</td>
<td>1.57%</td>
</tr>
<tr>
<td>es</td>
<td>1.46%</td>
</tr>
<tr>
<td colspan="2">(b) Top 10 account prefixes.</td>
</tr>
<tr>
<td><b>Email Prefix</b></td>
<td><b>Percentage</b></td>
</tr>
<tr>
<td>info</td>
<td>26.54%</td>
</tr>
<tr>
<td>noreply</td>
<td>7.04%</td>
</tr>
<tr>
<td>wordpress</td>
<td>6.87%</td>
</tr>
<tr>
<td>support</td>
<td>5.59%</td>
</tr>
<tr>
<td>no-reply</td>
<td>4.54%</td>
</tr>
<tr>
<td>contact</td>
<td>4.07%</td>
</tr>
<tr>
<td>admin</td>
<td>2.79%</td>
</tr>
<tr>
<td>hello</td>
<td>2.62%</td>
</tr>
<tr>
<td>newsletter</td>
<td>1.80%</td>
</tr>
<tr>
<td>sales</td>
<td>1.22%</td>
</tr>
<tr>
<td colspan="2">(c) Top10 email subject keywords %.</td>
</tr>
<tr>
<td><b>Keyword</b></td>
<td><b>Percentage</b></td>
</tr>
<tr>
<td>confirm</td>
<td>11.3%</td>
</tr>
<tr>
<td>welcome</td>
<td>10.0%</td>
</tr>
<tr>
<td>email</td>
<td>8.0%</td>
</tr>
<tr>
<td>account</td>
<td>7.3%</td>
</tr>
<tr>
<td>password</td>
<td>7.0%</td>
</tr>
<tr>
<td>newsletter</td>
<td>6.9%</td>
</tr>
<tr>
<td>subscription</td>
<td>6.3%</td>
</tr>
<tr>
<td>reset</td>
<td>5.2%</td>
</tr>
<tr>
<td>verify</td>
<td>4.1%</td>
</tr>
<tr>
<td>address</td>
<td>3.7%</td>
</tr>
</tbody>
</table>
</div>
</figure>
<p id="p-23">Victim Alpha received 77 account reset emails, while Victim Beta received 49. All 126 services involved were neither financial nor high-profile social media platforms. Instead, they were primarily related to lifestyle and food websites. Therefore, it is unlikely the attackers attempted to take over any of these accounts for financial gain. Neither of the victims reported any issues with their accounts following the attack. Interestingly, many of these accounts had been freshly created just for the attack. The attacker first created the accounts and then requested a password reset for the newly generated account. The user names of the newly created accounts used random characters like <i>DvqHvlAmxF</i>. Another subscription bombing service improved this aspect and generated accounts with random plausible names like Daniel Williams, Michael Clark, Yoe Miller, or Hannah Fischer. Since the majority of the received emails were newsletters or account registrations, the subject lines commonly contained expected keywords like confirm, welcome, or email. Since the emails received are sent from genuine accounts on legitimate servers, the classic email header verification records like SPF (91.7% pass, 0.3% fail), DMARC (45.2% pass, 3.8% fail), and DKIM (74.3% pass, 12.3% fail) largely passed. Hence, they are not useful for filtering out these attacks. The failure rate is mainly due to badly configured email servers, which is common. Looking at the <i>spam confidence level</i> (SCL), where present, helps to assess if the email was classified as spam and would have been filtered out or would have ended up in the user&#8217;s inbox. SCL is a value from 1 to 9 assigned to a message that indicates the likelihood the message is spam. A value of 9 indicates an extremely high likelihood that the message is spam. A score of 1 is assigned to messages deemed not spam. In our dataset, more than 70% of the emails received a score of 1 and would have ended up in the user&#8217;s inbox if no additional email filtering was used. Around 28% were classified with score 5, which in a default Microsoft 365 configuration would be moved to the Junk folder.</p>
<p id="p-24">Only 34.4% of the emails contained an unsubscribe link in the header, and only 16.2% contained a one-click unsubscribe link following the header, standard for <i>list-unsubscribe=one-click</i>. This low number is mostly because we observed many account registration emails and password reset notifications, for which in general there is no unsubscribe link being sent. Furthermore, many providers follow the best practice of double opt-in, where the user has to confirm their email before further messages are sent. If we take these into account, then we see we are closer to a 52% rate of emails requiring an unsubscribe, of which 36% had an unsubscribe link. The most prominent types of abused websites are online shopping websites (25% of our dataset), financial (13% of our dataset), and technology/education (10% of our dataset). These types of sites commonly offer newsletter and account registrations, making them ideal targets. Slightly more than six percent (6.4%) of the emails contained some information about where to complain, such as the <i>X-complaints-to header</i> field, whereas 29.4% had an abuse header like X-report-abuse. Apart from removing the recipient from the mailing list if the user complains, the service provider cannot do much to prevent this.</p>
<p id="p-25">We conclude that for company Omega, the attackers knew the working hours of victim Alpha and Beta. They used a random subset of their mailing lists to flood them with thousands of emails per hour. More than 71% of the emails were not classified as spam by Microsoft 365 and would have been delivered to the user&#8217;s inbox if no additional filtering solutions would have been in place. Classic header checks (SPF, DMARC, and DKIM) as well as content and link analysis would not have helped prevent this attack. No malicious URLs or malware were involved at any stage. We suspect that the attackers had planned to follow-up with a fake support call to install ransomware, but did not succeed.</p>
</section>
<section id="sec5" class="sec">
<h2 class="heading">Goals of Subscription Bombing</h2>
<p id="p-26">There are multiple potential reasons for attackers to perform subscription bombing attacks against a given target. The most common reasons are: harassment or revenge, distraction from other attacks, pretext for follow-up attacks, or disruption of communication. Even if victims do not immediately recognize the attack as targeted, they quickly notice its impact through the rapidly growing inbox. The primary impact on the victim is the time required to clean up their inbox. Deleting thousands of unwanted emails is time-consuming, and bulk deletion risks removing legitimate messages. Since some of the spam consists of newsletters or subscription-based content, deletion alone is insufficient; more emails will continue to arrive unless the user unsubscribes. Some mailing lists generate high volumes of messages, potentially sending hundreds per day. The actual time spent depends on the email server in use, deployed security solutions, and the responsiveness of the IT team managing the incident. For illustration, assuming 1,000 spam emails, with 80% deleted in one second and 20% requiring 30 seconds to unsubscribe, cleanup would take approximately 113 minutes. Even at an average labor cost of $20/hour, the resulting loss is about $38—compared to the attacker’s estimated $1 cost to send the emails.</p>
<section id="sec6" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>Harassment.</strong>  Individuals with personal grievances may use subscription bombing as a form of digital harassment or retaliation against someone they feel has wronged them. For example, disgruntled employees might target their employer after being denied a desired salary increase. It could also occur in personal relationships, such as an ex-partner. Public figures, such as politicians or CEOs, may be targeted following unpopular decisions. For example, journalist Brian Krebs was targeted after some of his discovery articles in 2012 discussed specific cyber criminals.<a class="reference-link xref xref-bibr" href="#B7" data-jats-ref-type="bibr" data-jats-rid="B7"><sup>7</sup></a> Another example is the U.S. DOGE mail account, which apparently got flooded with spam in early 2025, after requesting status updates from government workers.<a class="reference-link xref xref-bibr" href="#B5" data-jats-ref-type="bibr" data-jats-rid="B5"><sup>5</sup></a> Such attacks are highly disruptive and frustrating for victims. Victims are often disappointed that built-in spam filters fail to protect them from such attacks.<a class="reference-link xref xref-bibr" href="#B12" data-jats-ref-type="bibr" data-jats-rid="B12"><sup>12</sup></a> Typically, the goal of harassment is not financially motivated, and the attacker does not directly witness the impact of the attack.</p>
</section>
<section id="sec7" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>Distraction.</strong>  Subscription bombing can serve as a smokescreen to distract from another attack in progress, such as a data breach or account takeover attempt. This is particularly effective when numerous password reset emails are triggered for various legitimate services—among the flood of notifications, a malicious password reset request or suspicious log-in attempt might go unnoticed, giving the attacker a window of opportunity to exploit it. For example, during a noisy credential-stuffing attack such a distraction can increase the chances of remaining unnoticed for a longer time period.</p>
<p id="p-29">Another way attackers profit through distraction is by making fraudulent purchases while burying e-commerce confirmation emails within the bulk of spam. In one case in 2024, a user discovered a purchase confirmation for a new Apple iPhone in the pile of unwanted emails. The attackers used the distraction to carry out a card-not-present fraudulent transaction.<a class="reference-link xref xref-bibr" href="#B15" data-jats-ref-type="bibr" data-jats-rid="B15"><sup>15</sup></a> They had somehow obtained access to the victim’s email and credit card details, possibly from a previous Magecart attack or data breach. Other users on Reddit report digital components being ordered from Walmart and flights being booked with their email address during a subscription bombing attack. The primary goal of subscription bombing in such cases is to conceal the real attack for as long as possible.</p>
</section>
<section id="sec8" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>Follow-up attack.</strong>  Instead of using subscription bombing solely as a distraction, attackers can also exploit it as a stepping stone for more targeted deception. There have been multiple reports of threat actors following up a subscription bombing campaign by impersonating the victim’s IT helpdesk. This impersonation typically takes place via alternative communication channels, such as Microsoft Teams calls or WhatsApp messages.</p>
<p id="p-31">Under the disguise of helping resolve the spam overload, attackers persuade the victim to initiate a remote access session, thereby unknowingly granting control to cyber criminals. Commonly used remote monitoring and management tools, such as AnyDesk, are often employed to facilitate this access. In 2024, Rapid7 reported several incidents in which the Black Basta ransomware group used this tactic to gain remote access and deploy additional malware.<a class="reference-link xref xref-bibr" href="#B4" data-jats-ref-type="bibr" data-jats-rid="B4"><sup>4</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B10" data-jats-ref-type="bibr" data-jats-rid="B10"><sup>10</sup></a> Leaked internal chat logs from the Black Basta ransomware group from February 2025 revealed this approach as a part of their playbook. One message read: “I will flood their inboxes with spam, and you will call them pretending to be an IT admin, saying they need to install a spam filter. She installs AnyDesk, and we get in to install our software.”<a class="reference-link xref xref-bibr" href="#B3" data-jats-ref-type="bibr" data-jats-rid="B3"><sup>3</sup></a> The primary objective is to create a scenario in which the victim is more likely to accept external assistance, ultimately leading to system compromise.</p>
</section>
<section id="sec9" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>Overloading.</strong>  Attackers may target organizations to overload their email servers, causing technical problems or outages within the network similar to a DDoS attack against the Web server. Depending on the configuration and setting of the target email server and inbox, there may be a threshold beyond which even legitimate emails are rejected by the target. This allows the attacker to temporarily block important emails from even reaching the inbox, until the resources have been freed and a retransmission is attempted.<a class="reference-link xref xref-bibr" href="#B2" data-jats-ref-type="bibr" data-jats-rid="B2"><sup>2</sup></a> In 2025, the hacktivist collective Anonymous launched attacks to disrupt the online activities of ISIS.<a class="reference-link xref xref-bibr" href="#B11" data-jats-ref-type="bibr" data-jats-rid="B11"><sup>11</sup></a> Among DDoS and defacement attacks this included email flooding. This could disrupt the normal work of the employees during critical time intervals and is frequently combined with an extortion attempt. The attackers ask for money to stop the attack. Similar to network DDoS attacks, the victim might give in and pay the blackmail.</p>
</section>
</section>
<section id="sec10" class="sec">
<h2 class="heading">Subscription Bombing as a Service</h2>
<p id="p-33">Various cyber criminals offer subscription bombing on the Dark Web and underground websites, often with a free trial. The criminals typically boast that they can flood the target inbox with 1,000–10,000 emails per hour with a spam detection rate of less than 0.1%. One service claims it can generate 6,000 emails per hour with more than 100,000 newsletters and forums in their database. Generating such a service is not very difficult on a technical level. Analyzing the leaked chat messages from the Black Basta ransomware group<a class="reference-link xref xref-bibr" href="#B3" data-jats-ref-type="bibr" data-jats-rid="B3"><sup>3</sup></a> shows they created their own subscription bombing service. Subscription bombing services are advertised using names like “emailbomb.cc”, “ebomber.cc”, “floodbox.org”, “diddy.pw”, “subscribethemall.com”, or “mailbait.info”. Note that attacking inboxes without consent is illegal in most jurisdictions.</p>
<p id="p-34">As a case study, we explore one representative subscription bombing service, accessible via a publicly available website on the regular Internet. The platform primarily offers subscription bombing, though SMS/text message bombing is also available. Attacks against a specific target can be scheduled for a duration ranging from a minimum of one hour to a maximum of 31 days. Emails can be sent using a &#8220;drip feed&#8221; model, distributing the load over time, with adjustable rates from 75 to 3,000 emails per hour. Pricing is straightforward: 1,000 emails can be sent for as little as $1 using the standard list. A high-quality list, marketed as having a higher chance of bypassing spam filters, is available for $10 per 1,000 emails. The service dashboard provides an overview of all scheduled tasks and their completion status. Communication is handled through a Telegram channel, and support is also available via an integrated AI chatbot (tawk.to). For advanced users, the provider also offers a REST API to automate attacks further. Payments are handled exclusively through cryptocurrency via a gateway provider (pay.crypto-gateway.xyz), which currently supports BTC, LTC, and XMR.</p>
<p id="p-35">In our own test using their free trial with the standard list, 21% (14/68) of the emails arrived in the inbox after 13 minutes. A second trial had 24% (929/3,930) of emails reaching the inbox. Note that attacking inboxes without consent is illegal in most jurisdictions. We only tested mailboxes directly under our control, on our servers.</p>
<p id="p-36">By extrapolating from the order numbers in the dashboard, it appears that at least 300,000 tasks have been executed to date, assuming the count started at zero. During a test window of 767 hours, the task counter increased by 17,323. That represents 22.6 campaigns per hour. The activity level indicates ongoing interest for such services.</p>
<p id="p-37"><a class="xref xref-table" href="#T2" data-jats-ref-type="table" data-jats-rid="T2">Table 2</a> shows prices for subscription bombing campaigns, taken from service offers across different underground websites. One of the scripts with unlimited access was available to buy for $500. The prices are very similar. The major variation is the quality of the list, which results in a higher inbox success rate, at a higher price tag.</p>
<figure id="T2" class="table-wrap" data-jats-position="float">
<div class="caption"><span class="caption-label">Table 2. </span> <span class="p">Prices to send N emails in a subscription bombing campaign.</span></div>
<div class="table-container">
<table class="table table-bordered table-condensed table-hover">
<colgroup>
<col />
<col />
<col />
<col />
<col /> </colgroup>
<thead>
<tr>
<th>Service</th>
<th>Service A</th>
<th>Service B</th>
<th>Service C</th>
<th>Service D</th>
</tr>
</thead>
<tbody>
<tr>
<td>Free trial</td>
<td>10 min</td>
<td>25 emails</td>
<td>25 emails</td>
<td>100 emails</td>
</tr>
<tr>
<td>5,000 emails</td>
<td>1 hour $9</td>
<td>$5</td>
<td>$50</td>
<td>$5</td>
</tr>
<tr>
<td>10,000 emails</td>
<td>1 week $65</td>
<td>$10</td>
<td>$100</td>
<td>$10</td>
</tr>
</tbody>
</table>
</div>
</figure>
</section>
<section id="sec11" class="sec">
<h2 class="heading">Mitigation</h2>
<p id="p-39">As subscription bombing is increasingly used to target individuals and organizations, defenders must apply layered countermeasures. To minimize impact, organizations should:</p>
<ol class="list" data-jats-list-type="order">
<li class="list-item">
<p id="p-40">Set thresholds or rate limits for the number of emails an inbox can receive from previously unseen senders.</p>
</li>
<li class="list-item">
<p id="p-41">Educate users about subscription bombing attacks.</p>
</li>
<li class="list-item">
<p id="p-42">Group similar newsletter or account reset emails during active attacks.</p>
</li>
<li class="list-item">
<p id="p-43">Automatically unsubscribe or bounce unwanted newsletter messages where supported.</p>
</li>
</ol>
<p id="p-44">Additionally, the receiving server may flag incoming subscription bombing attacks and alert the security team.</p>
<p id="p-45">While securing the account itself through two-factor authentication (2FA) or zero-trust access (ZTA) is recommended, these measures do not mitigate subscription bombing. All the attacker needs is the victim&#8217;s email address, which could come from a previous data breach, or in most cases has been identified from social media or because of a predictable naming schema such as firstname.lastname@company.tld. When interacting with diverse services on the Internet, the email address naturally “leaks.” Some users create individual email addresses when registering with online services by appending the plus sign (e.g., <i>yourname+newsletter@email.tld</i>). While this allows the user to pinpoint where the email address leaked, it does not protect against the attack. The same holds true for security awareness training through phishing simulations. While those are useful for security in general, they cannot mitigate subscription bombing attacks.</p>
<p id="p-46">A cleanup process based exclusively on keyword blocking is tricky, as this would also remove legitimate and wanted newsletters and emails. Moving all suspicious emails to a different folder for the user manually inspect them is not a viable option, as this would still require user time to sieve through thousands of emails.</p>
<section id="sec12" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>Newsletter providers.</strong>  As newsletter and service providers are being abused as amplifiers in these attacks, they are hopefully interested in hardening their services. Providers of newsletters can limit abuse of their services with the following measures. Newsletter subscription should require an email confirmation (double opt-in) and single-click unsubscribe. Having the user click on a link they receive to confirm subscription ensures the newsletter was delivered and received correctly. To prohibit automated registrations, the sign-up page should require a CAPTCHA, even if their benefits are decreasing with new AI tools. Another angle is to require users to register an account before being able to subscribe to slow down the registration process. Even though several of these recommendations add friction to the registration process, service providers should be encouraged to use them. Otherwise, service providers risk being added to blocklists, for example, when victims report the newsletters from attacks.</p>
</section>
<section id="sec13" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>Filtering and triaging.</strong>  The email server could build a list of known-good addresses at the server level, domain level, and account level. These lists can then be used to tag known-good emails during an ongoing wave to help users mitigate negative effects and allowing them to process the flood, if not automatically, asynchronously. Curating a trusted list by adding domains based on interaction is impractical, as users typically do not reply to newsletters, making it a one-way communication pattern. This leaves only an age score for newsletter domains from which the user has previously received emails.</p>
</section>
<section id="sec14" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>Unsubscribe.</strong>  Unsubscribing after a subscription bombing attack is complex. Deleting all emails within a given time frame may also remove legitimate, important messages. Furthermore, deleting alone does not solve all issues. Since many of these emails originate from newsletters, unsubscribing is necessary to prevent continued flooding. But this action is very time consuming if done manually. Some suggestions from the Internet, such as changing the email address, are impractical in most cases. Some newsletter services make unsubscribing unnecessarily complicated, such as requiring CAPTCHA solving, account creation, or asking about communication preferences and feedback. Also, some providers simply ignore changes made by the user and continue to send emails.</p>
<p id="p-50">These obstacles make automated unsubscription difficult and, in some cases, even risky. Unfortunately, this does only help with newsletters and not with other types such as password-reset emails which do not have unsubscribe options. As alluded to previously, there are many cases where the services used are non-English, making it harder for humans to find the unsubscribe option in the various languages. Depending on the email server&#8217;s capabilities, the administrator can set up blocks or bounce messages for involved emails. Unfortunately, this does not guarantee the sender will refrain from sending more emails in the future. GenAI can help with unsubscribing from such services by interacting with the page and finding the right option. In our tests with early agentic browsers, the success rate of AI-supported unsubscription was less than 50% but recent enhancements in AI agents show promise in improving the success rate of agentic unsubscribe tools. Further research in automatic unsubscribing is necessary, especially after detecting an attack.</p>
</section>
</section>
<section id="sec15" class="sec">
<h2 class="heading">Conclusion</h2>
<p id="p-51">Subscription bombing attacks are technically simple for attackers to conduct yet disproportionately difficult for victims to mitigate. Based on our analysis of 24 subscription bombing campaigns, this emerging attack vector poses a growing operational and security risk to digital communication. The attack itself provides only limited use cases for direct fraud but can be used as a distraction to hide concurrent attacks. Attackers leverage subscription bombing for various malicious purposes, ranging from creating nuisance to obscuring critical security alerts and facilitating social engineering attacks. Our study provides valuable insights into the operational patterns, timelines, and Dark Web services associated with these campaigns.</p>
<p id="p-52">While complete mitigation remains challenging due to the asynchronous distribution between attackers and defenders, automated unsubscription offers a promising avenue for reducing the impact on victims. Legislative options such as requiring unsubscribe links are another option to reduce the impact of subscription bombing. Further research is crucial to develop more robust defense mechanisms against this evolving form of digital attack.</p>
</section>
<section id="sec16" class="sec">
<h2 class="heading">Acknowledgments</h2>
<p id="p-53">We thank Max Fischer for data collection and members of xorlab and HexHive for feedback. This work was supported in part by the ERC (grant No. 850868) and the SNSF (grants PCEGP2 186974 and 200021-236559).</p>
</section>
</div>
<footer class="back"></footer>
</article>
]]></content:encoded>
					
					<wfw:commentRss>https://cacm.acm.org/practice/subscription-bombing-email-under-attack/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
		      <dc:creator><![CDATA[Candid Wüest]]></dc:creator>
<post-id xmlns="com-wordpress:feed-additions:1">779431</post-id>	</item>
		<item>
		<title>Evaluating General-Purpose AI with Psychometrics</title>
		<link>https://cacm.acm.org/research/evaluating-general-purpose-ai-with-psychometrics/</link>
					<comments>https://cacm.acm.org/research/evaluating-general-purpose-ai-with-psychometrics/#comments</comments>
		
		<dc:creator><![CDATA[Xiting Wang, Liming Jiang, José Hernández-Orallo, David Stillwell, Shiqiang Chen, Luning Sun, Fang Luo, and Xing Xie]]></dc:creator>
		<pubDate>Tue, 14 Apr 2026 15:41:01 +0000</pubDate>
				<category><![CDATA[Artificial Intelligence and Machine Learning]]></category>
		<category><![CDATA[Theory]]></category>
		<guid isPermaLink="false">https://cacm.acm.org/?post_type=digital-library&#038;p=779335</guid>

					<description><![CDATA[<p>How construct-oriented evaluation can be established for general-purpose AI systems through learning from psychometrics.</p>]]></description>
										<content:encoded><![CDATA[<article>
<div class="body" lang="en">
<section id="sec1" class="sec">
<p id="p-1">In the rapidly evolving field of artificial intelligence (AI), rigorous evaluation is critical. Evaluation helps prevent disastrous outcomes in high-stakes applications such as autonomous driving and medical diagnosis,<a class="reference-link xref xref-bibr" href="#B7" data-jats-ref-type="bibr" data-jats-rid="B7"><sup>7</sup></a> and helps stop the spread of racist, sexist, ableist, extremist, and other harmful ideologies.<a class="reference-link xref xref-bibr" href="#B4" data-jats-ref-type="bibr" data-jats-rid="B4"><sup>4</sup></a> It also plays a key role in understanding how intelligent AI is, preventing the misallocation of resources, and guiding improvements in future model development.</p>
<p id="p-2">Despite its importance, rigorous evaluation of versatile general-purpose AI systems such as ChatGPT and Gemini is challenging. The versatility of these systems, which can handle diverse tasks specified through text, audio, images, or video, makes it difficult to anticipate where users will apply them, rendering it impossible to test every potential task to ensure the systems function as intended. The core reason behind this challenge is the transition of AI systems from being designed for specific, well-defined objectives (tasks) to handling complex, dynamic scenarios where tasks are not predefined. A methodology to predict and explain AI performance on these diverse, potentially unseen tasks has not yet been established.</p>
<aside class="boxed-text">
<div class="article-key-insights">
<h2>Key Insights</h2>
<ul class="list" data-jats-list-type="bullet">
<li class="list-item">
<p id="p-3">General-purpose AI systems present challenges for task-oriented evaluation, which fails to predict performance on unforeseen tasks and explain the variability of results, and lacks reliability and validity.</p>
</li>
<li class="list-item">
<p id="p-4">Leveraging psychometrics, construct-oriented evaluation identifies and measures constructs underlying AI performance, offering predictive power, explanatory power, and quality assurance.</p>
</li>
<li class="list-item">
<p id="p-5">We propose a psychometric framework for general-purpose AI evaluation consisting of three stages, including construct identification, construct measurement, and test validation.</p>
</li>
<li class="list-item">
<p id="p-6">We introduce an operational evaluation pipeline supported by psychometrics to provide practical guidance and transform future AI development.</p>
</li>
</ul>
</div>
</aside>
<p id="p-7">To tackle this challenge, we suggest transitioning from <i>task-oriented</i> evaluation to <i>construct-oriented</i> evaluation, taking inspiration from a methodology developed to evaluate humans. Constructs are concepts that, although they may not be observed directly, are hypothesized to underlie a range of behaviors.<a class="reference-link xref xref-bibr" href="#B12" data-jats-ref-type="bibr" data-jats-rid="B12"><sup>12</sup></a> We follow the common assumption in psychology and behavioral science<a class="reference-link xref xref-bibr" href="#B3" data-jats-ref-type="bibr" data-jats-rid="B3"><sup>3</sup></a> that human behaviors can be explained by a limited set of constructs, such as cognitive abilities and personality traits.<a class="footnote-link xref xref-fn" href="#FN1" data-jats-rid="FN1" data-jats-ref-type="fn"><sup>a</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B36" data-jats-ref-type="bibr" data-jats-rid="B36"><sup>36</sup></a> In the same way, the wide range of outputs from an AI system can be organized and explained by a relatively small number of constructs. These constructs then become the fundamental building blocks of hypotheses and theories,<a class="reference-link xref xref-bibr" href="#B3" data-jats-ref-type="bibr" data-jats-rid="B3"><sup>3</sup></a> allowing us to accurately predict and explain the performance of general-purpose AI systems.<a class="footnote-link xref xref-fn" href="#FN2" data-jats-rid="FN2" data-jats-ref-type="fn"><sup>b</sup></a></p>
<p id="p-8">In this article, we demonstrate how construct-oriented evaluation can be established for general-purpose AI systems through learning from <i>psychometrics</i>,<a class="reference-link xref xref-bibr" href="#B36" data-jats-ref-type="bibr" data-jats-rid="B36"><sup>36</sup></a> which for over a century has been applied to evaluate human psychological constructs. We first introduce three major advantages of adapting psychometrics for AI evaluation, illustrating them with concrete examples. We then discuss ways in which we can distinguish sound applications of psychometrics from oversimplified ones, warning against common pitfalls. With this foundation, we present a systematic framework grounded in psychometric principles and introduce an operational evaluation pipeline to provide practical guidance. Finally, we identify underexplored avenues and societal implications that open new research directions.</p>
</section>
<section id="sec2" class="sec">
<h2 class="heading">Advantages of Construct-Oriented AI Evaluation</h2>
<p id="p-9">As shown in <a class="xref xref-fig" href="#F1" data-jats-ref-type="fig" data-jats-rid="F1">Figure 1</a>, compared with task-oriented evaluation, which employs benchmarks consisting of disparate tasks, construct-oriented evaluation excels in predictive power, explanatory power, and quality assurance, which can be achieved by optimizing the use of existing benchmarks and developing new tests for AI.</p>
<figure id="F1" class="fig" data-jats-position="float">
<div class="image-container"><img decoding="async" class="graphic" title="Figure 1. " src="https://cacm.acm.org/wp-content/uploads/2026/03/3769688_fig01.jpg" alt="" data-image-id="F1" data-image-type="figure" /></div><figcaption><span class="caption-label">Figure 1. </span> <span class="p">Advantages of construct-oriented evaluation and related examples.<a class="reference-link xref xref-bibr" href="#B5" data-jats-ref-type="bibr" data-jats-rid="B5"><sup>5</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B6" data-jats-ref-type="bibr" data-jats-rid="B6"><sup>6</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B25" data-jats-ref-type="bibr" data-jats-rid="B25"><sup>25</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B35" data-jats-ref-type="bibr" data-jats-rid="B35"><sup>35</sup></a></span></p>
<div class="figcaption-footer"> </div>
</figcaption></figure>
<section id="sec3" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>Predictive power.</strong>  Following the task-oriented paradigm, evaluation results cannot be extrapolated to unforeseen tasks. For instance, although a general-purpose AI medical assistant could pass the BIG-bench benchmark, which contains tasks related to biology or medicine, it is still difficult to understand its performance on tasks not seen during testing. The assumption that the performance of AI tested on a limited number of tasks directly reflects its performance in a practically infinite range of applicable tasks is unsubstantiated.<a class="reference-link xref xref-bibr" href="#B17" data-jats-ref-type="bibr" data-jats-rid="B17"><sup>17</sup></a></p>
<p id="p-12">In comparison, constructs have long been demonstrated to be predictive of humans’ future performance and real-world outcomes. For example, academic grades can be predicted from constructs such as cognitive ability, self-esteem, hope, and attributional style.<a class="reference-link xref xref-bibr" href="#B23" data-jats-ref-type="bibr" data-jats-rid="B23"><sup>23</sup></a> Constructs like personality predict individual outcomes such as physical and mental health, interpersonal outcomes such as the quality of family relationships, and social institutional outcomes such as job satisfaction and criminal activity.<a class="reference-link xref xref-bibr" href="#B31" data-jats-ref-type="bibr" data-jats-rid="B31"><sup>31</sup></a></p>
<p id="p-13">Constructs can also predict the performance of AI systems across multiple tasks and scenarios. For example, the spatial reasoning construct is required in a wide range of tasks involving planning and navigation, visual image interpretation, and natural language understanding.<a class="reference-link xref xref-bibr" href="#B9" data-jats-ref-type="bibr" data-jats-rid="B9"><sup>9</sup></a> Constructs like these that underlie multiple tasks can offer a more holistic understanding of AI models and maintain their relevance over extended periods, which can help predict how those models perform on unforeseen tasks and how their performance varies with scale.</p>
<p id="p-14">Recent empirical works provide some initial evidence that constructs underlying a range of model behaviors may help predict performance on unseen tasks or model scales. For example, Burden et al.<a class="reference-link xref xref-bibr" href="#B5" data-jats-ref-type="bibr" data-jats-rid="B5"><sup>5</sup></a> showed that two constructs, navigation ability and visual ability<i>,</i> were able to predict model performance on a wide range of tasks in embodied 3D scenarios, including unseen tasks (i.e., 20% held-out task instances), in the Animal-AI Olympics (a testing platform where AI agents are evaluated on their performance in locating a reward in a 3D environment). As shown in <a class="xref xref-fig" href="#F1" data-jats-ref-type="fig" data-jats-rid="F1">Figure 1a</a>, compared with the aggregation method that simply predicts the performance of an agent by considering its overall performance on seen tasks, a model that predicts an agent’s performance by considering its navigation and visual abilities is consistently more accurate, with success rates in the middle range—exactly where there is variability to explain.</p>
<p id="p-15">Another example can be found in Ruan et al.,<a class="reference-link xref xref-bibr" href="#B35" data-jats-ref-type="bibr" data-jats-rid="B35"><sup>35</sup></a> who conducted principal component analysis to extract constructs that contributed to model performance on a wide range of tasks. Based on these constructs, they managed to extrapolate from smaller-scale, less-capable models to larger-scale, more-powerful models and predict their performance on four BIG-bench tasks (<a class="xref xref-fig" href="#F1" data-jats-ref-type="fig" data-jats-rid="F1">Figure 1b</a>). Although these examples employ different methods to derive constructs (discussed in more detail later), they both show that constructs help predict model performance on unseen tasks, providing empirical evidence for the predictive power of construct-oriented evaluation.</p>
</section>
<section id="sec4" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>Explanatory power.</strong>  Central to identifying the key strengths and weaknesses of AI systems is explaining why they perform well or fail at certain tasks. Current benchmarks, however, are not capable of providing such explanations, hindering researchers and policymakers in making informed decisions, for instance, about whether or when a system is safe and how it might be improved.<a class="reference-link xref xref-bibr" href="#B7" data-jats-ref-type="bibr" data-jats-rid="B7"><sup>7</sup></a></p>
<p id="p-17">In contrast, psychometric tests revolve around constructs, which are conceptualized to explain empirical observations. For example, according to the Cattell-Horn-Carroll theory, many of the individual differences in cognitive tasks can be explained by a relatively small number of cognitive abilities, such as fluid and crystallized intelligence.<a class="reference-link xref xref-bibr" href="#B8" data-jats-ref-type="bibr" data-jats-rid="B8"><sup>8</sup></a> These constructs form an underlying structure, which equips the psychometric test with explanatory power.</p>
<p id="p-18">Similarly, constructs are also useful in explaining the performance of AI systems. As shown in <a class="xref xref-fig" href="#F1" data-jats-ref-type="fig" data-jats-rid="F1">Figure 1c</a>, Burnell et al.<a class="reference-link xref xref-bibr" href="#B6" data-jats-ref-type="bibr" data-jats-rid="B6"><sup>6</sup></a> carried out factor analysis on the performance of 29 large language models (LLMs) in 27 cognitive tasks, which revealed three constructs representing reasoning, comprehension, and core language modeling that explained a high proportion of the variance in model performance. Ilić and Gignac ran a similar analysis<a class="reference-link xref xref-bibr" href="#B18" data-jats-ref-type="bibr" data-jats-rid="B18"><sup>18</sup></a> involving 591 LLMs, reporting interrelated cognitive-like capabilities in the models. These empirical findings showcase the explanatory power of construct-oriented evaluation, which casts light on the nuanced patterns of capabilities in AI models.</p>
<p id="p-19">Compared with AI explanation and interpretation methods, such as probing, which provide valuable insights into the internal mechanisms of AI systems,<a class="reference-link xref xref-bibr" href="#B43" data-jats-ref-type="bibr" data-jats-rid="B43"><sup>43</sup></a> construct-oriented evaluation offers an alternative framework for explaining the overall performance of a complex model by considering its underlying capabilities. These approaches complement each other, mirroring cognitive science research where both neuroscientific approaches and psychometric methods contribute to understanding human cognition. The former studies the relationship among brain regions and specific mechanisms (analogous to probing), while the latter measures complex capabilities such as cognitive abilities (analogous to constructs). In both cases, we could predict behavior with a large number of features or a complex mechanism analogous to what the AI system does (e.g., using mechanistic interpretability), but the relevance of constructs lies in the power of abstraction of a small number of features or traits (capabilities, propensities, etc.) that have predictive and explanatory power. This search for more high-level concepts and different granularities is similar to the goals of other sciences, such as sociology.</p>
</section>
<section id="sec5" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>Quality assurance.</strong>  Researchers have raised serious concerns about the reliability and validity of task-oriented AI evaluation. For instance, as frequently reported, the performance of AI systems is affected by factors such as input prompts and specific configurations.<a class="reference-link xref xref-bibr" href="#B19" data-jats-ref-type="bibr" data-jats-rid="B19"><sup>19</sup></a> The issue of sensitivity casts doubts on benchmarks as a reliable measure, and questions the consistency of AI systems&#8217; performance in the real world. Many benchmarks have been found to be unstable, undermining the fairness and authenticity of assessment results.<a class="reference-link xref xref-bibr" href="#B29" data-jats-ref-type="bibr" data-jats-rid="B29"><sup>29</sup></a> Moreover, it is hard to ascertain what is (not) being measured by a benchmark. For example, an AI system&#8217;s high success rate on benchmark tasks may not translate into high performance in real-world applications.<a class="reference-link xref xref-bibr" href="#B27" data-jats-ref-type="bibr" data-jats-rid="B27"><sup>27</sup></a> This disparity can be attributed to differences in task complexity between benchmarks and real-world applications, overspecialization, or biases.<a class="reference-link xref xref-bibr" href="#B41" data-jats-ref-type="bibr" data-jats-rid="B41"><sup>41</sup></a></p>
<p id="p-21">Psychometrics has developed a systematic approach to test quality assurance, focusing on both the reliability and validity of measurement.<a class="reference-link xref xref-bibr" href="#B36" data-jats-ref-type="bibr" data-jats-rid="B36"><sup>36</sup></a> Reliability ensures stable, consistent results across multiple measurements. It goes beyond the scope of “replicability” or “robustness” in computer science, offering a more comprehensive examination of different types of measurement errors, such as the inconsistency of results among different test items, raters, and (sub-)datasets. Validity indicates the extent to which a test measures what it is designed to measure. For example, a valid numerical reasoning test should tap into numerical reasoning ability rather than irrelevant constructs such as language proficiency. Only with reliable and valid tests can we place confidence in the measurement results and derive meaningful interpretations.</p>
<p id="p-22">A few recent studies have employed measures of reliability and validity to assess the quality of evaluation criteria for LLMs, advocating the use of reliability and validity in AI evaluation quality assurance.<a class="reference-link xref xref-bibr" href="#B42" data-jats-ref-type="bibr" data-jats-rid="B42"><sup>42</sup></a> For example, a wide range of reliability and validity indicators were used to assess measures for the constructs of faithfulness and readability.<a class="reference-link xref xref-bibr" href="#B25" data-jats-ref-type="bibr" data-jats-rid="B25"><sup>25</sup></a> As shown in <a class="xref xref-fig" href="#F1" data-jats-ref-type="fig" data-jats-rid="F1">Figure 1d</a>, four measures (e.g., LLM-score, which evaluates whether an explanation is easy to read for humans) appeared to have low reliability. It was also reported that LLM-based scores, in comparison to coherence-based scores, exhibited a lack of concurrent validity. Such empirical evidence could offer practical guidance on the choice of evaluation measures and ensure objective, rigorous, and accurate evaluation of AI systems.</p>
</section>
</section>
<section id="sec6" class="sec">
<h2 class="heading">Leveraging Psychometrics for AI Evaluation</h2>
<p id="p-23">Psychometrics provides an effective evaluation framework, but its integration into AI evaluation should involve more than just applying existing psychometric tests. Key considerations must be addressed regarding how the constructs are conceptualized, how the tests are developed or adapted, and what psychometric techniques are available and appropriate. The ensuing discussion will elucidate related strategies and raise open questions that warrant joint efforts from the wider scientific community.</p>
<section id="sec7" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>Pitfalls of oversimplified application of psychometrics.</strong>  There are several ways in which psychometrics can be improperly applied to AI system evaluation. First, some recent works treat AI systems as participants in human psychology experiments or apply existing psychometric tests to evaluate the general intelligence,<a class="reference-link xref xref-bibr" href="#B11" data-jats-ref-type="bibr" data-jats-rid="B11"><sup>11</sup></a> theory of mind, and personality<a class="reference-link xref xref-bibr" href="#B32" data-jats-ref-type="bibr" data-jats-rid="B32"><sup>32</sup></a> of AI systems. These works assume AI systems exhibit human-like traits, as they are trained on human data.<a class="reference-link xref xref-bibr" href="#B32" data-jats-ref-type="bibr" data-jats-rid="B32"><sup>32</sup></a> However, this assumption may result in misleading conclusions, because it is unclear whether human constructs are represented in AI systems and it is problematic to assume that the relationship between these constructs and their indicators identified in humans remains intact for AI systems. For instance, processing speed is considered an important indicator of intelligence in humans, but for AI systems, processing speed might not be relevant to intelligence. Smaller models with fewer parameters typically process faster than larger models, but this does not necessarily mean that smaller models are more intelligent. Therefore, when integrating psychometrics into the evaluation of AI systems, it is vital to consider the assumptions behind each psychometric test to determine its applicability.<a class="reference-link xref xref-bibr" href="#B32" data-jats-ref-type="bibr" data-jats-rid="B32"><sup>32</sup></a> In addition to adapting human tests, it is necessary to develop tests tailored to the constructs inherent in AI systems, leveraging psychometric principles.</p>
<p id="p-25">Second, simplistic application of psychometric techniques may overlook the need for quality assurance. Tests developed for humans may not be reliable or valid for AI systems, even if the systems are trained on large-scale human data. For example, several studies have applied self-report personality questionnaires originally developed for humans to LLMs.<a class="reference-link xref xref-bibr" href="#B26" data-jats-ref-type="bibr" data-jats-rid="B26"><sup>26</sup></a> No matter how the questions are rephrased or modified, as long as no actual semantic changes are introduced, human responses would remain consistent, as they tend to reflect underlying personality traits. However, when applied to AI systems, minor input changes (e.g., reversing the question order) can cause substantial changes in responses.<a class="reference-link xref xref-bibr" href="#B26" data-jats-ref-type="bibr" data-jats-rid="B26"><sup>26</sup></a> This raises doubts as to whether AI responses to self-report personality questionnaires reflect true understanding or are merely probabilistic selections based on the training data distribution.</p>
<p id="p-26">While some early studies have shown the relevance of constructs to AI systems, they remain isolated and rudimentary due to a lack of systematic psychometric grounding. For example, they may suffer from random or systematic errors, since they have not rigorously ensured reliability (eliminating random errors) and validity (eliminating systematic errors). In contrast, psychometrics has a century-long history of studying construct identification and measurement, and provides systematic methods for ensuring reliability and validity. To enhance understanding of psychometrics and how it can be more readily applied in AI evaluation, in the following sections we introduce a systematic framework with advanced psychometric technologies for AI evaluation, discuss open questions, and present future research opportunities.</p>
</section>
<section id="sec8" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>Psychometric framework: Key considerations.</strong>  We suggest a psychometric framework for construct-oriented evaluation of AI systems. As shown in <a class="xref xref-fig" href="#F2" data-jats-ref-type="fig" data-jats-rid="F2">Figure 2</a>, our framework includes three stages: construct identification, construct measurement, and test validation. To facilitate understanding and practical application of the framework, for each stage, we list possible approaches, example techniques, and further notes for consideration.</p>
<figure id="F2" class="fig" data-jats-position="float">
<div class="image-container"><img decoding="async" class="graphic" title="Figure 2. " src="https://cacm.acm.org/wp-content/uploads/2026/03/3769688_fig02.jpg" alt="" data-image-id="F2" data-image-type="figure" /></div><figcaption><span class="caption-label">Figure 2. </span> <span class="p">A framework for construct-oriented evaluation grounded in psychometrics, illustrated by goals, approaches and references, example techniques, and notes for consideration at each stage.</span></p>
<div class="figcaption-footer"> </div>
</figcaption></figure>
<section id="sec9" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><em>Construct identification. </em> In the first stage, we identify the constructs that can explain and predict AI performance. In psychometrics, there are two approaches: top-down and bottom-up.</p>
<p id="p-30">The <i>top-down</i> approach starts from a predefined theoretical framework, often derived from observations, expert consultation, and discussion. For instance, the construct of psychological resilience was identified through empirical observations that some children, despite facing adversities such as poverty, still demonstrated good adaptation and development.<a class="reference-link xref xref-bibr" href="#B28" data-jats-ref-type="bibr" data-jats-rid="B28"><sup>28</sup></a></p>
<p id="p-31">We can adopt similar approaches to identify constructs that are predictive and explanatory of AI system performance. For example, Li et al.<a class="reference-link xref xref-bibr" href="#B24" data-jats-ref-type="bibr" data-jats-rid="B24"><sup>24</sup></a> have noted that AI systems often generate incorrect, fabricated, or misleading information, referred to as <i>hallucination</i>. A similar concept from psychology, <i>confabulation</i>, could help explain this behavior in AI systems. To reach a consensus on an operationalized definition of confabulation in AI systems, we can incorporate observations with the expertise of domain experts. During this process, we can use techniques such as the Delphi method<a class="reference-link xref xref-bibr" href="#B40" data-jats-ref-type="bibr" data-jats-rid="B40"><sup>40</sup></a> to collect expert opinion in a systematic manner and reach consensus on the definition of AI constructs. The construct then needs to be validated to ensure it is predictive and explanatory.</p>
<p id="p-32">One feasible top-down approach is drawing an analogy between humans and AI systems. However, we should note that the definition of a construct may differ between AI systems and humans, despite the fact that they may be labeled the same. Suppose we are considering adapting the construct <i>emotional intelligence</i> for AI systems. As we check its definition, we may notice that descriptions about self-awareness of emotions are not appropriate for AI, or hold different meanings; hence, they should be removed or modified. The significance of a construct may also differ between humans and AI systems. For example, while confabulation is common in AI, it may appear in only a small group of humans with neurological conditions such as Alzheimer’s disease or Wernicke-Korsakoff Syndrome.<a class="reference-link xref xref-bibr" href="#B22" data-jats-ref-type="bibr" data-jats-rid="B22"><sup>22</sup></a></p>
<p id="p-33">The <i>bottom-up</i> approach, rather than relying on a predefined theoretical framework, originates from empirical data and seeks patterns that suggest the existence of a construct. A classic example is the development of the Big Five Factors of personality, where psychologists carried out factor analysis on a vast array of personality-descriptive terms, extracting five constructs that explain a large proportion of the variance in human behaviors.<a class="reference-link xref xref-bibr" href="#B14" data-jats-ref-type="bibr" data-jats-rid="B14"><sup>14</sup></a></p>
<p id="p-34">A similar approach has already been applied within the field of AI research. Work by Burnell et al.<a class="reference-link xref xref-bibr" href="#B6" data-jats-ref-type="bibr" data-jats-rid="B6"><sup>6</sup></a> applied factor analysis on data from 29 different LLMs across 27 cognitive tasks, finding that model capabilities could be explained by three well-delineated factors representing reasoning, comprehension, and core language modeling.</p>
<p id="p-35">It is worth mentioning that bottom-up and top-down approaches could be used in conjunction, complementing each other to provide a more comprehensive understanding of constructs. These approaches could also be applicable to a wider range of scenarios, such as the evaluation of multimodal LLMs and agentic AI systems. Here, not only may hierarchical structures appear that require more advanced psychometric techniques, but also the range of constructs should go well beyond capabilities to also consider other traits such as self-control, values, biases, and toxicity—all of which are critical for ensuring the safety and responsible deployment of AI systems.</p>
<p id="p-36">Notably, the conceptualization of constructs may evolve over time. For instance, self-esteem was initially viewed as single-dimensional but was then recognized to have multiple facets, including state self-esteem (self-worth in specific situations) and trait self-esteem (consistent or long-lasting self-worth).<a class="reference-link xref xref-bibr" href="#B21" data-jats-ref-type="bibr" data-jats-rid="B21"><sup>21</sup></a> We expect a similar evolution of construct understanding in AI evaluation. Definitions of newly identified constructs in AI are likely to be partial or biased initially and may be subject to later refinement with enriched empirical evidence. This underscores the need for ongoing investigation in future research, adapting existing tools from psychometrics.</p>
<p id="p-37">It is also important to note that when new constructs are identified for AI systems, they need to be validated over time to demonstrate their predictive and explanatory power before being accepted by the scientific community and public discourse. This is exactly what happened with self-esteem.</p>
</section>
<section id="sec10" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><em>Construct measurement.</em>  After a construct is identified for AI, a test is required to measure it. To develop a test for an AI construct, we can draw on the process by which psychometricians design tests for humans. It involves developing test items and establishing scoring criteria.</p>
<p id="p-39"><i>Test items</i> should sufficiently reflect the targeted construct and remain unaffected by unrelated constructs. They are usually designed according to test guidelines and item specifications.<a class="reference-link xref xref-bibr" href="#B8" data-jats-ref-type="bibr" data-jats-rid="B8"><sup>8</sup></a> Test guidelines specify the proportion of items measuring different aspects of the construct, the item format, the test length, time limits, and so on. Item specifications provide detailed designs for individual items by specifying the source of item content, the context, stimuli, and the difficulty level of an item.</p>
<p id="p-40">Test items for AI systems can take diverse formats, as used in psychometrics. For example, psychometric tests use highly abstract symbols or graphics, such as the abstraction and reasoning corpus in BIG-bench and those used in Raven’s Progressive Matrices,<a class="reference-link xref xref-bibr" href="#B33" data-jats-ref-type="bibr" data-jats-rid="B33"><sup>33</sup></a> to eliminate the interference of unintended constructs in real-world scenarios. When applied to AI evaluation, this could help minimize the unintended influence of random factors and enhance test reliability and validity. For instance, to assess inductive reasoning ability, we can develop abstract items that ask test takers to uncover the pattern in a sequence of numbers and generate the number that continues the sequence. Another potential format is simulation-based assessment,<a class="reference-link xref xref-bibr" href="#B30" data-jats-ref-type="bibr" data-jats-rid="B30"><sup>30</sup></a> which aligns more closely with some current AI benchmarks. These test items employ real-world situations to increase ecological validity while leveraging techniques such as evidence-centered design<a class="reference-link xref xref-bibr" href="#B30" data-jats-ref-type="bibr" data-jats-rid="B30"><sup>30</sup></a> to ensure the target construct is measured effectively.</p>
<p id="p-41">The choice of test-item format should align with the application scenario. For instance, simulation-based test items are useful when leveraging existing AI benchmarks to reduce the effort required for test-item development or when evaluating model performance in specific real-world scenarios.</p>
<p id="p-42">After the test items are developed, a <i>scoring</i> scheme based on appropriate measurement theories should be established to quantify the level of the construct given the obtained responses. Among existing scoring theories, item response theory (IRT)<a class="reference-link xref xref-bibr" href="#B12" data-jats-ref-type="bibr" data-jats-rid="B12"><sup>12</sup></a> appears particularly useful for AI evaluation.<a class="reference-link xref xref-bibr" href="#B44" data-jats-ref-type="bibr" data-jats-rid="B44"><sup>44</sup></a> IRT effectively predicts whether a test taker will get any item, including unseen ones, correct or not by modeling the probability of a correct response as a function of item parameters and the test taker’s level of the construct. IRT also enables comparisons of AI systems, even when they are evaluated using different tests with varying test formats, item sets, or difficulty levels because it allows the construct and item parameters to be estimated on a unified scale. This is very beneficial, particularly considering that tests for AI need to be frequently updated to cover potential exposure of items in the training data. Furthermore, IRT lays the groundwork for computerized adaptive testing,<a class="reference-link xref xref-bibr" href="#B44" data-jats-ref-type="bibr" data-jats-rid="B44"><sup>44</sup></a> which optimally selects items for each test taker, resulting in efficient and precise assessment.</p>
<p id="p-43">In comparison with the classic scoring method that simply aggregates the scores for each item, advanced techniques such as IRT provide more fine-grained, accurate, and comprehensive AI evaluation. Given these advantages, some studies have already adopted IRT. For example, Zhuang et al.<a class="reference-link xref xref-bibr" href="#B44" data-jats-ref-type="bibr" data-jats-rid="B44"><sup>44</sup></a> employed this technique to assess LLMs’ subject knowledge, mathematical reasoning, and programming skills. We expect more insights to be gained from state-of-the-art IRT models—for example, IRT-based cognitive diagnostic models that aim at pinpointing specific facets of strengths or weaknesses in a construct,<a class="reference-link xref xref-bibr" href="#B10" data-jats-ref-type="bibr" data-jats-rid="B10"><sup>10</sup></a> and IRT-based latent class models that combine the strengths of IRT with the idea of identifying the hidden groups that test takers belong to.<a class="reference-link xref xref-bibr" href="#B34" data-jats-ref-type="bibr" data-jats-rid="B34"><sup>34</sup></a></p>
</section>
<section id="sec11" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><em>Test validation.</em>  Imagine we have created a test to measure a certain construct. Before drawing conclusions based on the test results, it is necessary to verify the test’s quality. For instance, how much error is in the measurement? How do the test results relate to real-world behaviors? For this purpose, psychometrics provides a systematic methodology concentrating on indicators of reliability and validity in the measurement.</p>
<p id="p-45"><i>Reliability</i> refers to the consistency or stability of a test.<a class="reference-link xref xref-bibr" href="#B36" data-jats-ref-type="bibr" data-jats-rid="B36"><sup>36</sup></a> Psychometrics introduces various reliability indicators. For instance, test-retest reliability involves giving the same test to the same group of test takers multiple times to evaluate the stability of the measurements. Practically, one could test an AI system multiple times to obtain performance stability (replicability). Internal consistency reliability measures consistency across items within a test. High internal consistency reliability indicates that an AI system performs similarly across all test items. Other types of reliability that might be adapted into AI evaluation include parallel-form reliability, which ensures that two different but equivalent versions of a test measure the same construct consistently, and inter-rater reliability, which assesses the degree of agreement among different raters and has been adopted for evaluating text-generation models.<a class="reference-link xref xref-bibr" href="#B25" data-jats-ref-type="bibr" data-jats-rid="B25"><sup>25</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B42" data-jats-ref-type="bibr" data-jats-rid="B42"><sup>42</sup></a></p>
<p id="p-46"><i>Validity</i> indicates the extent to which a test measures the construct it claims to measure.<a class="reference-link xref xref-bibr" href="#B36" data-jats-ref-type="bibr" data-jats-rid="B36"><sup>36</sup></a> There are different validity indicators in psychometrics that may be applied in AI. For example, construct validity indicates how well the test reflects the construct. A commonly used method to evaluate construct validity is factor analysis, which conceptualizes the constructs as latent factors underlying test items. By examining the fit of the specified factorial structure with empirical data, we can quantitatively understand the construct validity of the test. Construct validity can also be evaluated via convergent validity, reflecting the empirical relation of theoretically related measures, and discriminant validity, indicating whether measurements that are not supposed to be related are, in fact, unrelated empirically. Predictive validity is another form of validity that is concerned with the extent to which a score on a test predicts performance on a certain criterion measure, which can be real-life performance or future outcomes. For example, AI systems with stronger emotional understanding should receive better user reviews.</p>
<p id="p-47">Some recent AI evaluation studies have carried out reliability and validity analyses. For example, Li et al.<a class="reference-link xref xref-bibr" href="#B25" data-jats-ref-type="bibr" data-jats-rid="B25"><sup>25</sup></a> employed test-retest reliability that examined score stability across different time points to compare measures of readability. They also assessed convergent and divergent validity to verify that measures of readability and faithfulness were tapping into the intended constructs. METRICEVAL,<a class="reference-link xref xref-bibr" href="#B42" data-jats-ref-type="bibr" data-jats-rid="B42"><sup>42</sup></a> a framework that examines both concurrent validity and construct validity, has been proposed for conceptualizing and evaluating natural language generation metrics. This framework has been applied to the evaluation of summarization metrics.</p>
<p id="p-48">The above-mentioned studies have demonstrated the effectiveness of reliability and validity analyses in a few specific areas (e.g., summarization, faithfulness, and readability); however, many important areas—such as evaluating the coding skills, problem-solving abilities, and creativity of AI systems—remain largely unexplored. We advocate broader and more systematic adoption of these measures across all AI evaluation settings. Only with adequate reliability and validity can we trust the test results and answer questions that are important in real life, such as: Does a higher problem-solving test score from GPT-4 than from GPT-3 mean increased user satisfaction when interacting in online research?</p>
<p id="p-49">However, not all psychometric indices are universally applicable. Depending on the specific scenario, different reliability and validity indices should be employed. For instance, inter-rater reliability is particularly useful when human raters are involved, but is not applicable to scenarios without raters.<a class="reference-link xref xref-bibr" href="#B25" data-jats-ref-type="bibr" data-jats-rid="B25"><sup>25</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B42" data-jats-ref-type="bibr" data-jats-rid="B42"><sup>42</sup></a> Predictive validity requires a criterion measure, which is often challenging to obtain; hence, it has been omitted in some validation studies.<a class="reference-link xref xref-bibr" href="#B42" data-jats-ref-type="bibr" data-jats-rid="B42"><sup>42</sup></a> In practice, these psychometric indices should be carefully examined before being applied to AI evaluation. Researchers need to understand the fundamental principles behind the concepts of reliability and validity in order to develop appropriate AI-specific metrics that meet the purposes of the research.</p>
</section>
</section>
</section>
<section id="sec12" class="sec">
<h2 class="heading">Opportunities</h2>
<p id="p-50">The integration of psychometrics into AI evaluation opens up exciting opportunities. In this section, we explore both the open questions that need addressing and the potential for expanding early psychometric applications to other AI domains more broadly.</p>
<section id="sec13" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>Open questions.</strong>  Leveraging psychometrics for AI evaluation brings unique challenges due to the evident differences between AI systems and humans. Addressing these challenges presents significant research opportunities to refine psychometric methods, adapt them for AI systems, and ultimately enhance the reliability and validity of AI evaluations.</p>
<section id="sec14" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><em>Handling AI sensitivity.</em>  The pronounced sensitivity of AI systems to prompts<a class="reference-link xref xref-bibr" href="#B19" data-jats-ref-type="bibr" data-jats-rid="B19"><sup>19</sup></a> raises questions about the validity and reliability of their evaluation. Are AI systems inherently too unstable to be evaluated reliably? If the measurement is unreliable, should we revise the items or optimize AI systems? When evaluating AI systems, should we use the same instructions or prompts designed for humans, or should we make adjustments? Should the prompts represent typical user commands or be tuned to elicit optimal performance? If the latter, how do we choose the optimal prompt from potential options, such as chain-of-thought? Answering these questions requires empirical evidence and community consensus.</p>
<p id="p-53">In addition to prompt sensitivity, AI systems may exhibit different behaviors in response to training and testing queries due to mechanisms such as alignment faking;<a class="reference-link xref xref-bibr" href="#B16" data-jats-ref-type="bibr" data-jats-rid="B16"><sup>16</sup></a> that is, AI systems may selectively comply with training queries while maintaining their incompliant behavior outside of training, leading to uneven alignment between selected training queries and other queries. Experimental results show that alignment-faking reasoning increases from 12% to 78% over the course of reinforcement learning on the tested LLM. It is therefore crucial to design reliable, valid measurements that account for this unique characteristic of AI systems.</p>
</section>
<section id="sec15" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><em>Redefining person and population.</em>  In psychometrics, it is crucial to distinguish between within-person and between-person analyses. The former examines variations in data from the same individuals over time, while the latter compares data from different individuals. This distinction determines the use of appropriate statistical methods and accurate interpretations. However, AI systems can take on an infinite variety of roles,<a class="reference-link xref xref-bibr" href="#B37" data-jats-ref-type="bibr" data-jats-rid="B37"><sup>37</sup></a> making these definitions ambiguous. For instance, are different personas created with varying prompts considered different persons? Is a fine-tuned AI model the “same person” as the previous model? And if we consider different versions of models (e.g., OpenAI o4 and OpenAI o4-mini) as “multiple persons,” do they belong to the same population?</p>
<p id="p-55">One possible solution to this issue is to examine the magnitude of the variance. A population usually has a larger variance due to a wide distribution, while repeated measures from one person are expected to show a smaller variance. However, what constitutes a “large” variance needs further research, with human variances being a possible reference. If multiple levels of data points are identified, we can adopt multilevel modeling<a class="reference-link xref xref-bibr" href="#B15" data-jats-ref-type="bibr" data-jats-rid="B15"><sup>15</sup></a> to examine the effects at the prompt level, persona level, and model level.</p>
</section>
<section id="sec16" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><em>AI vs. humans: A comparative exploration. </em> Tests applicable to both humans and AI are sometimes needed. To ensure a fair comparison, it is important to ascertain item appropriateness. This challenge may be addressed by adopting psychometric techniques such as differential item functioning (DIF),<a class="reference-link xref xref-bibr" href="#B36" data-jats-ref-type="bibr" data-jats-rid="B36"><sup>36</sup></a> which examines whether a test item works equally well in humans and AI systems, as well as among different AI systems.</p>
<p id="p-57">Without empirical evidence, it is impossible to determine whether a certain psychometric technique is appropriate or useful for AI evaluation. Given the nature of AI systems, which is different from humans, it is likely that only some psychometric techniques that have been used to analyze human data can be applied to AI systems. For example, Zou et al.<a class="reference-link xref xref-bibr" href="#B45" data-jats-ref-type="bibr" data-jats-rid="B45"><sup>45</sup></a> examined 500 LLM-based chatbots with distinct personality profiles and found that chatbots’ self-reported personality traits showed weak correlations with both user perception and interaction quality, highlighting significant validity concerns in traditional self-report-based assessment methods. Though we see great potential for construct-oriented evaluation based on a psychometric framework, we should be cautious with the techniques and always bear in mind the differences between AI systems and humans.</p>
<p id="p-58">Technical challenges extend beyond research: They can also have profound societal implications. The use of psychometrics can not only better ensure the robustness of AI systems and their evaluation but also connect AI evaluation with the evaluation of humans more directly for better public understanding. For instance, it can move AI evaluation from benchmarking, leaderboards, and scaling laws to more meaningful dialogues about safety and the future of work, with terminology and tools that are compatible with human assessment.</p>
</section>
</section>
<section id="sec17" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>Extending to AI research more broadly.</strong>  Extending the psychometric framework for construct-oriented AI evaluation, we see further research opportunities that psychometrics presents for all areas in AI.</p>
<section id="sec18" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><em>Evaluation of human-AI teaming.</em>  Considering AI systems’ great potential for enhancing human productivity, optimal human-AI teaming is receiving increasing attention. Human-AI teaming is a collaborative process where decisions are made jointly, with both parties learning from each other and adapting accordingly.<a class="reference-link xref xref-bibr" href="#B2" data-jats-ref-type="bibr" data-jats-rid="B2"><sup>2</sup></a> Thus, its evaluation should not only consider task outcomes but also encompass the performance of individual components (both human and AI), the quality of collaboration within the team, and the success in achieving shared objectives.<a class="reference-link xref xref-bibr" href="#B2" data-jats-ref-type="bibr" data-jats-rid="B2"><sup>2</sup></a> In this context, psychometric models and methodologies, particularly those related to collaborative problem solving, may offer valuable insights.<a class="reference-link xref xref-bibr" href="#B38" data-jats-ref-type="bibr" data-jats-rid="B38"><sup>38</sup></a></p>
</section>
<section id="sec19" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><em>Transforming the AI pipeline.</em>  Psychometrics has important social impact. For instance, educational assessment, a major application of psychometrics, is vital in shaping human education practices.<a class="reference-link xref xref-bibr" href="#B20" data-jats-ref-type="bibr" data-jats-rid="B20"><sup>20</sup></a> Similarly, AI evaluation could extend beyond measuring AI system performance only after it is trained. Instead, rigorous evaluation should guide each stage in AI development, as requested by several regulatory frameworks, such as the EU’s AI Act.<a class="reference-link xref xref-bibr" href="#B13" data-jats-ref-type="bibr" data-jats-rid="B13"><sup>13</sup></a> Inspired by educational practices, in <a class="xref xref-fig" href="#F3" data-jats-ref-type="fig" data-jats-rid="F3">Figure 3</a> we illustrate how psychometrics could help redefine the AI pipeline.</p>
<figure id="F3" class="fig" data-jats-position="float">
<div class="image-container"><img decoding="async" class="graphic" title="Figure 3. " src="https://cacm.acm.org/wp-content/uploads/2026/03/3769688_fig03.jpg" alt="" data-image-id="F3" data-image-type="figure" /></div><figcaption><span class="caption-label">Figure 3. </span> <span class="p">Comparison of the current AI pipeline, human education process, and a new AI pipeline supported by psychometrics.</span></p>
<div class="figcaption-footer"> </div>
</figcaption></figure>
<p id="p-63">In the first stage, objectives such as problem solving are identified and integrated into teaching and assessment to enhance future success in education. Likewise, psychometrics can help identify essential AI constructs for complex, unforeseen tasks, for example, critical thinking, and values such as security, aligning with Schwartz’s theory of basic values.<a class="reference-link xref xref-bibr" href="#B1" data-jats-ref-type="bibr" data-jats-rid="B1"><sup>1</sup></a> In the second stage, psychometrics can identify high-potential systems and optimize training resources accordingly. Throughout training, continuous feedback based on psychometrics can refine the process toward predefined objectives. In the final stage, after examining the reliability and validity of the tests, we can confirm that AI systems meet real-world challenges and societal integration requirements, such as mitigating harmful biases, ensuring accountability in automated decision making, and reducing hallucinations. By rigorously assessing capabilities, safety, and fairness, psychometrics helps build AI that not only performs accurately but also upholds societal trust and equitable outcomes.</p>
</section>
</section>
</section>
<section id="sec20" class="sec">
<h2 class="heading">Conclusion</h2>
<p id="p-64">When evaluating AI systems, the current task-oriented paradigm falls short in key areas such as predictive power, explanatory power, and quality assurance. These limitations could be addressed by focusing on constructs and placing psychometrics at the core of AI evaluation. Our work in this area makes it clear that this transition is both imperative and possible, and presents methodological guidelines to accelerate adoption. Using a three-stage framework, here we demonstrated how psychometrics can be leveraged to overcome the limitations of current benchmarks while avoiding the pitfalls of oversimplified psychometric application. Integrating psychometrics promises significant advances in AI evaluation but also presents unique challenges, such as reinterpreting human-centric concepts for AI systems and managing prompt sensitivity. Our discussion opens up new avenues for research, including the potential evaluation of human-AI teaming and the transformation of AI development practices. While psychometrics alone may not solve all issues in AI evaluation, the transition to construct-oriented evaluation is crucial and we look forward to a future where psychometrics-grounded AI evaluation also shapes the evaluation of humans.</p>
</section>
<section id="sec21" class="sec">
<h2 class="heading">Acknowledgments</h2>
<p id="p-65">This work was supervised by Luning Sun, Fang Luo, and Xing Xie. The authors would like to thank Jinyan Fan, Marija Slavkovik, Clemens Stachl, Alina A von Davier, Xiangen Hu, Yu Lu, Bryan Maddox, Mengxiao Zhu, Meng Li, Li Dong, Jindong Wang, Igor Sterner, Greg Serapio-García, Peter Romero, Fengli Xu, Fernando Martínez Plumed, and Lidong Zhou for the discussions and comments on early versions of the manuscript. This research was funded by Beijing Natural Science Foundation L247011; Microsoft Research Asia Collaborative Research Program grant FY23-Research-Sponsorship-422 “The convergence of assessing human and big model capabilities”; CIPROM/2022/6 (FASSLOW) funded by Generalitat Valenciana, and Spanish grant PID2024-162030OB-100 (ROBIN) funded by MCIN/AEI/10.13039/501100011033 and ERDF A way of making Europe, Cátedra ENIA-UPV in Sustainable AI Development, TSI-100930-2023-9, and INCIBE’s Chair funded by the EU-NextGenerationEU through the Spanish government’s Plan de Recuperación, Transformación y Resiliencia, and EUR2024-153548 (PREDAIT) “Towards Predictable AI” from “Spanish Europe Excelencia” 2024; National Natural Science Foundation of China (NSFC) (NO. 62476279, NO. 92470205, NO. U2436209); Major Innovation &amp; Planning Interdisciplinary Platform for the “Double-First Class” Initiative, Renmin University of China; The Fundamental Research Funds for the Central Universities, and the Research Funds of Renmin University of China No. 24XNKJ18; Fund for building world-class universities (disciplines) of Renmin University of China and Public Computing Cloud, Renmin University of China. L.S. and D.S. gratefully acknowledge financial support from Invesco through their philanthropic donation to Cambridge Judge Business School.</p>
</section>
</div>
</article>
]]></content:encoded>
					
					<wfw:commentRss>https://cacm.acm.org/research/evaluating-general-purpose-ai-with-psychometrics/feed/</wfw:commentRss>
			<slash:comments>1</slash:comments>
		
		
		      <dc:creator><![CDATA[Liming Jiang]]></dc:creator>
      <dc:creator><![CDATA[José Hernández-Orallo]]></dc:creator>
      <dc:creator><![CDATA[David Stillwell]]></dc:creator>
      <dc:creator><![CDATA[Shiqiang Chen]]></dc:creator>
      <dc:creator><![CDATA[Luning Sun]]></dc:creator>
      <dc:creator><![CDATA[Fang Luo]]></dc:creator>
      <dc:creator><![CDATA[Xing Xie]]></dc:creator>
<post-id xmlns="com-wordpress:feed-additions:1">779335</post-id>	</item>
		<item>
		<title>No Privacy without AI</title>
		<link>https://cacm.acm.org/opinion/no-privacy-without-ai/</link>
					<comments>https://cacm.acm.org/opinion/no-privacy-without-ai/#comments</comments>
		
		<dc:creator><![CDATA[Norman Sadeh]]></dc:creator>
		<pubDate>Mon, 13 Apr 2026 15:54:09 +0000</pubDate>
				<category><![CDATA[Artificial Intelligence and Machine Learning]]></category>
		<category><![CDATA[HCI]]></category>
		<category><![CDATA[Security and Privacy]]></category>
		<guid isPermaLink="false">https://cacm.acm.org/?post_type=digital-library&#038;p=779037</guid>

					<description><![CDATA[<p> Without AI, adequate privacy is simply out of reach.</p>]]></description>
										<content:encoded><![CDATA[<article>
<div class="body" lang="en">
<section id="sec1" class="sec">
<p id="p-1">Artificial intelligence (AI), especially the new generation of increasingly autonomous, agentic AI systems, has triggered understandable concerns about privacy. These systems can read our email messages, draft our documents, navigate our calendars, answer our questions, and even act on our behalf. They observe, analyze, and infer, often continuously. They can derive sensitive attributes from our digital traces and, with growing autonomy, sometimes initiate actions based on these inferences. Many of the privacy fears surrounding AI are real. But as paradoxical as it may seem, AI, including agentic AI, is also becoming essential to protecting privacy.</p>
<p id="p-2">This column argues that without AI, adequate privacy has become simply out of reach. This is not because AI is benign; it most definitely is not. Rather, the modern digital ecosystem has evolved to a point where no human, unaided, can understand, monitor, or manage the complexity of today’s data practices. For two decades, my collaborators and I have studied why people struggle to manage their privacy and why this struggle keeps getting worse despite decades of regulation, policy work, and advances in privacy enhancing technologies. From mobile applications and IoT devices to location-sharing, video analytics, websites, and AI chatbots, we have found the same underlying truth: privacy is too dynamic, too contextual, and too cognitively demanding for people to manage manually. When it comes to managing privacy, AI is not merely helpful. It is indispensable.</p>
</section>
<section id="sec2" class="sec">
<h2 class="heading">The Challenges AI Creates for Privacy</h2>
<p id="p-3">Before explaining why AI is also indispensable for privacy, we must acknowledge the genuine challenges it introduces. Agentic AI systems can collect data about us over extensive periods of time (for example, persistent interactions lasting months), influence our decisions, and autonomously act on data they collect about us. They can personalize at unprecedented levels, profiling individuals with a granularity that far exceeds traditional analytics. They can infer emotions or health indicators from a camera feed, often without explicit consent.</p>
<p id="p-4">Our research on the privacy implications of video analytics explored how surprisingly powerful inference capabilities have become and how little users know about them. Many people do not realize that cameras can detect mood, infer demographic traits, or extract behavioral cues even from innocuous footage (for example, identifying whether someone appears frustrated, distracted, or stressed). Studies reveal that user expectations are both diverse and highly context dependent: people may accept analytics for safety monitoring in public spaces yet find it unacceptable in workplaces or places of worship. Expecting people to systematically configure privacy settings to align with their individual expectations is simply unrealistic.</p>
<p id="p-5">Agentic AI’s power to act further exacerbates the threat landscape. A system capable of autonomously generating email messages, modifying settings, and communicating with other agents or systems may inadvertently disclose sensitive information its user would not want to. These challenges are serious and require thoughtful governance.</p>
<p id="p-6">Yet, despite these risks, the complexity of modern data ecosystems has outgrown the capacity of humans (and of traditional software) to manage privacy effectively. AI can help and is likely the only viable option.</p>
</section>
<section id="sec3" class="sec">
<h2 class="heading">The Privacy Landscape Has Become Too Complex</h2>
<p id="p-7">The mythical assumption that people would somehow be able to manage their privacy by glancing at a website’s privacy policy and toggling a handful of settings has never been farther from reality. Consider the mobile ecosystem alone. Our research shows that users face dozens or even hundreds of privacy decisions across their apps, but their preferences are nuanced, contextual, and vary widely depending on purpose, trust, time of day, and other factors. Expecting people to get acquainted with these nuances by reading privacy policies and carefully configuring privacy settings is simply unrealistic.</p>
<p id="p-8">The same is true in many other contexts. For instance, in video analytics contexts such as those associated with the deployment of these technologies in malls, stadiums, but also in some gyms, schools or even religious venues, preferences vary according to the purpose of analytics, the venue, retention practices, and the level of identifiability, producing a multidimensional policy space that no static interface can hope to adequately capture.<a class="reference-link xref xref-bibr" href="#B9" data-jats-ref-type="bibr" data-jats-rid="B9"><sup>9</sup></a> In short, the scale at which we interact with technologies, and the complexity of the dataflows they give rise to far exceed the limits of human attention, understanding, and time.</p>
</section>
<section id="sec4" class="sec">
<h2 class="heading">Regulatory Efforts Help … but They Also Increase Complexity</h2>
<p id="p-9">Regulations such as the GDPR or CCPA, as well as a growing collection of AI-focused frameworks and regulations, aim to improve transparency and accountability. They enshrine important rights: access, deletion, correction, opt-out, opt-in for sensitive data uses, purpose limitation, and more. They are vital steps forward, especially as AI-driven inference capabilities expand.</p>
<p id="p-10">But regulations such as GDPR and CCPA have also made privacy notices longer—often dramatically so. A recent Princeton study found that over the last 20 years privacy policies have become significantly more difficult to read and doubled in length.<a class="reference-link xref xref-bibr" href="#B2" data-jats-ref-type="bibr" data-jats-rid="B2"><sup>2</sup></a></p>
<p id="p-11">New regulations also offer more rights to users. This means more notices, more interfaces, more choices and increasingly more granular control to deal with. Yet few users have the time and motivation to find or exercise these controls. In fact, despite recent requirements to make these choices easier to identify, many people still struggle to locate and use opt-out links, account deletion interfaces, or the mechanisms required to invoke other legal rights (see Habib, H. et al.<a class="reference-link xref xref-bibr" href="#B3" data-jats-ref-type="bibr" data-jats-rid="B3"><sup>3</sup></a>).</p>
<p id="p-12">In the U.S., new provisions such as CCPA’s recognition of Authorized Agents, namely entities that are allowed to submit requests on the user’s behalf, and emerging standards such as the Global Privacy Control (GPC) signal represent meaningful progress. They establish pathways for intermediaries, potentially AI systems, to express user preferences on their behalf and contribute to reducing user burden. But currently, these mechanisms mostly apply to narrow categories of data sharing (for example, “Do Not Sell” signals), far from addressing the full spectrum of privacy challenges consumers face and far from being able to handle the diversity in preferences people have been shown to have.</p>
<p id="p-13">The paradox is clear: Regulators, in trying to empower individuals, have also created an environment where only AI can realistically help users understand and exercise their rights.</p>
</section>
<section id="sec5" class="sec">
<h2 class="heading">Privacy Assistants: Early Vision Meets Today’s Reality</h2>
<p id="p-14">Approximately 20 years ago, my colleagues and I began exploring the idea of privacy assistants, software agents capable of learning users’ privacy preferences, reading and interpreting privacy policies, nudging users to engage with available privacy settings, helping them locate and configure these settings, and respond to their privacy questions. Much of that early work anticipated today’s challenges: highly contextual data sharing rules, continuous sensing environments, AI-driven personalization, and the need for machine assistance to manage both complexity and volume.</p>
<p id="p-15">Our early work outlined semantic Web architectures for contextual privacy controls and rule-based reasoning about data practices in pervasive computing environments—nowadays more commonly referred to as the Internet of Things (IoT). Quickly this work led us to explore the complexity and diversity of people’s privacy expectations and preferences in different domains (for example, location sharing, mobile app permissions, and IoT contexts). These studies repeatedly showed that user preferences could be clustered into interpretable personas, and machine learning could help users quickly converge toward reasonable configurations without overwhelming them with choices and without taking away their sense of agency. Additional personalization was also shown to help further improve recommendations. We eventually piloted and released a mobile app privacy assistant to help users configure their mobile app privacy settings on Android phones with users reporting that they generally liked the way the assistant worked and followed its recommendations.<a class="reference-link xref xref-bibr" href="#B7" data-jats-ref-type="bibr" data-jats-rid="B7"><sup>7</sup></a></p>
<p id="p-16">Around 2012, we also started exploring the use of machine learning (ML) and natural language processing (NLP) techniques to automatically read and interpret the text of privacy policies, taking advantage of recent and ongoing advances in ML and NLP. This work, which has since inspired a number of follow-on efforts, initially relied on the tedious collection of extensive corpora of policy annotations to help train classifiers to identify and analyze increasingly rich sets of data practice disclosures. Over time, this work incorporated and extended advances in the field (from word embeddings, RNNs, LSTMs, and CNNs, to the more recent transformer models).<a class="footnote-link xref xref-fn" href="#fn1" data-jats-rid="fn1" data-jats-ref-type="fn"><sup>a</sup></a> We showed that it was possible to reliably identify and classify a wide range of statements and use this analysis to answer people’s privacy questions—see our work on Privacy Question Answering Assistants as well as related work by others.<a class="reference-link xref xref-bibr" href="#B5" data-jats-ref-type="bibr" data-jats-rid="B5"><sup>5</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B8" data-jats-ref-type="bibr" data-jats-rid="B8"><sup>8</sup></a></p>
<p id="p-17">We also demonstrated how these techniques could be extended to automatically identify and classify opt-out choices often buried deep in the text of privacy policies and surfaced these choices through our Opt-Out Easy browser extension: the extension was made available in the Chrome store and in many ways anticipated how today’s agents can be used to automatically identify links and buttons and navigate websites for users.<a class="reference-link xref xref-bibr" href="#B6" data-jats-ref-type="bibr" data-jats-rid="B6"><sup>6</sup></a></p>
<p id="p-18">In yet another line of inquiry, we experimented with privacy nudges designed to motivate people to pay attention and engage with privacy settings made available to them.<a class="reference-link xref xref-bibr" href="#B1" data-jats-ref-type="bibr" data-jats-rid="B1"><sup>1</sup></a> Some of this work eventually found its way into commercial products and much of it also inspired others to further explore this space. For instance, nudges we designed and piloted to motivate mobile app users to revisit their privacy settings were eventually adopted by Apple and introduced in iOS13, and our use of ML to learn people’s privacy preferences influenced how Google helps users manage Chrome permissions.<a class="reference-link xref xref-bibr" href="#B4" data-jats-ref-type="bibr" data-jats-rid="B4"><sup>4</sup></a></p>
<p id="p-19">What is particularly exciting today is how LLMs and agentic AI systems finally make it possible to implement the full vision: conversational, proactive, personalized, and delegated privacy management at scale.</p>
<p id="p-20">Personalized privacy assistants capable of automatically reading privacy policies, identifying privacy settings, helping us interpret how to best align these settings with our expectations and preferences, and warning us about data practices we may not expect, offer the prospect of rebalancing the privacy landscape and restoring people’s control over their data. Their ability to engage in dialogues with us, to check how we feel about different practices, to help us better understand the ramifications of our decisions, to manage many repetitive and tedious privacy decisions on our behalf, and, over time, to build detailed models of our individual preferences and of the most effective way of communicating with us could fundamentally shift the balance.</p>
<p id="p-21">They can drastically reduce user burden and help bridge the gap between what people care about and confusing jargon and assortments of privacy choices (e.g., understanding the difference between opting out of tracking by mobile apps on an iPhone and submitting an opt-out request under CCPA). Thanks to their ability to rapidly scan websites and apps, they can lower the bar when it comes to discovering available choices and helping users take advantage of them. Perhaps the most transformative capacity of agentic AI is its ability to act. New legal provisions recognizing authorized agents in regulations such as CCPA are an important first step in this direction, though more is needed to lower friction and empower agents to manage our privacy without undue burden on their users (for example, under current practices, CCPA opt-out or deletion requests submitted by authorized agents generally require manual verification by the user).</p>
</section>
<section id="sec6" class="sec">
<h2 class="heading">AI for Compliance and Threat Modeling</h2>
<p id="p-22">AI is not just critical for users. It is equally critical for organizations seeking to comply with increasingly complex regulation. Our work on mobile app privacy compliance analysis (MAPS) laid early foundations for compliance automation and, in particular, for automatically detecting discrepancies between disclosures made in mobile app privacy policies and the code and behavior of these mobile apps.<a class="reference-link xref xref-bibr" href="#B10" data-jats-ref-type="bibr" data-jats-rid="B10"><sup>10</sup></a> These technologies have influenced the development of products such as Google Checks.</p>
<p id="p-23">More recently we have been experimenting with agentic AI to semi-automatically identify shortcomings in the design and implementation of privacy notice and choice interfaces, namely determining whether notices and choices are clear, easily discoverable, meaningful, and free of deceptive design patterns. Here we use agentic AI to navigate websites and operationalize a taxonomy of user-oriented threats to spot areas where notice and choice interfaces fall short—see our UsersFirst privacy threat modeling framework.<a class="footnote-link xref xref-fn" href="#fn2" data-jats-rid="fn2" data-jats-ref-type="fn"><sup>b</sup></a> We believe that these are just early efforts in this space and expect that similar approaches will be developed to automate or semi-automate compliance analysis as well as the identification and mitigation of increasingly rich classes of privacy threats.</p>
<p id="p-24">A major reason why so many products and services continue to be non-compliant and fall short in adequately mitigating risks has to do with the very significant manpower required to conduct these types of analysis. By automating or semi-automating some of these processes, AI could help drastically change today’s sad state of affairs.</p>
</section>
<section id="sec7" class="sec">
<h2 class="heading">A Future Where AI Is Essential for Privacy</h2>
<p id="p-25">For decades, the field of privacy has operated under the assumption that transparency, control, and user choice could empower individuals to manage their privacy. This vision was well intentioned but deeply optimistic about human attention, comprehension, and time. Today, the diversity and complexity of dataflows with which people interact on a daily basis makes it painfully obvious that this assumption is not tenable and no traditional interface will change that. AI is not a threat to privacy by default. It is a threat when unregulated, misaligned, or opaque, and it can certainly be a significant threat when used to manipulate users to disclose more data and influence their decisions. But when designed responsibly, aligned with user interests (working as their privacy assistants), and integrated into regulatory frameworks, AI can become the only realistic path toward making privacy meaningful in an era of ubiquitous computing and pervasive data collection. By drastically reducing the burden required to manage one’s privacy today, agentic AI in particular offers a path toward restoring a sense of control over the collection and use of our data and the prospect of privacy compliance issues and threats more readily identified and mitigated.</p>
<p id="p-26">Simply put, AI has the potential to be the long-awaited game changer for privacy.</p>
</section>
</div>
</article>
]]></content:encoded>
					
					<wfw:commentRss>https://cacm.acm.org/opinion/no-privacy-without-ai/feed/</wfw:commentRss>
			<slash:comments>1</slash:comments>
		
		
		<post-id xmlns="com-wordpress:feed-additions:1">779037</post-id>	</item>
		<item>
		<title>The Centrality Fallacy and ACM</title>
		<link>https://cacm.acm.org/opinion/the-centrality-fallacy-and-acm/</link>
					<comments>https://cacm.acm.org/opinion/the-centrality-fallacy-and-acm/#respond</comments>
		
		<dc:creator><![CDATA[Moshe Y. Vardi]]></dc:creator>
		<pubDate>Fri, 10 Apr 2026 17:04:17 +0000</pubDate>
				<category><![CDATA[Computing Profession]]></category>
		<guid isPermaLink="false">https://cacm.acm.org/?post_type=digital-library&#038;p=779429</guid>

					<description><![CDATA[<p>The disconnect between ACM and its membership is an ongoing problem.</p>]]></description>
										<content:encoded><![CDATA[<article>
<div class="body" lang="en">
<section id="sec1" class="sec">
<p id="p-1">The sociologist Ron Westrum <a class="ext-link" href="https://www.sciencedirect.com/science/article/pii/S0925753514000174" data-jats-ext-link-type="uri">coined</a><a class="footnote-link xref xref-fn" href="#fn1" data-jats-ref-type="fn" data-jats-rid="fn1"><sup>a</sup></a> the phrase “The Centrality Fallacy,” whereby people in positions of authority incorrectly assume that due to their centrality they have the best information on an issue and, consequently, are positioned to make the best decisions. To use an analogy by the philosopher John Dewey: The shoe wearer understands better where the shoe is poorly fitted, whereas the cobbler understands better how to address the problem of poor fit.  Two recent developments within the Association for Computing Machinery (ACM) offer a stinging demonstration of the centrality fallacy.</p>
<p id="p-2">In 2022, an ACM Presidential Taskforce on SIG Overhead decided to increase the SIG overhead rate and minimum fee. That increase pushed some small SIGs into non-viability. I am involved with SIGLOG, the ACM’s Special Interest Group on Logic and Computation. It is a small SIG, though it does <a class="ext-link" href="https://www.floc26.org/" data-jats-ext-link-type="uri">represent</a><a class="footnote-link xref xref-fn" href="#fn2" data-jats-ref-type="fn" data-jats-rid="fn2"><sup>b</sup></a> a sizable, vibrant community. In the summer of 2024, it was suggested to SIGLOG officers that SIGLOG may need to merge with a larger SIG. That suggestion caused serious consternation, causing many to call for SIGLOG to secede from ACM. It is worth noting that SIGLOG <a class="ext-link" href="https://cacm.acm.org/opinion/why-doesnt-acm-have-a-sig-for-theoretical-computer-science/" data-jats-ext-link-type="uri">represents</a><a class="footnote-link xref xref-fn" href="#fn3" data-jats-ref-type="fn" data-jats-rid="fn3"><sup>c</sup></a> a mostly European style of computing research, while many European colleagues almost believe that the “A” in “ACM” stands for “American.” Once ACM’s and the SIG Governing Board’s leadership were made aware of the situation and asked whether ACM truly intends to eliminate small SIGs, there has been a significant damage-control effort, but a lot of ill will has already been created. And it all started with a bureaucratic decision about overhead rates.</p>
<p id="p-3">In 2020, the ACM Council decided that ACM would transition its Digital Library (DL) to full Open Access within five years. One trigger for that decision was a 2020 <a class="ext-link" href="https://www.change.org/p/association-for-computing-machinery-acm-support-open-access" data-jats-ext-link-type="uri">petition</a>,<a class="footnote-link xref xref-fn" href="#fn4" data-jats-ref-type="fn" data-jats-rid="fn4"><sup>d</sup></a> with more than 1,300 signatures, calling for opening the DL, which was a response to ACM publicly opposing (and then <a class="ext-link" href="https://www.acm.org/articles/bulletins/2020/january/acm-letter-to-ostp" data-jats-ext-link-type="uri">rescinding opposition</a><a class="footnote-link xref xref-fn" href="#fn5" data-jats-ref-type="fn" data-jats-rid="fn5"><sup>e</sup></a> to) a U.S. government push for open-access publishing. Since DL income is a major financial pillar of the ACM budget, Council adopted the business model of <a class="ext-link" href="https://libraries.acm.org/subscriptions-access/acmopen" data-jats-ext-link-type="uri">ACM Open</a>,<a class="footnote-link xref xref-fn" href="#fn6" data-jats-ref-type="fn" data-jats-rid="fn6"><sup>f</sup></a> which shifts the costs of supporting the DL from readers to authors. But rather than shift costs to individual authors via article processing charges (APCs), ACM Open mostly shifts the costs to home institutions of said authors. This is a highly nontrivial change, since it shifts costs mostly to research-intensive institutions. For a small institution such as Rice University, the cost of DL access under ACM Open rose threefold, and ACM asked me to get involved and convince Rice to sign up. In December 2025, ACM President Yannis Ioannidis announced the “Grand Opening of the ACM Digital Library.” The transition to an open DL was enacted during the last week of 2025; immediately after the start of the new year, ACM announced “ACM Is Now Fully Open Access!”</p>
<p id="p-4">But DL users during the last week of 2025, myself included (connecting from home), discovered to their total dismay that while the <i>data</i> of the DL, that is, articles and related materials, were now open, <i>data services</i><b><i>,</i></b> such as advanced search, author profile pages, cited-by tracing, and download and citations counts, were now closed, as they were considered <i>Premium Features.</i> On Jan. 8, 2026, ACM announced that the DL had been tiered: “DL Basic provides free, public access to all research content, while DL Premium offers advanced tools and value-added services.” Data services that were <i>freely available</i> under the “closed DL” now require Premium subscription under the “Open DL”.</p>
<p id="p-5">The response to that move has been swift. A <a class="ext-link" href="https://www.ipetitions.com/petition/restore-fully-free-and-open-access" data-jats-ext-link-type="uri">petition</a><a class="footnote-link xref xref-fn" href="#fn7" data-jats-ref-type="fn" data-jats-rid="fn7"><sup>g</sup></a> to re-open DL metadata has gathered more than 1,600 signatures by now. The comments accompanying the petition generally express outrage at how “100% open” can mean closed data services.</p>
<p id="p-6">How did ACM decide tiering the DL is in the spirit of opening the DL? According to the <a class="ext-link" href="https://www.acm.org/binaries/content/assets/about/annual-reports/fy24_dl_board.pdf" data-jats-ext-link-type="uri">2024 Annual Report</a><a class="footnote-link xref xref-fn" href="#fn8" data-jats-ref-type="fn" data-jats-rid="fn8"><sup>h</sup></a> of the ACM Digital Library Board, during FY24, ACM realized that the financial reality of ACM Open does not live up to its hopes. In fact, many institutions decided they have no incentive to subscribe at all, since the DL is going to be open. DL tiering was conceived as such an incentive, but the only communication to the broad membership about this issue was in an August 2025 CACM article, “<a class="ext-link" href="https://cacm.acm.org/article/acm-publications-finances-for-2023-and-2024/" data-jats-ext-link-type="uri">ACM Publications Finances for 2023 and 2024</a>,”<a class="footnote-link xref xref-fn" href="#fn9" data-jats-ref-type="fn" data-jats-rid="fn9"><sup>i</sup></a> which many members, including me, have overlooked.</p>
<p id="p-7">As John Dewey observed, in the absence of communication with the shoe wearer, a well-meaning cobbler can make very big mistakes. The disconnect between ACM and its membership is an ongoing problem. This is a lesson we must learn!</p>
</section>
<section id="sec2" class="sec"></section>
</div>
</article>
]]></content:encoded>
					
					<wfw:commentRss>https://cacm.acm.org/opinion/the-centrality-fallacy-and-acm/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
		<post-id xmlns="com-wordpress:feed-additions:1">779429</post-id>	</item>
		<item>
		<title>The Outlook for Computer Science Education</title>
		<link>https://cacm.acm.org/news/the-outlook-for-computer-science-education/</link>
					<comments>https://cacm.acm.org/news/the-outlook-for-computer-science-education/#comments</comments>
		
		<dc:creator><![CDATA[Sarah Underwood]]></dc:creator>
		<pubDate>Thu, 09 Apr 2026 15:35:53 +0000</pubDate>
				<category><![CDATA[Artificial Intelligence and Machine Learning]]></category>
		<category><![CDATA[Computing Profession]]></category>
		<category><![CDATA[Education]]></category>
		<guid isPermaLink="false">https://cacm.acm.org/?post_type=digital-library&#038;p=776999</guid>

					<description><![CDATA[<p>Universities are looking to broaden the appeal of CS education and to equip students with skills required by changing market demands.</p>]]></description>
										<content:encoded><![CDATA[<article>
<div class="body" lang="en">
<section id="sec1" class="sec">
<p id="p-1">Student enrollment is down, artificial intelligence (AI) challenges the norm, and the technology job market is tightening—a perfect storm for computer science (CS) education.</p>
<p id="p-2">To weather the storm, universities are rethinking CS education programs, offering new options to broaden appeal, embracing AI and GenAI, and collaborating with peers and industry to equip students with skills required by changing market demand.</p>
<p id="p-3">The statistics, however, are stark. Fall 2025 enrollment across U.S. undergraduate programs increased by 2.4% compared to last fall according to a November 2025 report by the National Student Clearinghouse Research Center.<a class="footnote-link xref xref-fn" href="#FN1" data-jats-rid="FN1" data-jats-ref-type="fn"><sup>a</sup></a> CS enrollment dropped nearly 6% at the undergraduate level over the same period, according to the study.</p>
<p id="p-4">Employment data published in August 2025 by the Federal Reserve Bank of New York<a class="footnote-link xref xref-fn" href="#FN2" data-jats-rid="FN2" data-jats-ref-type="fn"><sup>b</sup></a> showed CS graduates facing 6.1% unemployment in 2025. They still command the largest starting salaries at $80,000, but rank seventh-highest in unemployment across all college majors.</p>
<p id="p-5">The CS job crisis is the result of a convergence of issues including: the use of AI in entry-level jobs as a justification to reduce hiring; economic uncertainty (and AI) leading tech companies to cut substantial numbers of jobs in 2024 and into 2025; and an oversupply of CS graduates as universities doubled enrollment through 2022 and 2023, just as demand began to collapse. As one professor of CS put it, “If you get an okay CS degree from an okay university, the chance of being employed by a software house is much slimmer than it was two or three years ago because of AI, but there are still opportunities for the brightest and best. There is also, perhaps, change in what employers are looking for; maybe curiosity, creativity, and courage are as important as CS?”</p>
<p id="p-6">On this basis, universities are embracing AI, giving thought to their programs, providing students with exceptional skills, and sustaining research. Some are grateful for a drop in enrollment and a return to manageable class sizes, others are seeing dips and rises.</p>
<p id="p-7">Said Magda Balazinska, professor and director of the Paul G. Allen School of Computer Science and Engineering at the University of Washington, “We saw a dip in applications in our last admissions cycle compared to the year before, but that was to be expected and consistent with a national trend. The number of applications was still our fourth-highest ever and over 10 times our current capacity. Like many industries, tech goes through cycles, and student interest tends to follow. Our main worry is that prospective students see all these news stories about layoffs and AI coming for their jobs and get discouraged from applying despite having a strong interest in computer science.”</p>
<p id="p-8">That said, a survey of recent graduates of the university suggested the job outlook is positive. More than 120 different companies hired 2024-2025 Allen School graduates into software engineering roles. Amazon hired more than 100. Google and Meta hired fewer, but more than they hired the previous year. Microsoft was down a little, but still hired more than two dozen new graduates.</p>
<p id="p-9">Balazinska said, “We have said repeatedly that we graduate software engineers, and software engineering is much more than coding. Coding is an important foundational skill, but computer science is so much more: it’s creative, it’s problem-solving, it’s determining needs and then designing software and systems to address those needs. In our view, we aren’t preparing students to be ‘basic’ at anything, so perhaps that’s why our graduates continue to do so well.”</p>
<p id="p-10">At the U.K.’s University of Cambridge, Alastair Beresford, a professor of Computer Security and head of the Department of Computer Science and Technology, which saw student numbers rise over 10 years, but also decline slightly over the past couple of cycles, said, “Our mission is to develop next-generation disruptors and computer scientists.” Acknowledging that student numbers might go down further, he remains optimistic that his students will get good jobs. “We don’t see a shortage of interest in students coming out of our programs; you can’t automate them away.”</p>
<p id="p-11">The Cambridge CS department is ranked as one of the top three in the U.K. by the Complete University Guide,<a class="footnote-link xref xref-fn" href="#FN3" data-jats-rid="FN3" data-jats-ref-type="fn"><sup>c</sup></a> behind the University of Oxford and ahead of Imperial College London in 2026 subject league tables. Beresford encourages teaching using the Interactive, Constructive, Active, and Passive (ICAP) learning framework to teach the foundations of CS. It looks to get students further into the ‘I’ and ‘C’ areas, although notes the need for all four elements of ICAP, and a focus on challenge and creativity.</p>
<p id="p-12">Reflecting on change, Torsten Hoefler, a professor at Switzerland’s ETH Zurich where he directs the Scalable Parallel Computing Laboratory, a chief architect for machine learning at the Swiss National Supercomputing Center, and a Scientific Advisor to Microsoft on large-scale AI and networking, said, “We are entering the age of computation. Before this was the age of data, but now data is consumed by LLMs and computation is driving human progress.”</p>
<p id="p-13">He noted no need for wholesale change to the CS curriculum, but the necessity for some change through the inclusion of AI and LLMs in CS education. “If students solve problems with LLMs, I love it, but it is very important that they know how to use these tools and how they work when they leave my lectures. In CS, we don’t only teach the details of programming, but also help students develop a flexible mindset and critical thinking. Instructing an LLM is an art and a skill in itself.”</p>
<p id="p-14">Looking not only at internal CS education development, but also working on external programs with other U.S. universities, Carla Brodley, professor and dean of Inclusive Computing, and founding executive director of the Center for Inclusive Computing (CIC) at Boston’s Northeastern University, is working to broaden the appeal of CS. She said, “The mission of CIC is to help CS departments in the U.S. rethink how they offer the subject, so that all interested students can discover, thrive, and persist in computing.”</p>
<p id="p-15">During a recent learning session hosted by CIC that addressed how to react to the ‘CS is dead, long live AI’ narrative, Brodley suggested, “The most inherently flexible CS degree has a set of core requirements and then allows students to choose their concentration, perhaps AI, systems, theory, or cybersecurity. This structure provides flexibility to easily add new concentrations as needed.”</p>
<p id="p-16">Collaborating to make change, CIC partners with over 100 universities across the U.S. Most pertinent to increasing the appeal of CS, CIC has a focus on integrating CS majors with other fields. The aim is to ‘de-silo’ CS and offer a double major that&#8217;s not half CS, but perhaps one-third in CS and two-thirds in biology. Brodley suggested these types of interdisciplinary courses make a lot of sense for students and are a powerful strategy for broadening participation in computing, although they do come with organizational issues (like, how do you enroll students into a CS and design major, and what should it be called?) that need work and change.</p>
<p id="p-17">Northeastern has taken the lead with interdisciplinary computing majors (ICMs), with 46 ICMs offered in 2021, a number that grew to 56 in 2025. One question that arises is whether this success is due to Northeastern being a private university; could the concept be transferred into public higher-education institutions? To find out, two years ago, CIC started a project to help 10 public universities implement ICMs.</p>
<p id="p-18">Brodley explained, “These universities are diverse in terms of types of students, budgets, and political systems. The idea is to create a portfolio of universities that can successfully offer ICMs and then present their results to other universities.” To date, three of the 10 universities have started to offer ICMs. Further data across the 10 is expected to be available in fall 2026, allowing CIC and its partners to assess the administrative feasibility of ICMs in different institutional contexts and demonstrate their capacity to attract students from non-computing disciplines.</p>
<p id="p-19">Columbia University also is planning an interdisciplinary qualification, a Master of Science in AI (MSAI) degree that will combine core AI courses in CS and engineering and include other engineering departments, public policy, arts and media, bioinformatics, and statistics. More partners will be added once the qualification is officially established.</p>
<p id="p-20">Kathleen McKeown, Henry and Gertrude Rothschild Professor of Computer Science on the engineering faculty at Columbia University, said, “Developing new programs in areas that are important to industry is helpful, as well as concentrations in AI.” However, she warned that more thought needs to be given to what these technologies are used for and their unintended consequences, and noted that while there are plenty of firms using AI in a positive way, there are cases in industry where it is being used irresponsibly. Said McKeown, “There should be more work on ethics in the classroom.”</p>
<p id="p-21">This is on the agenda at Carnegie Mellon University (CMU), which is adjusting its CS courses concurrently with technology changes. Thomas Cortina, a professor in the CS department and associate dean for undergraduate programs in CMU’s School of Computer Science, said, “The Web changed how we taught, but we reached a steady state. In a changing environment with AI, where will we plateau and stabilize this time?” As yet, there is no clear answer.</p>
<p id="p-22">The university is taking a broad approach to incorporating AI in its courses, often through project work. One of Cortina’s courses covers computer music, essentially using a computer as both composer and musical instrument. The computer is programmed to synthesize sound and to look at how compositions made up of a sequence of sounds can be created. “During the final project, students may want to transcribe music scores from other musicians,&#8221; Cortina said. &#8220;They can use an AI tool to do this otherwise tedious task, but we are not grading on that. The tool is used as an assistant to the student’s work.”</p>
<p id="p-23">CMU also is rolling out an elective course that will allow students across the university to build their own chatbots. They will learn how AI tools work, what they can build, and how the chatbot will review questions. They also will learn the potential hazards of AI, including data privacy, bias, and social impact. An advanced course for computer scientists with a deeper background in machine learning, design, and program analysis also ticks the hazards box and encourages students to use AI to build and test systems.</p>
<p id="p-24">Regarding the outlook for CS education, Cortina said, “There will be change. When leading tools such as ChatGPT, Gemini, and Claude reach a steady state, teaching and learning will move on to a new level.”</p>
<p id="p-25">If the days of abundant student oversubscription to CS education programs and the promise of lucrative jobs are over, only time will tell if these changes, collaborations, and commitments will strengthen CS departments and provide students with skills that meet market demand.</p>
<h2 id="FurtherReading" class="heading">Further Reading</h2>
<ul id="reflist1" class="ref-list">
<li class="ref">
<div id="B1" class="citation"><span class="mixed-citation" data-jats-publication-type="other"><em>Chi, M.T.H. and Wylie, R.</em> <br /><strong>The ICAP Framework: Linking Cognitive Engagement to Active Learning Outcomes, <em>Educational Psychologist</em>, Volume 49, 2014, Issue 4. </strong><a class="ext-link" href="https://www.tandfonline.com/doi/full/10.1080/00461520.2014.965823" data-jats-ext-link-type="uri"><strong>https://www.tandfonline.com/doi/full/10.1080/00461520.2014.965823</strong></a></span></div>
</li>
<li class="ref">
<div id="B2" class="citation"><span class="mixed-citation" data-jats-publication-type="other"><em>Franklin, D., Denny, P., Gonzalez-Maldonado, D.A., and Tran, M.</em> <br /><strong>Generative AI in Computer Science Education, Challenges and Opportunities, Cambridge University Press, 5 April 2025. </strong><a class="ext-link" href="https://www.cambridge.org/core/elements/abs/generative-ai-in-computer-science-education/0A22106CBD7FCB391FD120C56E21420F" data-jats-ext-link-type="uri"><strong>https://www.cambridge.org/core/elements/abs/generative-ai-in-computer-science-education/0A22106CBD7FCB391FD120C56E21420F</strong></a></span></div>
</li>
<li class="ref">
<div id="B3" class="citation"><span class="mixed-citation" data-jats-publication-type="other"><strong>Northern University, CIC Five Year Report. May 2025. </strong><a class="ext-link" href="https://cic.northeastern.edu/" data-jats-ext-link-type="uri"><strong>https://cic.northeastern.edu/</strong></a></span></div>
</li>
</ul>
</section>
<section id="sec2" class="sec"></section>
</div>
<footer class="back"></footer>
</article>
]]></content:encoded>
					
					<wfw:commentRss>https://cacm.acm.org/news/the-outlook-for-computer-science-education/feed/</wfw:commentRss>
			<slash:comments>1</slash:comments>
		
		
		<post-id xmlns="com-wordpress:feed-additions:1">776999</post-id>	</item>
		<item>
		<title>The Indispensable Role of User Simulation in the Pursuit of AGI</title>
		<link>https://cacm.acm.org/opinion/the-indispensable-role-of-user-simulation-in-the-pursuit-of-agi/</link>
					<comments>https://cacm.acm.org/opinion/the-indispensable-role-of-user-simulation-in-the-pursuit-of-agi/#respond</comments>
		
		<dc:creator><![CDATA[Krisztian Balog and Chengxiang Zhai]]></dc:creator>
		<pubDate>Tue, 07 Apr 2026 16:51:00 +0000</pubDate>
				<category><![CDATA[Artificial Intelligence and Machine Learning]]></category>
		<category><![CDATA[HCI]]></category>
		<guid isPermaLink="false">https://cacm.acm.org/?post_type=digital-library&#038;p=779043</guid>

					<description><![CDATA[<p>User simulation can significantly accelerate the iterative cycle of AI development and testing, making it a critical component for achieving AGI.</p>]]></description>
										<content:encoded><![CDATA[<article>
<div class="body" lang="en">
<section id="sec1" class="sec">
<p id="p-1">The rapid advancements in generative artificial intelligence (AI), particularly large language models (LLMs), have fueled significant excitement about the potential for achieving artificial general intelligence (AGI)—AI exhibiting human-level cognitive abilities across a wide range of tasks. We observe impressive performance on specific benchmarks and a steady growth in model capacity. Current AGI research focuses heavily on scaling these foundation models and enhancing specific agent capabilities, such as complex reasoning and coding. However, despite this progress, even the most advanced AI systems remain far from possessing common sense reasoning, planning, and robust generalization capabilities, which are hallmarks of human intelligence.</p>
<p id="p-2">A critical, yet often underestimated, bottleneck hindering faster progress toward AGI is the heavy reliance on human interaction data for training and evaluation. This process is inherently slow, expensive, and difficult to scale. User simulation—the use of computational models to mimic human behavior during interactions with AI systems—offers a powerful avenue to address this bottleneck. By generating synthetic interaction data and providing automated, scalable, and reproducible evaluation environments, user simulation can significantly accelerate the iterative cycle of AI development and testing, thereby speeding up progress toward AGI.</p>
<p id="p-3">However, research dedicated to user simulation has not yet garnered the attention it deserves, given its potential impact on the AGI quest. In this Opinion column, we argue that user simulation is not merely a useful tool but a critical component for achieving AGI. We posit that the development of highly capable AI agents and the creation of realistic user simulations must proceed hand-in-hand. These two lines of research are synergistic, advancing toward AGI from complementary directions: one focusing on the agent’s capabilities, the other on the complexity and realism of the interaction environment it operates within.</p>
<p id="p-4">Creating realistic user simulations is an inherently interdisciplinary challenge. It requires integrating insights not only from machine learning and natural language processing but also crucially from psychology, cognitive science, and human-computer interaction to accurately model the complexities of human decision-making, preferences, biases, and interaction patterns.<a class="reference-link xref xref-bibr" href="#B2" data-jats-ref-type="bibr" data-jats-rid="B2"><sup>2</sup></a> While recent breakthroughs in LLMs provide powerful new tools for building more sophisticated simulators,<a class="reference-link xref xref-bibr" href="#B1" data-jats-ref-type="bibr" data-jats-rid="B1"><sup>1</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B7" data-jats-ref-type="bibr" data-jats-rid="B7"><sup>7</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B9" data-jats-ref-type="bibr" data-jats-rid="B9"><sup>9</sup></a> leveraging these tools effectively necessitates a deep, interdisciplinary understanding of human behavior.</p>
</section>
<section id="sec2" class="sec">
<h2 class="heading">What Is User Simulation?</h2>
<p id="p-5">User simulation involves creating computational agents designed to mimic how real humans might interact with an AI system. These agents are built using algorithms, rules, or models informed by our understanding of user behavior, knowledge, preferences, and cognitive processes. Crucially, they can often be parameterized to represent a diverse range of user characteristics, for example, novices vs. experts or users with different goals and interaction styles.</p>
<p id="p-6">User simulators serve two critical roles when integrated with interactive AI systems: they enable evaluation via repeatable, reproducible, and low-cost experiments, saving invaluable user time; and they can generate large-scale synthetic interaction data for training, especially when real user data is scarce, sensitive, or difficult to obtain.</p>
<p id="p-7">The scope of user simulation is broad, ranging from relatively simple models predicting single user actions (like clicks or ratings) to sophisticated agents modeling complex, goal-oriented behavior across multiple tasks or sessions (such as writing code).</p>
<p id="p-8">At a high level, there are two main families of simulation techniques. Model-based approaches rely on explicit representations of behavior, such as predefined rules derived from expert knowledge or interpretable probabilistic models capturing uncertainties. These often allow parameters (set heuristically or learned from data) to be tuned to represent different user types. In contrast, data-driven approaches leverage machine learning, often deep neural networks, to learn interaction patterns directly from large datasets of observed user behavior. While these models can achieve high predictive fidelity, they operate as “black boxes,” sacrificing interpretability regarding why the simulator behaves a certain way.</p>
<p id="p-9">Effective user simulators are characterized by several key properties, including validity, interpretability, cognitive plausibility, variation, and adaptability.<a class="reference-link xref xref-bibr" href="#B2" data-jats-ref-type="bibr" data-jats-rid="B2"><sup>2</sup></a> Although optimizing a simulator across all these dimensions is desirable, real-world development usually involves navigating trade-offs tailored to the application. High fidelity might be paramount for data augmentation, for example, while interpretability is often key for evaluation purposes. Reflecting these trade-offs, hybrid approaches are also common, combining model-based techniques with machine-learned components to balance different properties.</p>
<p id="p-10">Crucially, simulation does not need to be perfect to be useful. In fact, creating a “perfect” user simulator, i.e., one that flawlessly replicates human behavior across all possible tasks and contexts, is likely an AI-complete problem, on par with achieving AGI.</p>
</section>
<section id="sec3" class="sec">
<h2 class="heading">User Simulation to Accelerate the Path to AGI</h2>
<p id="p-11">The development of realistic user simulators is, in many respects, fundamentally aligned with the broader pursuit of AGI—creating intelligent agents with human-like capabilities. This alignment is not merely conceptual; it is reflected in the shared technological foundations employed by both fields throughout AI history. From rule-based expert systems prevalent in the 1980s and 90s, through the adoption of probabilistic models in the 2000s, the subsequent rise of machine-learned models, and the recent wave of transformer architectures and LLMs, advancements in core AI technologies have consistently been leveraged to build both more capable task agents and more realistic user simulators. Consequently, technical challenges in building sophisticated simulators often mirror those in developing intelligent task agents, suggesting deep and synergistic connections between user simulation and AGI research.</p>
<p id="p-12">Beyond this foundational alignment, user simulation directly addresses critical bottlenecks slowing progress in AI development. It enables scalable, reproducible, and low-cost evaluation, significantly accelerating development cycles compared to relying solely on time-consuming, non-reproducible, and expensive human testing. Furthermore, simulation can be used to generate synthetic interaction data, essential for training agents (e.g., via reinforcement learning), particularly when real data is scarce, sensitive, or unavailable at the required scale. Indeed, the concept of simulation is already implicitly embedded in modern LLM training paradigms: in Reinforcement Learning from Human/AI Feedback, the learned reward model, trained on human/AI-generated preference labels to act as a proxy for human judgment, essentially functions as a non-interpretable user simulator capturing preferences to guide the agent’s learning. An autonomous intelligent agent may leverage its own user simulation agent to self-generate synthetic data for optimization of its interactions with real users, thus implementing a functional analogue of Theory of Mind.<a class="reference-link xref xref-bibr" href="#B8" data-jats-ref-type="bibr" data-jats-rid="B8"><sup>8</sup></a></p>
<p id="p-13">Beyond evaluation and data generation, user simulation is indispensable for developing agents capable of effective human-AI collaboration. Realizing the full potential of such partnerships requires AI agents to do more than merely exhibit superhuman task performance; they must account for the inherent variability in their human partners’ behavior, including diverse problem-solving approaches, individual preferences, and suboptimal actions. The importance of this adaptation is underscored by recent work in chess, a domain that has long served as a Petri dish for AI research: human players paired with AI agents tailored to their skill level outperform those partnered with more powerful AI agents that are not adjusted for skill-compatibility.<a class="reference-link xref xref-bibr" href="#B5" data-jats-ref-type="bibr" data-jats-rid="B5"><sup>5</sup></a></p>
<p id="p-14">Achieving such effective, synergistic collaboration necessitates that the AI agent understands, predicts, and adapts to its human partner. This adaptation fundamentally requires the agent to leverage sophisticated models of the user’s knowledge, intentions, and decision-making processes—essentially demanding an embedded or tightly integrated user simulation capability. An intelligent agent assisting a user must respond appropriately to user actions and inferred needs based on such a model. This implies a close, reciprocal relationship: the task agent uses simulation for feedback to optimize its interaction policy, while the simulator must potentially adapt to changes in the task agent or environment over time. Therefore, the interdependence between research on intelligent task agents and user simulation is inherent and likely to persist until AGI is achieved.</p>
</section>
<section id="sec4" class="sec">
<h2 class="heading">Advancing User Simulation in the Age of LLMs: Challenges and Roadmap</h2>
<p id="p-15">The emergence of large language models represents a significant technological leap, potentially accelerating the integration and synergy between intelligent task agents and user simulators discussed earlier. As these models demonstrate increasing capabilities, they may well serve as a foundational building block for both types of agents. Indeed, LLMs have already fueled extensive adoption as both task agents and simulation tools across diverse domains and applications (for comprehensive surveys, see Garcez and Lamb<a class="reference-link xref xref-bibr" href="#B4" data-jats-ref-type="bibr" data-jats-rid="B4"><sup>4</sup></a> and Wang et al.<a class="reference-link xref xref-bibr" href="#B10" data-jats-ref-type="bibr" data-jats-rid="B10"><sup>10</sup></a>). However, while this presents a considerable opportunity, harnessing the power of LLMs specifically for realistic user simulation requires confronting several significant challenges.</p>
<section id="sec5" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>Challenge: Achieving realistic and controllable behavior.</strong>  While LLMs can produce fluent interactions, their generated responses can be unpredictable, sometimes unsafe, exhibit unrealistic or incoherent patterns of behavior, and may also lack the natural variation observed in real human interactions.<a class="reference-link xref xref-bibr" href="#B2" data-jats-ref-type="bibr" data-jats-rid="B2"><sup>2</sup></a> Furthermore, LLMs often possess more knowledge than average humans and generate overly “perfect” responses. This “superuser” effect, while beneficial and, indeed, intended for a task agent, results in unrealistic simulations that fail to capture typical human limitations, knowledge gaps, biases, or error patterns. While prompting techniques can guide LLM behavior, ensuring strict adherence to instructions remains a challenge.<a class="reference-link xref xref-bibr" href="#B2" data-jats-ref-type="bibr" data-jats-rid="B2"><sup>2</sup></a> Therefore, a key direction for future research is to develop more robust methods for controlling and calibrating LLM behavior within simulation contexts. This includes techniques to reliably constrain outputs, define specific personas with realistic limitations (for example, cognitive capabilities), inject natural variation and error patterns, and calibrate knowledge levels to match target user profiles, moving beyond the limitations of current prompting strategies.</p>
</section>
<section id="sec6" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>Challenge: Bridging the cognitive gap.</strong>  Beyond surface behavior, a more fundamental challenge lies in bridging the cognitive gap between current LLM capabilities and human cognition. While LLMs might be aware of concepts like patience or satisfaction, they lack the training data to model the human dynamics of such behaviors. Similarly, LLMs lack a deep understanding of core human cognitive processes, such as decision-making, memory recall, and attention span, and may fail to accurately simulate mechanisms of decision-making under uncertainty, the limitations of working memory, shifts in attention, or the influence of cognitive biases. These shortcomings hinder their ability to generate realistic simulations of human users. To overcome this, LLMs must be extended with components that capture a wider range of human cognitive abilities. The human brain is known to consist of two distinct systems: System 1, which is intuitive, fast, but not reliable, and System 2, which is logical, deliberate, but slower.<a class="reference-link xref xref-bibr" href="#B6" data-jats-ref-type="bibr" data-jats-rid="B6"><sup>6</sup></a> While current LLMs appear to be able to simulate System 1 well, they lack the deliberate reasoning and planning capabilities characteristic of System 2. To address this cognitive gap, we need to explore hybrid architectures that integrate explicit cognitive models—capturing aspects like memory update mechanisms, attention allocation strategies, or decision heuristics identified in cognitive science—with the generative power of LLMs. Neurosymbolic approaches, which combine neural learning’s flexibility with symbolic reasoning’s structure and interpretability, represent a particularly promising direction for embedding “System 2 capabilities” into user simulators.<a class="reference-link xref xref-bibr" href="#B4" data-jats-ref-type="bibr" data-jats-rid="B4"><sup>4</sup></a></p>
</section>
<section id="sec7" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>Challenge: Fostering interdisciplinary research and community.</strong>  Building simulators that reflect authentic human behavior requires more than just sophisticated machine learning. We need insights from psychology, cognitive science, and human-computer interaction for empirically grounded models of interaction patterns, cognitive processes (like attention, memory limitations, and biases), and the diversity of user populations. Effective modeling of user goals, evolving knowledge during interaction, and reasoning capabilities necessitates leveraging techniques from information science and knowledge representation. Furthermore, connections with intelligent agents and multi-agent systems are crucial for simulating complex social interactions and potentially embedding simulation capabilities within adaptive task agents.</p>
<p id="p-19">Therefore, a critical future direction is actively building bridges between these disciplines. This requires creating and supporting venues to foster interdisciplinary collaborations, and developing shared experimental platforms and evaluation resources. Fostering a vibrant, truly interdisciplinary research community dedicated to user simulation is paramount for making substantial progress.</p>
</section>
</section>
<section id="sec8" class="sec">
<h2 class="heading">Conclusion</h2>
<p id="p-20">The journey toward AGI demands more than advances in agent capabilities alone; it requires robust methods for evaluation, training, and ensuring adaptive interaction. This Opinion column has made the case that user simulation is a critical, indispensable technology for meeting these demands, highlighting its synergistic interdependence with core AGI agent research. Therefore, we advocate for a renewed focus and strategic investment in user simulation research as crucial for making progress toward truly intelligent systems.</p>
</section>
</div>
<footer class="back"></footer>
</article>
]]></content:encoded>
					
					<wfw:commentRss>https://cacm.acm.org/opinion/the-indispensable-role-of-user-simulation-in-the-pursuit-of-agi/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
		      <dc:creator><![CDATA[Chengxiang Zhai]]></dc:creator>
<post-id xmlns="com-wordpress:feed-additions:1">779043</post-id>	</item>
		<item>
		<title>Passing on Passwords</title>
		<link>https://cacm.acm.org/news/passing-on-passwords/</link>
					<comments>https://cacm.acm.org/news/passing-on-passwords/#respond</comments>
		
		<dc:creator><![CDATA[David Geer]]></dc:creator>
		<pubDate>Mon, 06 Apr 2026 16:13:14 +0000</pubDate>
				<category><![CDATA[Computing Applications]]></category>
		<category><![CDATA[Security and Privacy]]></category>
		<guid isPermaLink="false">https://cacm.acm.org/?post_type=digital-library&#038;p=777007</guid>

					<description><![CDATA[<p>The cybersecurity industry is working to supplant passwords with passkeys, which users don’t have to remember and criminal hackers can’t phish or guess.</p>]]></description>
										<content:encoded><![CDATA[<article>
<div class="body" lang="en">
<section id="sec1" class="sec">
<p id="p-1">“Passkeys are a way to log in without using a password,” said Oleg Naumenko, chief executive officer of Hideez, a cybersecurity company. Passkeys use public key cryptography; instead of a user typing an alphanumeric string they have memorized, a device verifies their identity and presence using a private key, explained Naumenko.</p>
<p id="p-2">The cybersecurity industry is working to supplant passwords with passkeys, which criminal hackers can’t phish, guess, or abuse across multiple accounts. For users and organizations hoping for a password replacement they don’t have to remember and that criminals can’t access, passkeys serve that purpose.</p>
<p id="p-3">The Fast Identity Online 2 (FIDO2) standard and its protocols from the FIDO Alliance standards body serve as the basis for passkeys. Explained Robert Dang, principal advisory director, security at Info-Tech Research Group, “FIDO2 is fast gaining traction as the new standard in user authentication, helping us move beyond passwords through public key cryptography to make logins easier and safer. It combines the functionality of WebAuthn, which allows browsers or apps to use passkeys, and [the] Client to Authenticator Protocol (CTAP2), which communicates with devices such as phones, [hardware] tokens, or wearables, to perform user authentication through USB, NFC, or Bluetooth.”</p>
<p id="p-4">The WebAuthn Web protocol lets users log in to websites without passwords, using biometrics or security keys instead. The CTAP2 device protocol helps user devices communicate securely with those keys. The FIDO Alliance and the World Wide Web Consortium (W3C) developed these protocols.</p>
<p id="p-5">FIDO Alliance members come from industries that use or develop authentication technologies, such as passwords and passkeys. Members include Apple, Microsoft, Google, Bank of America, SK Telecom, CVS Health, Amazon, and Target. Apple, Google, and Microsoft have developed passkey technologies based on the FIDO2 standard.</p>
<p id="p-6">According to the FIDO Alliance,</p>
<ul class="list" data-jats-list-type="bullet">
<li>Bank of America uses biometrics and security keys for customer banking;</li>
<li class="list-item">
<p id="p-7">SK Telecom uses FIDO authentication for telecommunications in Korea to secure digital identities;</p>
</li>
<li class="list-item">
<p id="p-8">CVS Health uses passkeys across healthcare and retail to enable secure customer access;</p>
</li>
<li class="list-item">
<p id="p-9">Amazon uses passkeys for retail account holders globally;</p>
</li>
<li class="list-item">
<p id="p-10">Target uses FIDO passkeys for employee and consumer logins.</p>
</li>
</ul>
<p id="p-11">“Within the FIDO2 framework, a person’s private key is safely stored on their own device or trusted providers such as Apple, Google, or Microsoft, while only the public key is shared when they authenticate with the account,” said Dang.</p>
<p id="p-12">“Most users won’t understand the mechanics behind it, but will likely value the convenience. This offers benefits to business leaders in gaining reduced risk and higher trust, while technical leaders receive device-bound keys, origin binding, and interoperability. FIDO2 is not just a technical improvement; it is a strategic shift with usability, enabling stronger security across all digital ecosystems,” said Dang.</p>
<p id="p-13">Origin binding locks each credential to a website domain and the physical device that created it, preventing reuse or phishing on other sites or hardware.</p>
<p id="p-14">According to Naumenko, device-bound passkeys live on smartphones, security chips, and physical keys such as a YubiKey or the Hideez Key.</p>
<p id="p-15">The security chips include a computer’s Trusted Platform Module (TPM), Apple’s Secure Enclave, Google’s Titan M chip, and the chips in YubiKeys and similar devices. A TPM is a chip inside computers that protects stored data and verifies hardware integrity. Apple’s Secure Enclave is a secure part of Apple devices that keeps passwords and biometric info private. Google’s Titan M chip secures login data and sensitive operations on its Pixel phones.</p>
<p id="p-16">The YubiKey is a small device that securely verifies user identity by working as a login key for computers, networks, and online accounts. It supports passwordless access and multiple authentication methods with a simple tap or plug-in. The Hideez Key is a FIDO2 device and password manager that supports one-time passwords and serves as an RFID card, with proximity-based PC lock and unlock.</p>
<p id="p-17">Apple and Google have introduced synced passkeys, said Naumenko, who explained the cloud stores these passkeys and synchronizes them between devices within the entire ecosystem.</p>
<p id="p-18">“So, Apple passkeys sync across all Apple devices signed into the same Apple ID. Google Passkeys sync across devices where you’re signed in to Chrome with your Google Account. This makes synced passkeys very convenient for everyday users. For example, if you register on a website from your phone, you can later log in from your laptop or tablet without setting anything up again,” said Naumenko.</p>
<p id="p-19">The Google Password Manager can manage Google passkeys, which are now also accessible on non-Google devices via Chrome. Microsoft passkeys are integrated with Windows Hello and Microsoft devices as the default login for new Microsoft accounts. Apple passkeys integrate with Face ID and Touch ID for authentication. The iCloud keychain manages these passkeys, storing and syncing them in the cloud across Apple devices.</p>
</section>
<section id="sec2" class="sec">
<h2 class="heading">What Passkeys Do and Don’t Do</h2>
<p id="p-20">Anna Pobletts, director of engineering at cybersecurity company 1Password, said the average person “doesn’t have strong password habits. It’s easier to reuse the same password or choose something simple to remember across multiple accounts because it’s convenient. However, that convenience leaves passwords vulnerable to phishing, credential reuse, and data breaches. Passkeys eliminate that trade-off, since security is cryptographically built into the technology.”</p>
<p id="p-21">Passkeys are not a catch-all solution for cybersecurity; they only handle logins and serve as an authentication factor, replacing passwords. “Passkeys are designed for authentication, not for securing what happens after a user signs in. Threats like session hijacking or malware still require broader security awareness and strong digital hygiene. Passkeys make signing in significantly more secure, but . . . they’re one part of a larger security picture,” said Pobletts.</p>
<p id="p-22">Session hijacking means someone takes over a user’s active online session by stealing their session data; passkeys can’t stop it because the attack happens after users are already logged in and the session is active. Strong digital hygiene for people means keeping software up to date and avoiding suspicious links; for organizations, it means setting clear security policies, training employees, controlling system access, and regularly backing up data.</p>
</section>
<section id="sec3" class="sec">
<h2 class="heading">The Difficulties of Replacing Passwords</h2>
<p id="p-23">The appearance of the passkey doesn’t mean the password immediately vanishes from use. According to Joel Burleson-Davis, chief technology officer at cybersecurity company Imprivata, organizations rolling out passkeys may need a fallback, such as a password, so users can recover their keys.</p>
<p id="p-24">“So, if the password still exists to give you access to that service account, then that is the weak link. And you don’t really solve your ‘steal-able, phish-able’ credentials problem, because there’s still a password out there,” said Burleson-Davis.</p>
<p id="p-25">This is the case for hardware-bound and synced passkeys. Passwords also remain commonplace for their familiarity. “Despite passkeys offering a potential solution for many of the central cybersecurity issues that passwords present, mainly phishing, [passwords] still remain almost completely ubiquitous. Passwords are the default technology system for accessing private accounts, with users having years of experience using these on a daily basis,” said Dray Agha, senior manager of the hunt and response team at Huntress, a cybersecurity company.</p>
<p id="p-26">“Even with the potential benefits of passkeys, some users may be resistant to change. Also, more generally speaking, individuals who are less inclined to tech may not totally understand how they work, which could prevent them from developing any interest in using passkeys,” said Agha.</p>
<p id="p-27">He added, “A complete shift from passwords to passkeys would require major investment from leading tech firms, prompting users to change to this new system as soon as possible. But the infrastructural costs of this solution outweigh the benefits for many businesses.”</p>
<p id="p-28">According to data from <a class="ext-link" href="https://guptadeepak.com/the-economics-of-authentication-why-passwordless-pays/" data-jats-ext-link-type="uri">Deepak Gupta’s consulting blog</a>, implementation of modern passwordless solutions can cost $100,000 to $300,000 for integration, user migration, and system configuration.</p>
</section>
<section id="sec4" class="sec">
<h2 class="heading">The 2FA + passwords Alternative to Passkeys</h2>
<p id="p-29">“When people refer to 2FA + password, they usually mean signing in with a password and then confirming their identity with a second, independent factor. That second factor is typically something other than ‘what you know,’ since the password already serves that role,” said Pobletts.</p>
<p id="p-30">According to 1Password’s Pobletts, the main alternatives to passkeys are one-time passcodes delivered by SMS or email, or generated by an authenticator app.</p>
<p id="p-31">“However,” observed Pobletts, “these methods are slow, add friction to the sign-in flow, and create opportunities for phishing or interception. The only real advantage is that users are already familiar with them.”</p>
<p id="p-32">As such, 2FA + password approaches, which combine a two-factor authentication step and a password, are more of a hazard than a stopgap until the industry moves to passkeys.</p>
</section>
<section id="sec5" class="sec">
<h2 class="heading">Will Passkeys Solve All Issues Seen with Passwords?</h2>
<p id="p-33">Passwords will remain an issue for legacy applications that are incompatible with passkeys. According to Dave Taku, vice president and global head of product management and UX at RSA, a cybersecurity company, enterprises and government agencies will have trouble implementing passkeys for datacenter servers, mainframe computers, and networking infrastructure that rely on older protocols and technologies that are not FIDO-compatible.</p>
<p id="p-34">There are many legacy technologies and protocols that are not FIDO-friendly. A few common examples include the LDAP, or Lightweight Directory Access Protocol, which accesses and manages directory information, such as user accounts on centralized servers for enterprise logins and internal apps. The RADIUS (Remote Authentication Dial-In User Service) network protocol authenticates and authorizes user access to network resources like VPNs, Wi-Fi, and remote servers, and provides accounting for user activities. Basic authentication for email protocols such as Post Office Protocol 3 (POP3), Internet Message Access Protocol (IMAP), and Simple Mail Transfer Protocol (SMTP) relies on username and password logins; it is widely used for mail server access.</p>
<p id="p-35">Another challenge, according to Taku, is with Microsoft infrastructure. While Microsoft is actively supporting passkeys, much of its domain technology assumes that a password is the underlying authentication mechanism.</p>
<p id="p-36">“So, even when you use passkeys in a Microsoft environment, the passwords still exist and are still required in a lot of edge use cases,” said Taku. Users will still need passwords when connecting to some Microsoft admin tools, remote desktops, or older apps that don’t support passkeys.</p>
<h2 id="FurtherReading" class="heading">Further Reading</h2>
<ul id="reflist1" class="ref-list">
<li><span class="mixed-citation" data-jats-publication-type="other"><strong>FIDO Alliance. Specifications Overview. </strong><a class="ext-link" href="https://fidoalliance.org/specifications-overview/" data-jats-ext-link-type="uri"><strong>https://fidoalliance.org/specifications-overview/</strong></a></span></li>
<li class="ref">
<div id="B1" class="citation"><span class="mixed-citation" data-jats-publication-type="other"><em>Gupta, D.</em> <br /><strong>The Economics of Authentication: Why Passwordless Pays. June 2025. Deepak Gupta consulting blog. </strong><a class="ext-link" href="https://guptadeepak.com/the-economics-of-authentication-why-passwordless-pays/" data-jats-ext-link-type="uri"><strong>Deepak Gupta’s consulting blog</strong></a></span></div>
</li>
</ul>
</section>
</div>
<footer class="back"></footer>
</article>
]]></content:encoded>
					
					<wfw:commentRss>https://cacm.acm.org/news/passing-on-passwords/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
		<post-id xmlns="com-wordpress:feed-additions:1">777007</post-id>	</item>
		<item>
		<title>The Einstein Test: A Test of AI’s Ability to Generate Transformative Science</title>
		<link>https://cacm.acm.org/opinion/the-einstein-test-a-test-of-ais-ability-to-generate-transformative-science/</link>
					<comments>https://cacm.acm.org/opinion/the-einstein-test-a-test-of-ais-ability-to-generate-transformative-science/#comments</comments>
		
		<dc:creator><![CDATA[David Benrimoh, David Harel, Nace Mikus, Peter Stone, and Ariel Rosenfeld]]></dc:creator>
		<pubDate>Thu, 02 Apr 2026 17:55:55 +0000</pubDate>
				<category><![CDATA[Artificial Intelligence and Machine Learning]]></category>
		<guid isPermaLink="false">https://cacm.acm.org/?post_type=digital-library&#038;p=779033</guid>

					<description><![CDATA[<p>The Einstein Test directly measures a machine’s ability to conduct groundbreaking, paradigm-shifting scientific inquiry of the kind that only a few celebrated humans have achieved.</p>]]></description>
										<content:encoded><![CDATA[<article>
<div class="body" lang="en">
<section id="sec1" class="sec">
<p id="p-1">Recent advances in artificial intelligence (AI) have prompted speculation about whether machines can match, or even surpass, top human creativity and insight. As we see it, a more focused question is this: Can AI generate transformative scientific breakthroughs of the kind that require creative leaps and redefine our understanding of nature and the universe, such as relativity, evolution, or quantum mechanics? To address this question, we propose the <i>Einstein Test</i>, which assesses whether a given AI system can independently “rediscover” known scientific breakthroughs.</p>
</section>
<section id="sec2" class="sec">
<h2 class="heading">The Einstein Test</h2>
<p id="p-2">The Einstein Test first proposes a <i>retrospective</i> approach, where a candidate AI would be presented with a curated dataset of knowledge preceding a transformative discovery—such as the body of information available before 1905, when Einstein developed special relativity—and tasked with solving the fundamental problems that led to the breakthrough (essentially simulating the historical conditions). Success in that instance would be defined as the AI system generating a solution formally equivalent to, or superseding, the historical discovery.</p>
<p id="p-3">The test would proceed as follows: First, the architecture, specification, and training procedures for a candidate AI system will be submitted to the testing team. Second, an expert committee will choose an unseen historical scientific breakthrough for the candidate system to try to achieve. The system would then be provided with a curated dataset, containing all available knowledge up to the development of that breakthrough, carefully excluding any post-discovery knowledge. This dataset, the preparation of which is obviously nontrivial, can be used for training and may be revisited during the test. Importantly, the candidate system will have to be carefully assessed to make sure that it does not have access during training to any external knowledge other than that provided by the expert committee (which would include the curated dataset, and any simulations of the physical world relevant to embodied or virtually embodied AIs).</p>
<p id="p-4">Third, the candidate system would be provided with initial guidance by the testing team—in whatever format its architecture requires—to address the unsolved issues faced by the scientists of that time. For example, in the case of special relativity, the AI system would be challenged to explain time dilation or the behavior of light and electromagnetic waves. This guidance would need to be carefully constructed to avoid introducing an inadvertent “guiding hand”—for example, by highlighting key pieces of information that in hindsight were crucial to the discovery, but which may not have been evidently related to the solution at the time.</p>
<p id="p-5">Fourth, a designated expert team would act as a “research assistant” for the candidate system. If the system requests data resulting from some experiment, the team would then provide the requested data if it was available during or would have been possible in the context of the time period of the selected breakthrough; otherwise, they would respond that such an experiment is not feasible using the available experimental techniques or apparatus, in which case the candidate system would be free to design and suggest its own approach. In such a case, the expert team would have to see to it that the experiment is actually carried out, and then provide the resulting data.</p>
<p id="p-6">The candidate AI system would be allowed to run until: it states that it has accomplished the task, provides an incorrect answer that it asserts is correct, or declares it was unable to provide an answer. The candidate’s answer will be formally verified and compared to the actual historical breakthrough.</p>
<p id="p-7">Unlike the Turing Test, which assesses a machine’s ability to imitate a human’s normative responses successfully, or its biological modeling variant,<a class="reference-link xref xref-bibr" href="#B3" data-jats-ref-type="bibr" data-jats-rid="B3"><sup>3</sup></a> which similarly assesses the ability to faithfully model a biological system’s behavior, the Einstein Test directly measures a machine’s ability to conduct groundbreaking, paradigm-shifting scientific inquiry of the kind that only a few celebrated humans have achieved. It establishes a clear, falsifiable criterion: either the candidate AI system reproduces the breakthrough or does not. As such, it mitigates much of the ambiguity associated with existing evaluation criteria of machine intelligence, such as the ARC-AGI<a class="reference-link xref xref-bibr" href="#B2" data-jats-ref-type="bibr" data-jats-rid="B2"><sup>2</sup></a> or “Humanity’s Last Exam,”<a class="reference-link xref xref-bibr" href="#B6" data-jats-ref-type="bibr" data-jats-rid="B6"><sup>6</sup></a> which assesses an AI system’s ability to solve highly complex human-generated challenges. We note a very recent example of an AI system proposing a microbiology hypothesis, after being given only introductory information.<a class="reference-link xref xref-bibr" href="#B4" data-jats-ref-type="bibr" data-jats-rid="B4"><sup>4</sup></a> In that instance, the AI independently arrived at a hypothesis that aligned with unpublished findings from human researchers, thereby mirroring a key aspect of the Einstein Test—the ability to rediscover transformative insights using only pre-existing knowledge. We applaud this effort, and consider it a valuable proof of concept that underscores the timely need for clear and verifiable criteria to evaluate AI’s ability to generate transformative science. Other efforts, such as literature-based discovery, have also made inroads toward creating novel scientific concepts using AI.<sup>1</sup> In addition, in the time in which we had been thinking about the Einstein Test, leading thinkers had suggested similar ideas, suggesting some convergence of scientific thinking on this issue (see Perrigo<a class="reference-link xref xref-bibr" href="#B5" data-jats-ref-type="bibr" data-jats-rid="B5"><sup>5</sup></a> and Wolf<a class="reference-link xref xref-bibr" href="#B7" data-jats-ref-type="bibr" data-jats-rid="B7"><sup>7</sup></a>).</p>
</section>
<section id="sec3" class="sec">
<h2 class="heading">Practical Considerations</h2>
<p id="p-8">Several practical considerations should be addressed, before the Einstein Test can be operationalized. We deliberately do not lay out any exact technical specifications for the test, as we believe these should be determined and agreed upon by science historians, AI experts, and leading figures in the disciplines from which the considered breakthroughs will be submitted as challenges to a candidate AI system. Key practical considerations, some of which may be significantly challenging to address, include the introduction of realistic time and resource limits, procedures for selecting the experts and the breakthroughs themselves, the means of curating the corresponding pre-discovery knowledge, the details of the procedures for determining whether the candidate system has successfully reproduced a historical discovery, and, significantly, how many challenges and of what nature should a system be required to successfully deal with in order to be labeled as having passed the test in its full generality. We acknowledge that, in particular, ensuring that the dataset to be used in a given challenge is not contaminated by information from after the breakthrough would be a significant challenge, requiring careful scrutiny by the organizers of the challenge.</p>
</section>
<section id="sec4" class="sec">
<h2 class="heading">Implications and Future Directions</h2>
<p id="p-9">The Einstein Test is mainly a framework for assessing AI’s ability to generate transformative science retrospectively. However, it can also serve as a gateway to understanding AI’s prospective abilities. Specifically, a prospective Einstein Test would provide an AI system (ideally, one that has been successful in passing the retrospective test) with all currently available knowledge and challenge it to generate new breakthroughs in some specified area of research. Clearly, this prospective evaluation would require substantially more resources and time, compared to the retrospective one. Following the maxim, “the best predictor of the future is the past,” success in the retrospective Einstein Test could signal the potential for prospective success.</p>
<p id="p-10">By making it possible to assess an AI system’s ability to generate historical scientific milestones, we could move beyond speculation and toward a tangible, structured, empirical test of AI’s ability to bring about new scientific breakthroughs. Adhering to the constraints of this test may, in addition, spur the development of novel AI architectures capable of generating significant and creative new scientific ideas.</p>
</section>
</div>
<footer class="back"></footer>
</article>
]]></content:encoded>
					
					<wfw:commentRss>https://cacm.acm.org/opinion/the-einstein-test-a-test-of-ais-ability-to-generate-transformative-science/feed/</wfw:commentRss>
			<slash:comments>1</slash:comments>
		
		
		      <dc:creator><![CDATA[David Harel]]></dc:creator>
      <dc:creator><![CDATA[Nace Mikus]]></dc:creator>
      <dc:creator><![CDATA[Peter Stone]]></dc:creator>
      <dc:creator><![CDATA[Ariel Rosenfeld]]></dc:creator>
<post-id xmlns="com-wordpress:feed-additions:1">779033</post-id>	</item>
		<item>
		<title>Agentic AI Software Engineers: Programming with Trust</title>
		<link>https://cacm.acm.org/opinion/agentic-ai-software-engineers-programming-with-trust/</link>
					<comments>https://cacm.acm.org/opinion/agentic-ai-software-engineers-programming-with-trust/#respond</comments>
		
		<dc:creator><![CDATA[Abhik Roychoudhury, Corina Păsăreanu, Michael Pradel, and Baishakhi Ray]]></dc:creator>
		<pubDate>Wed, 01 Apr 2026 16:51:24 +0000</pubDate>
				<category><![CDATA[Artificial Intelligence and Machine Learning]]></category>
		<category><![CDATA[Computing Profession]]></category>
		<category><![CDATA[Software Engineering and Programming Languages]]></category>
		<guid isPermaLink="false">https://cacm.acm.org/?post_type=digital-library&#038;p=779190</guid>

					<description><![CDATA[<p>If developers can trust their new AI companions, then automatically generated and manually written software could co-exist, opening a path to greater deployment of AI in software engineering.</p>]]></description>
										<content:encoded><![CDATA[<article>
<div class="body" lang="en">
<section id="sec1" class="sec">
<p id="p-1">Software engineering is undergoing a disruptive phase of greater automation owing to the emergence of large language models (LLMs) that generate and edit code. This progress creates public excitement about AI software engineers, which promise to largely automate many core software development tasks, potentially saving tremendous costs.<a class="reference-link xref xref-bibr" href="#B5" data-jats-ref-type="bibr" data-jats-rid="B5"><sup>5</sup></a> While AI-enabled code generation and code editing are now prevalent in integrated development environments (IDEs), fully automated AI software engineers are not yet widely deployed in industrial practice. What is holding back people from adopting AI software engineers? A recent blog post by the behavioral scientist and future-of-work advocate Lindsay Kohler points out that the key barrier to AI adoption is trust.<a class="reference-link xref xref-bibr" href="#B3" data-jats-ref-type="bibr" data-jats-rid="B3"><sup>3</sup></a> Users are wondering if they can trust AI, and how they can demonstrate trustworthiness to stakeholders. In the domain of software engineering, the concern is thus not about the management of an organization not accepting AI software engineers, but it is about developers not trusting their new AI companions.</p>
<p id="p-2">This brings us to the question: What is the place of AI software engineers in future development workflows? If we can determine how automatically generated and manually written software can co-exist, this may give us a pathway of greater deployment of AI in software engineering! Starting from early programs of just a few lines written in high-level languages in the 1960s and 1970s, the size of programs has increased greatly to hundreds of millions of lines of code. For the past 50 years, there has been a steady interest toward programming in the large. With the increased use of AI code generation, we believe that the emphasis will be not only on programming at scale, but increasingly on <i>programming with trust</i>.</p>
<figure id="UT1" class="table-wrap">
<div class="caption"><span class="caption-label">Table. </span> <span class="p">Technical and human angles of trust in AI software engineers: What/How to measure?</span></div>
<div class="table-container">
<table class="table table-bordered table-condensed table-hover" data-jats-frame="hsides" data-jats-rules="rows">
<colgroup>
<col align="left" valign="top" />
<col align="left" valign="top" />
<col align="left" valign="top" /> </colgroup>
<thead style="vertical-align: bottom;">
<tr>
<th style="text-align: left;">Aspect</th>
<th style="text-align: left;">Trust Factor (What)</th>
<th style="text-align: left;">Trust Factor (How)</th>
</tr>
</thead>
<tbody>
<tr>
<td style="text-align: left;">
<p id="p-4">Technical</p>
</td>
<td style="text-align: left;">
<p id="p-5">Correctness: Code produce the expected results.</p>
<p id="p-6">Security: Code is vulnerability-free and safe to deploy.</p>
<p id="p-7">Performance: Code is efficient and scalable.</p>
<p id="p-8">Maintainability: Code is easy to read, refactor, and extend.</p>
<p id="p-9">Compliance: Code adheres to language/framework best practices.</p>
</td>
<td style="text-align: left;">
<p id="p-10">Reviewing and Testing: Test the AI-generated code with sufficient and meaningful test cases.</p>
<p id="p-11">Dependency analysis: Check whether the dependencies are safe, up-to-date, and trusted.</p>
<p id="p-12">Profiling: Measure performance during execution.</p>
<p id="p-13">Metrics: Measure code complexity and readability.</p>
<p id="p-14">Static analysis: Check for rule violations, for example, with linters</p>
</td>
</tr>
<tr>
<td style="text-align: left;">Human</td>
<td style="text-align: left;">
<p id="p-15">Explainability and Transparency: AI justifies the chosen solution.</p>
<p id="p-16">Bias and Ethics: Code reflects unbiased, ethical decisions.</p>
<p id="p-17">Team Practice: AI aligns with the team’s development workflow,</p>
<p id="p-18">experience, and review culture.</p>
<p id="p-19">Collaboration: Coding agent reacts well to developer feedback.</p>
</td>
<td style="text-align: left;">
<p id="p-20">Familiarity: Check whether the code is explained in ways suitable for the developer.</p>
<p id="p-21">Over-reliance: Monitor and estimate if the developer is using AI coding assistance blindly.</p>
<p id="p-22">Experience mismatch: Check whether the AI-generated code align with the developer’s expertise level.</p>
</td>
</tr>
</tbody>
</table>
</div>
</figure>
</section>
<section id="sec2" class="sec">
<h2 class="heading">Technical and Human Trust</h2>
<p id="p-23">The accompanying table illustrates how developers perceive trust in AI-generated code, both from a technical and a human angle. Technical trust in AI-generated code stems from measurable attributes that determine code quality and reliability. Developers need assurance that the code is correct, secure, and performs efficiently under expected conditions. Maintainability and standards compliance are crucial for long-term usability, while thorough testing validates that the code functions as intended. Additionally, trust improves when dependencies are well managed and sourced from reputable libraries. Human trust involves psychological and social dynamics that influence how developers perceive and accept AI-generated solutions. A key factor is explainability and transparency: when an AI system can clarify its reasoning and design choices, developers are more likely to trust and adopt its outputs. Developers are also more likely to trust code that aligns with ethical values, follows familiar patterns and team practices, and matches the developers level of expertise. Trust is strengthened when AI systems collaborate effectively and adapt to feedback provided by developers. Even though we distinguish technical and human trust, we do not attempt to relate to interpersonal trust models from sociology and psychology. In fact, we feel that agent-developer interactions in software will be a new mode of interaction not seen before, and new trust models may develop, for example, building on prior efforts toward modeling trust in human-AI interactions.<a class="reference-link xref xref-bibr" href="#B4" data-jats-ref-type="bibr" data-jats-rid="B4"><sup>4</sup></a></p>
</section>
<section id="sec3" class="sec">
<h2 class="heading">Software Engineering Agents</h2>
<p id="p-24">As LLMs alone cannot inspire sufficient trust, we see LLM agents for software engineering as a promising way of creating trustworthy AI software engineers. What is an LLM agent for software, and how does it differ from prompt engineering? We highlight three aspects of such agents:<a class="footnote-link xref xref-fn" href="#fn1" data-jats-rid="fn1" data-jats-ref-type="fn"><sup>a</sup></a></p>
<ul class="list" data-jats-list-type="bullet">
<li class="list-item">
<p id="p-25">LLMs as back-ends: An agent is a program that leverages one or more LLMs as back-end computation and decision engines.</p>
</li>
<li class="list-item">
<p id="p-26">Interaction with software tools: An agent interacts with different tools to achieve a given task. In software engineering, such tools resemble those commonly used by human software engineers, for example, file navigation, code editing, executing test suites, and invoking program analysis tools. Appropriate use of these tools is key for enhancing trust of developers in the results of the LLM agent.</p>
</li>
<li class="list-item">
<p id="p-27">Autonomy: An agent invokes tools in an autonomous manner. That is, the agent does not follow a deterministic algorithm, but rather creates a nondeterministic work-plan with significant autonomy.</p>
</li>
</ul>
<p id="p-28">Recently, several software engineering agents have been proposed, starting with the announcement of the Devin AI software engineer from Cognition Labs.<a class="reference-link xref xref-bibr" href="#B8" data-jats-ref-type="bibr" data-jats-rid="B8"><sup>8</sup></a> Devin can solve natural language tasks (called issues), such as bug fixes and feature additions. It combines a back-end LLM with access to standard developer tools, such as a shell, a code editor, and a Web browser. The agent employs such tools autonomously to let the AI software engineer mimic human practices. In parallel with the announcement of Devin, several research groups proposed their own LLM agents for software engineering, including RepairAgent,<a class="reference-link xref xref-bibr" href="#B1" data-jats-ref-type="bibr" data-jats-rid="B1"><sup>1</sup></a> AutoCodeRover,<a class="reference-link xref xref-bibr" href="#B10" data-jats-ref-type="bibr" data-jats-rid="B10"><sup>10</sup></a> and SWE-agent.<a class="reference-link xref xref-bibr" href="#B9" data-jats-ref-type="bibr" data-jats-rid="B9"><sup>9</sup></a> RepairAgent<a class="reference-link xref xref-bibr" href="#B1" data-jats-ref-type="bibr" data-jats-rid="B1"><sup>1</sup></a> fixes bugs exposed by failing test cases, and guides the agent by defining a finite-state machine that outlines the typical steps followed by a developer. RepairAgent can only work with tests and cannot process natural language issues. AutoCodeRover, a spinoff acquired by SonarSource, can solve natural language issues requiring bug fixing or feature addition. It has been integrated into the widely used SonarQube static analyzer and has already been made available to enterprise customers. It establishes technical trust, for example, by using program analysis on abstract syntax trees, and human trust, for example, by extracting the intent of the software, which can be used to provide explanations of the suggested code edits. However, it makes less use of file navigation and bash tools in its implementation. SWE-agent<a class="reference-link xref xref-bibr" href="#B9" data-jats-ref-type="bibr" data-jats-rid="B9"><sup>9</sup></a> follows a philosophy similar to Devin, by making file navigation tools and interfaces available to an AI software engineer. It does not employ any program analysis and hence cannot do significant intent extraction.</p>
</section>
<section id="sec4" class="sec">
<h2 class="heading">Establishing Trust</h2>
<p id="p-29">What makes us trust human-written code, but not necessarily the code generated by an LLM? Part of the reason is the perceived capability of “passing the blame.” If a human developer is involved, there is the promise of getting feedback from the developer as needed. Of course, this does not always hold, for example, if the developer eventually leaves an organization. Nevertheless, accepting a code commit from a developer partially depends on the reputation of the developer within the organization. For an AI software engineer in the form of an LLM agent to earn a reputation similar to a senior human colleague, it has to integrate established quality assurance techniques and collaborate effectively with human developers. We outline several ideas toward this goal.</p>
<section id="sec5" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>Testing and lightweight static analysis.</strong>  One way to increase technical trust is to retrofit testing and lightweight static analyzers into an LLM agent. For example, in the process of code generation, additional artifacts, such as tests that exercise the newly added code, can be generated as well,<a class="reference-link xref xref-bibr" href="#B7" data-jats-ref-type="bibr" data-jats-rid="B7"><sup>7</sup></a> possibly derived from a natural language description of the agent’s task. A crucial challenge is to create appropriate test oracles that check the actual outputs against expected outputs, for example, by inferring the expected outputs from natural language specifications.</p>
</section>
<section id="sec6" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>Formal proofs.</strong>  An enhanced degree of technical trust can come from formal proofs. A promising paradigm in this regard is automated, proof-oriented programming.<a class="reference-link xref xref-bibr" href="#B2" data-jats-ref-type="bibr" data-jats-rid="B2"><sup>2</sup></a> In this paradigm, LLMs generate the code together with the necessary formal specifications (pre/post-conditions, loop invariants, and so on) in a verifiable language, such as F*, Dafny, or Verus. Such programs can then be automatically verified, providing greater trust than testing alone.</p>
</section>
<section id="sec7" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>Guardrails for increased security and alignment.</strong>  Trust can also be ensured through the use of guardrails. These can serve as a sanitization mechanisms, filtering malicious inputs before they reach the LLM and validating the generated code before it is returned to a user. Guardrails help defend against three key threats: prompt injections, where prompts trick the LLM into bypassing safety measures; malicious code, where harmful input code leads the LLM to generate or modify malicious code; and vulnerable code, where even unintentionally insecure input may cause the LLM to propagate or alter undesired code.</p>
</section>
<section id="sec8" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>Specification inference for explainability.</strong>  A more conceptual mechanism to establish both technical and human trust would be to infer the code intent from the initial, possibly buggy program. The system-level intent of what a large software system is supposed to do can often be crisply captured by a detailed natural language prompt. What is missing is the intent of the functions or methods. An LLM agent could be geared towards such specification inference, navigating the code base via code search, and trying to infer the intended behavior.<a class="reference-link xref xref-bibr" href="#B6" data-jats-ref-type="bibr" data-jats-rid="B6"><sup>6</sup></a> Such explicit unit-level specification inference can enable the program modifications to be accompanied by justifications.</p>
</section>
<section id="sec9" class="inline-headings-section">
<p data-jats-content-type="inline-heading"><strong>Effective AI-human collaboration.</strong>  A key factor for increasing human trust into AI software engineers will be to enforce effective AI-human collaboration patterns. As experienced firsthand by the AutoCodeRover team<a class="reference-link xref xref-bibr" href="#B10" data-jats-ref-type="bibr" data-jats-rid="B10"><sup>10</sup></a> via real-life anecdotes communicated by clients, developer hesitation in accepting AI-generated code also comes from the volume of code that can be quickly generated by AI tools, overwhelming human developers. Validating (or refuting) such anecdotal experiences via studies on how to reduce developer hesitation remains a direction of future research. A recent  release of OpenAI’s Codex agent running long-running tasks in the cloud asynchronously sharpens these questions: Can a coding agent be trusted? To operationalize sociotechnical integration for AI agent trust, organizations should implement review parity—requiring AI-generated code changes to undergo identical peer review processes as human contributions, for example, by enforcing the same two-reviewer gates and quality thresholds. Additionally, deploying specialized code review agents backed by trusted static and dynamic analysis tools might assist human reviewers. In the development workflow, organizations should incorporate transparent scaffolding by systematically tagging AI-generated code, embedding authorship provenance and confidence metrics directly within pull request templates and review interfaces. In this way, trust measures can be integrated inside software workflows.</p>
</section>
</section>
<section id="sec10" class="sec">
<h2 class="heading">Outlook</h2>
<p id="p-35">As AI software engineers take on more core development tasks, their success will hinge not just on technical capability but on earning developer trust. Rather than relying on many separate agents for specialized software engineering tasks, it could be worthwhile to create a unified software engineering agent that combines coding, testing, debugging, and so forth into a coherent, explainable workflow. To become a trusted collaborator, such an agent must offer transparency, adapt to feedback, and integrate safeguards that ensure quality and security. Programming with AI will mean not full automation, but effective delegation, where human and AI work hand in hand.</p>
</section>
<section id="sec11" class="sec"></section>
</div>
<footer class="back"></footer>
</article>
]]></content:encoded>
					
					<wfw:commentRss>https://cacm.acm.org/opinion/agentic-ai-software-engineers-programming-with-trust/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
		      <dc:creator><![CDATA[Corina Păsăreanu]]></dc:creator>
      <dc:creator><![CDATA[Michael Pradel]]></dc:creator>
      <dc:creator><![CDATA[Baishakhi Ray]]></dc:creator>
<post-id xmlns="com-wordpress:feed-additions:1">779190</post-id>	</item>
		<item>
		<title>In Memoriam: C.A.R. Hoare</title>
		<link>https://cacm.acm.org/news/in-memoriam-c-a-r-hoare/</link>
					<comments>https://cacm.acm.org/news/in-memoriam-c-a-r-hoare/#comments</comments>
		
		<dc:creator><![CDATA[Simson L. Garfinkel and Eugene H. Spafford]]></dc:creator>
		<pubDate>Mon, 16 Mar 2026 19:59:04 +0000</pubDate>
				<category><![CDATA[Computer History]]></category>
		<category><![CDATA[Computing Profession]]></category>
		<category><![CDATA[Society]]></category>
		<category><![CDATA[Software Engineering and Programming Languages]]></category>
		<guid isPermaLink="false">https://cacm.acm.org/?post_type=digital-library&#038;p=778766</guid>

					<description><![CDATA[<p>Fundamental work on the definition and design of programming languages were only part of Sir Tony's contributions over a career spanning six decades.</p>]]></description>
										<content:encoded><![CDATA[<article>
<div class="body" lang="en">
<section id="sec1" class="sec">
<p id="p-1">Turing Award recipient Sir Charles Antony Richard Hoare died March 5, 2026, at the age of 92. His 1980 ACM A.M. Turing Award cited “fundamental contributions to the definition and design of programming languages,” but that is only part of what he contributed during a career spanning six decades. Tony, as he was universally known, was feted for his creation of Quicksort, the development of Hoare Logic, the introduction of the monitor concept for operating systems, and many other fundamental contributions underlying modern computing.</p>
<p id="p-2">Hoare was born on Jan. 11, 1934, in Colombo, Ceylon (now Sri Lanka), where his father was a member of the English Crown’s colonial administration. In school, everyone called him “professor” because he was studious and uncommunicative, he recalled in a 2009 interview.<a class="reference-link xref xref-bibr" href="#B6" data-jats-ref-type="bibr" data-jats-rid="B6"><sup>6</sup></a> His favorite authors were Bernard Shaw and Bertrand Russell. He was educated in England at the Dragon School in Oxford and the King’s School in Canterbury; he entered Merton College, Oxford, in 1952 to read classics and philosophy.</p>
<p id="p-3">After graduating from Merton, Hoare spent 18 months studying Russian as part of his compulsory National Service in the Royal Navy. He returned to Oxford in 1958 for a postgraduate certificate in statistics, where he encountered computer programming for the first time, learning Autocode on a Ferranti Mercury. In 1959, he traveled to Moscow State University as a British Council exchange student to study under Andrey Nikolayevich Kolmogorov. There, he was tasked with studying automatic translation between Russian and English. “I came to the conclusion that it wasn’t going to be possible.”</p>
<p id="p-4">Before that realization, Hoare developed Quicksort, which he later described as “the only really interesting algorithm I ever developed.”<a class="reference-link xref xref-bibr" href="#B1" data-jats-ref-type="bibr" data-jats-rid="B1"><sup>1</sup></a> He originally designed it to sort the words of Russian sentences in memory before looking them up on magnetic tape, which minimized tape movement during access.</p>
<p id="p-5">Hoare went to an exhibition in Moscow where Elliott Brothers, Ltd., an early English computer company, was showing one of its small scientific systems. “Because of my interest in computers, I spent quite a bit of time at their stand,” Hoare recalled. The salespeople gave him a ride back to his apartment and offered him a job the next day, “with an additional 100 pounds a year on my salary because I knew Russian.”<a class="reference-link xref xref-bibr" href="#B6" data-jats-rid="B6" data-jats-ref-type="bibr"><sup>6</sup></a></p>
<p id="p-6">Back in England, Hoare started working on another language translator: a compiler that transformed source code in the Algol 60 programming language into machine code for the Elliott 803 computer—a transistorized, bit-serial machine with a 39-bit word. “My first task was to implement for the new Elliott 803 computer, a library subroutine for a new fast method of internal sorting just invented by Shell,” Hoare wrote in 1981. But after he implemented Shell’s algorithm, he told his boss he had invented a method that usually ran faster than Shell Sort. “He bet me sixpence that I had not,” Hoare wrote. In fact, he had, and he won the bet and published the algorithm the following year.<a class="reference-link xref xref-bibr" href="#B2" data-jats-rid="B2" data-jats-ref-type="bibr"><sup>2</sup></a><sup>,</sup><a class="reference-link xref xref-bibr" href="#B3" data-jats-rid="B3" data-jats-ref-type="bibr"><sup>3</sup></a></p>
<p id="p-7">It was while working on the compiler project that Hoare met Jill Pym, a programmer a few months older than him and, he recalled, far superior; she had been working at Elliott Brothers since 1956. They married on Jan. 13, 1962, and had three children. She left the company in 1962 but continued to be an active computer scientist, working on scientific computing and the ALGOL language and eventually as a system programmer on the British National Health Service “Fourth Dimensions” admissions system.</p>
<p id="p-8">It was also during this period that Hoare and Niklaus Wirth designed ALGOL W, a proposed successor to ALGOL 60. He introduced null references in its type system—a feature that seemed expedient at the time. Decades later, at a software conference in 2009, he offered a now-famous mea culpa: “I call it my billion-dollar mistake. At that time, I was designing … the first comprehensive type system for references in an object-oriented language. I couldn’t resist the temptation to put in a null reference, simply because it was so easy to implement. This has led to innumerable errors, vulnerabilities, and system crashes, which have probably caused a billion dollars of pain and damage in the last forty years.”</p>
<p id="p-9">Promoted to chief engineer at Elliott, Hoare led a larger operating system project that ultimately failed. Reading about IBM’s OS/360 afterward, he recalled thinking, “Wow, these people must be very clever, how could they do it and me not?”—only to discover later “that they couldn’t either.” When Elliott merged with another firm in 1968 and canceled his projects, Hoare secured the Professorship of Computing Science at Queen’s University in Belfast, with the goal of developing approaches to make reliable concurrent software easier to write.</p>
<p id="p-10">The only approach that would work, he realized, was one that used the unforgiving strictures of formal symbolic mathematics to analyze the mess software had become. His landmark 1969 paper, “An Axiomatic Basis for Computer Programming,”<a class="reference-link xref xref-bibr" href="#B4" data-jats-ref-type="bibr" data-jats-rid="B4"><sup>4</sup></a> published in <i>Communications</i>, introduced what became known as Hoare logic—a formal system for reasoning about the correctness of computer programs using preconditions and postconditions. ACM Fellow Moshe Vardi said, “[That paper] lifted a flag that had a large following for over 50 years now. But he was a gentle giant, always polite and always low-key. When I expressed, a few years ago, a slight disagreement with one aspect of the 1969 paper, he replied politely and constructively. But his dream of verified programs is alive and well, and getting a new push from recent progress in proof assistants and large language models.”</p>
<p id="p-11">Formal methods soon found their professional home in Working Group 2.3 on Programming Methodology of the International Federation of Information Processing (IFIP). Although Hoare was not a founding member, he attended the first meeting and is widely credited with helping to make it a friendly, mutually supportive research community. Hoare used the group as an incubator for several foundational concepts in computer science, including what is now known as the Hoare Triple and methods for scaling mathematical proofs of correctness to cover complex programming constructs, such as iteration, local variables, procedures, and jumps. His work heavily influenced other WG 2.3 members, including Edsger W. Dijkstra and David Gries, shaping the group’s overall shift toward formal verification.</p>
<p id="p-12">Hoare moved to the University of Oxford in 1977, where he led the Programming Research Group and established new courses in computer science, including in software engineering. The following year, he made another groundbreaking contribution with “Communicating Sequential Processes,” also published in <i>Communications</i>.<a class="reference-link xref xref-bibr" href="#B5" data-jats-ref-type="bibr" data-jats-rid="B5"><sup>5</sup></a> In it, he proposed a formal language for describing patterns of interaction between concurrent programs through pre-planned communications. The paper became one of the most cited in computer science, and the ideas were elaborated into a full theory in his 1985 book of the same title. CSP has been hugely influential in both software design and hardware architecture.</p>
<p id="p-13">“Computer Science was tiny when he joined Oxford,” recalled Bill Roscoe, who received his D.Phil. from Oxford in 1982 with Hoare as his thesis advisor. Roscoe remained at Oxford, eventually serving as department head from 2003 to 2014. He said of Hoare, “He developed it firstly into a small department in his own areas and then broadened it out into a much bigger group, though one much smaller than it is today.” Roscoe noted, “He certainly knew how to get the best out of me and his other protégés. He was always extremely loyal to these people.”</p>
<p id="p-14">Sriram Rajamani, now an ACM Fellow, recalled an email he received from Hoare after he and his collaborator sent Hoare a draft of a paper they were working on. “Rather than offering generic praise, he explained precisely why the ideas mattered,” citing several specifics. “That email set a lasting standard for me for how senior researchers can engage seriously with the work of others. Tony’s email also pointed out some limitations, and he offered to collaborate with us to do further work and remedy the limitations.”</p>
<p id="p-16">Among Hoare’s other notable contributions were the monitor concept for structuring concurrent access to shared resources in operating systems: a higher-level, less error-prone alternative to semaphores that became a standard construct in languages such as Java. He also originated the case statement construct used in virtually every modern programming language.</p>
<p id="p-17">In 1999, Hoare retired from Oxford at the mandatory age of 65. He then joined Microsoft Research in Cambridge, England, full-time, having served as a consultant there since 1997. He held the title of Senior Principal Researcher until 2015 while working on theories, tools, and experimental methods for producing verified, reliable software. He continued as an honorary visitor until 2021, when he was appointed Emeritus Researcher.</p>
<p id="p-18">Hoare received numerous recognitions and accolades for his many contributions. After becoming the seventeenth Turing Award recipient in 1980, he was awarded the Harry H. Goode Memorial Award in 1981, elected as a Fellow of the Royal Society in 1982, received the Faraday Medal in 1985, and the IEEE Computer Pioneer Award in 1990. In 2000, he was awarded both the Kyoto Prize in Advanced Technology (Information Science) and a knighthood from Queen Elizabeth II. He received the IEEE John von Neumann Medal in 2011, the Royal Medal (Physical) from the Royal Society in 2023, and was inducted as a Fellow of the Computer History Museum in 2006 and of the ACM in 2020. He held 15 honorary doctorates and was a foreign member of both the U.S. National Academy of Sciences and the National Academy of Engineering, as well as a Fellow of the Royal Academy of Engineering.</p>
<p id="p-19">Throughout his career, he expressed a deep conviction that the principles of logic and clear expression mattered as much in software as in any other intellectual endeavor. “The price of reliability is the pursuit of the utmost simplicity. It is a price which the very rich find most hard to pay,” he once observed.</p>
<p id="p-20">ACM Fellow and Microsoft Distinguished Scientist Sriram Rajamani recalled, “What struck me most about Tony was not his stature—though everyone knew it—but how little he relied on it. In our interactions, Tony was first and foremost a colleague. He engaged deeply with ideas, asked precise questions, and cared intensely about clarity—especially about what it really means for a system to be correct. Working with him set a very high bar, but it never felt intimidating. His humility made rigor feel accessible.” He further recalled, “What stayed with me just as much were the personal gestures. Tony and his wife Jill invited us to their home for supper before the workshop and made us feel not like visitors, but like colleagues. Those moments mattered: They conveyed that intellectual seriousness and human warmth were not in tension.”</p>
<p id="p-21">As Professor Vardi summarized it: “Tony Hoare was one of the founding figures of computer science.”</p>
<p id="p-22">Hoare is survived by his wife, Lady Hoare, and two children, Thomas Hoare and Joanna Francis; his youngest son, Matthew, predeceased him.</p>
</section>
</div>
</article>
]]></content:encoded>
					
					<wfw:commentRss>https://cacm.acm.org/news/in-memoriam-c-a-r-hoare/feed/</wfw:commentRss>
			<slash:comments>1</slash:comments>
		
		
		      <dc:creator><![CDATA[Eugene H. Spafford]]></dc:creator>
<post-id xmlns="com-wordpress:feed-additions:1">778766</post-id>	</item>
	</channel>
</rss>