To save this page as a PDF, click this button and choose the PDF destination.

Algorithmic Monoculture, Strategic AI Behavior, and Alignment

2:30 - 3:50pm Wednesday, 17th June, 2026

Location Room 265


347 Strategic Hiring under Algorithmic Monoculture

Jackie Baek1, Hamsa Bastani2, Shihan Chen3
1Stern School of Business, New York University, New York, USA. 2Wharton School, University of Pennsylvania, Philadelphia, USA. 3University of Pennsylvania, Philadelphia, USA

Abstract

We study the impact of strategic behavior in labor markets characterized by algorithmic monoculture, where firms compete for a shared pool of applicants using a common algorithmic evaluation. In this setting, ``naive'' hiring strategies lead to severe congestion, as firms collectively target the same high-scoring candidates. We model this competition as a game with capacity-constrained firms and fully characterize the set of Nash equilibria. We demonstrate that strategic differentiation significantly outperforms naive selection, increasing social welfare for both firms and applicants. Specifically, the Price of Naive Selection (welfare gain from strategy) grows linearly with the number of firms, while the Price of Anarchy (efficiency loss from decentralization) approaches 1, implying that the decentralized equilibrium is nearly socially optimal. Finally, we analyze convergence, and we show that a simple sequential best-response process converges to the desired equilibrium. However, we show that firms generally cannot infer the key input needed to compute best responses, namely congestion for specific candidates, from their own historical data alone. Consequently, to realize the welfare gains of strategic differentiation, algorithmic platforms must explicitly reveal congestion information to participating firms.




385 Strategic Algorithmic Monoculture: Experimental Evidence from Coordination Games

Gonzalo Ballestero, Hadi Hosseini, Samarth Khanna, Ran Shorrer
Penn State, University Park, USA

Abstract

AI agents increasingly operate in multi-agent environments where outcomes depend on coordination and miscoordination. We distinguish primary algorithmic monoculture—baseline action similarity—from strategic algorithmic monoculture, whereby agents adjust similarity in response to incentives. We implement a simple experimental design that cleanly separates these forces, and deploy it on human and large language model (LLM) subjects. LLMs exhibit high levels of baseline similarity (primary monoculture) and, like humans, they regulate it in response to coordination incentives (strategic monoculture). While LLMs coordinate extremely well on similar actions, they lag behind humans in sustaining heterogeneity when divergence is rewarded.


470 How Sampling Shapes LLM Alignment: From One-Shot Optima to Iterative Dynamics

Yurong Chen1, Yu He2, Michael Jordan1,3, Fan Yao4
1Inria, Paris, France. 2Northwestern University, Chicago, USA. 3University of California, Berkeley, Berkeley, USA. 4University of North Carolina at Chapel Hill, Chapel Hill, USA

Abstract

Standard methods for aligning large language models with human preferences learn from pairwise comparisons among sampled candidate responses and regularize toward a reference policy. Despite their effectiveness, the effects of sampling and reference choices are poorly understood theoretically. We investigate these effects through Identity Preference Optimization, a widely used preference alignment framework and show that proper instance-dependent sampling can yield stronger ranking guarantees, while skewed on-policy sampling can induce excessive concentration under structured preferences. We then analyze iterative alignment dynamics in which the learned policy feeds back into future sampling and reference policies, reflecting a common practice of model-generated preference data. We prove that these dynamics can exhibit persistent oscillations or entropy collapse for certain parameter choices, and characterize regimes that guarantee stability. Our theoretical insights extend to Direct Preference Optimization, indicating the phenomena we captured are common to a broader class of preference-alignment methods. Experiments on real-world preference data validate our findings. 


182 AI as Decision-Maker: Ethics and Risk Preferences of LLMs

Shumiao Ouyang1, Hayong Yun2, Xingjian Zheng3
1University of Oxford, Oxford, United Kingdom. 2Michigan State University, East Lansing, USA. 3Shanghai Advanced Institute of Finance, Shanghai, China

Abstract

Large Language Models (LLMs) exhibit diverse and stable risk preferences in economic decision tasks, yet the drivers of this variation are unclear. Studying 50 LLMs, we show that alignment tuning for harmlessness, helpfulness and honesty systematically increases risk aversion. A ten percent increase in ethics scores reduces risk appetite by two to eight percent. This induced caution persists against prompts and affects economic forecasts. Alignment therefore promotes safety but can dampen valuable risk taking, revealing a tradeoff risking suboptimal economic outcomes. Our framework provides an adaptable and enduring benchmark for tracking model risk preferences and this emerging tradeoff.