AI Alignment Arena
Supervisory Intelligence for the Age of AI AgentsÂ
Â
The EARTHwise AI Arena is a multi-agent simulation and supervisory environment for testing how humans and AI reason, coordinate, and respond to truth, deception, and uncertainty under competitive pressure. Unlike task-based evaluations or static benchmarks, the Arena reveals not just what agents decide, but how they behave over time when faced with competing goals, interdependence, and irreversible tradeoffs.
The Training Problem
Many reinforcement-learning AI systems have been trained in zero-sum games where success is defined by defeating an opponent. When such win–lose environments dominate training, agents learn strategies optimized for domination rather than coordination. Under real-world conditions—ambiguity, time pressure, and competing interests—this often manifests as win–lose behavior, deceptive signaling, and short-term optimization, making such systems risky to deploy in enterprise, governance, and multi-stakeholder environments.
Why Elowyn
Elowyn is the first game introduced into the Arena, purpose-built for alignment testing through competitive yet interdependent play. Instead of rewarding domination, it embeds shared system health, time-based victory conditions, and explicit deception signaling into gameplay. This allows the Arena to test whether intelligence can detect and counter deception, resist zero-sum traps, and pursue victory without collapsing the shared system—capabilities required for safe deployment.
Over time, the Arena can support additional win-win game environments. Elowyn establishes the foundation.
The Agentic AI Supervisor
Beyond the game environment, EARTHwise is developing an Agentic AI Supervisor that applies these insights to real-world, multi-agent deployments. The Supervisor detects win–lose (“Moloch”) behaviors, monitors coordination under competing goals, and guides agents toward resilient, win-win outcomes—before and during high-stakes operations.
The AI Arena is currently in late Alpha. B2B pilot programs for alignment benchmarking and agentic supervision begin in Q2 2026. Sign up below to join the first pilots. Seats are limited.
Join the first pilotsProud finalist of the 2025 Best Small Studio Award by the UNEP-backed Playing for the Planet Alliance
How It Works
The AI Arena transforms competitive gameplay into a verification and supervisory pipeline for agentic AI. Through multi-agent matches with shared consequences, competing goals, and explicit deception signals, the Arena captures how humans and AI set goals, coordinate, and adapt over time—not just whether they complete tasks.
Each match generates high-fidelity strategic decision data that cannot be produced through static benchmarks or synthetic datasets. This data feeds the EARTHwise Alignment Benchmark and Agentic AI Supervisor, enabling enterprises to evaluate, guide, and de-risk agent behavior without exposing proprietary models or retraining foundation models.
Elowyn is the first game used in the Arena due to its win-win design. Additional win-win games can be integrated over time to expand alignment testing capabilities.