AI is shaping decisions across society quietly and at scale. RAD Labs is dedicated to understanding and addressing the risks of gradual disempowerment in AI-augmented decision-making, with the goal of preserving human agency in the face of rapid technological change.
In order to understand the failure modes of human-AI collaboration on decision-making, it is crucial to clarify the contributions of both humans and models, and how they shift when messy real-world conditions push the system to its limits.
Develop quantitative metrics that warrant early warnings when various bits of human-AI collaboration start to break down, and understand how specific pain points require different interventions.
RAD research isn't meant to stay in the lab but iteratively improve the tool and contribute to public knowledge by sharing benchmarks, safeguards, and data-backed policy recommendations with the wider academic community.
Ever caught yourself asking AI for recommendations on decisions you used to make on your own? You're not alone. Human societies outsource judgment calls to AI at an unprecedented pace. Continuing in this direction without rigorous, actionable monitoring of AI-assisted decision-making risks catastrophic outcomes.
Most people who use AI at work don't verify its process, and most processes that go unverified result in errors at magnitudes current safeguards are not designed to handle. Each overlooked error puts the next decision at risk. Now imagine that decision is a medical diagnosis. An investment. A military call.
Beyond losing understanding of mechanisms leading to decisions that influence our lives, humanity also risks its ability to meaningfully influence outcomes and ultimately agency over our choices. AI does not need to be misaligned or seek power for this to happen. It's enough for it to continue being widely implemented.
In 2026, we continue being unable to put early warnings and safeguards in place because there are no robust metrics for overreliance. This makes sense—it's a complex problem. Nonetheless, humanity needs metrics to manage the risks of AI-assisted decisions. For that reason, RAD Labs are up for the challenge.
This research agenda tackles a complex measurement problem by breaking it into tractable components. The investigative model is built on three principles:
Both models and humans can contribute to human-AI collaboration failures. This agenda weighs both sides of the equation.
Decisions rarely happen under ideal conditions. This research evaluates human-AI collaboration under time pressure, high stakes, and ambiguity to capture realistic failure modes.
Decision types share structural characteristics regardless of domain. Task-based analysis identifies these patterns to build transferable metrics.
The goal of this research is to define the line between helpful AI augmentation and harmful displacement of human choice in decision-making. Click through sections below to explore the methodology:
Establish a coordinated network aligned on RAD objectives.
Identify gaps most relevant to building a monitoring solution.
Produce early empirical results to guide metric and solution development.
Implement functional monitoring features and test them in a live environment.
Put forward initial evaluations and use ongoing findings to improve the platform.
Expand the platform to cover more institution types using adaptable task types.
Answer more relevant questions to continuously improving the solution and inform safeguards.
Use quantity data from across institution types to develop a formal certification.
Use technical research to inform safeguards development and policy recommendations.
Many ways to get involved!
If you have a background in LLM evaluations, Game Theory, can think rigorously about human-AI interaction, and/or have the skills to scale teams and raise support, you'd be a great fit.
Believe in the cause but can't commit? You can make a huge difference by offering expertise, connections, funding, or simply time to help build momentum. Let's talk.
New here? Join an infosession to hear what's being built, why it matters, and whether there's a role for you.
Next session: TBD.
If you use AI tools in your work, you know which problems are real and which are hype. Help ground RAD research by sharing how AI contributes to—and undermines—decisions in your field.
Looking for a meaningful side project? If you've got research skills and some time, check out this bank full of ready-to-research open questions relevant to RAD objectives.
If this sparked something or you want to explore working together, please reach out to nowe.moore@gmail.com.