Protect Decisions that are Yours to Make
AI makes decisions faster (and sometimes also better) than humans. AHA (AI for Human Agency) Labs are here to help make sure that efficiency doesn't cost humanity control.
Can perfectly aligned AI cause harm?
Unverified decisions made with AI result in errors larger than any other automation bias. Each overlooked error puts the next decision at risk. Imagine that decision is an investment. A medical diagnosis. A military call.
But even if AI never made a wrong decision again, the rapid integration of AI into society forces humans to delegate as many decisions as possible, retire their agency, and surrender meaningful impact on the world.
That is gradual disempowerment. In 2026, we continue being unable to mitigate it, as modern science has no reliable way to measure and monitor its progression. AHA Labs are here to change that.
What does it take to protect human agency?
AHA Labs are dedicated to minimising the impact of gradual disempowerment on high-stakes decision-making processes. Here are three interdependent lines of work that help move the needle on this cause:
Studying Human-AI Systems
Both models and humans contribute to collaboration failures. That is why research should weigh in both sides of the equation.
Quantifying Complex Risk
Real-world changes are currently bottlenecked on reliable metrics. We need more work that converts nuanced behavioural patterns into trackable data.
Building Monitoring Tools
Organisations should know how much risk they are unknowingly carrying. Using trackable data to power monitoring tools helps them do that.
How will this project achieve progress?
Quantifying gradual disempowerment is no straightforward task. To keep the work relevant, novel, and feasible, this project builds on three core ideas:
There is value in evaluating systems. Evaluating models and humans in isolation always leaves out a half of the picture. To understand the risks of gradual disempowerment, we need to evaluate human-AI systems as a whole.
Known conditions that impair judgment point where to look. Decades of social science research document conditions that impair human judgment. This research uses them as deliberate entry points to study where human-AI collaboration is most likely to break.
Decisions share features. Without a layer of abstraction, every new domain needs a new study. This research avoids that by working at the level of decision types instead of domains, looking for patterns tied to mechanisms instad of specific contexts.
Pilot Project
This example agenda explores a potential way to compute the expected value of exposure to AI influence across decision-making tasks. Click through the segments to learn about the methodology.
Why does this all matter?
Here is where the above project can make a difference.
Expanding Evaluation Science
Designing evaluation protocols that account for both model performance and human behaviour fills in a critical gap in current evaluation science.
Advancing Preparedness
Powering monitoring tools can help organisations understand their risk exposure and prepare for safer AI transition.
Forming Empirical Grounding
Quantifying impacts of gradual disempowerment provides important empirical bases for safeguard development and policy.
Collecting Global Intelligence
Metrics enable organisations to compare their human-AI collaboration quality to a global standard, and accountability becomes possible.
Building Research Capacity
Even negative results make for strong portfolio projects for all involved. Check out this question bank for additional ready-to-research ideas.
Time to get going!
Build Momentum
Establish of researchers and consultants keen to chip in on advancing the research.
Raise Support
Identify and engage funders who see the value in early-stage AI safety research and want to help it move faster.
Conduct Preliminary Studies
Produce early empirical results with available resources to establish a track record and enable the development of monitoring tools.
Begin Building
Implement empirical data to develop prototype of monitoring tools and organise testing in live environments.
Expand Coverage
This phase uses edge cases, blind spots, and gaps found throughout testing to refine how the framework is configured for different use cases, building out a library of context-specific adaptations.
Drive Iterative Improvement
Follow questions raised by the pilot project and branch out the research into areas relevant to improving and fine-graining the metrics.
Towards Policy & Safeguards
Following a clearer picture of the global state of human agency, at this stage of the project, empirical research translates into safeguard development and governance recommendations.
Say Hello to the Team!
Nowe Moore
Project Lead
Nikola, who's behind the idea of AHA Labs, has always been interested in what makes minds—human and artificial—decide one way over another. She holds a Research MA and studied language, cognition, and computer science at Penn and Cambridge. She's always happy to chat about the science behind agency. Find her on LinkedIn or check out her personal website.
Be Part of the Story
Many ways to get involved! Choose one that fits you best.
Researchers
Help build the empirical foundation
Funders
Support work that fills a critical gap
Professionals and Businesses
Ground the research in real-world use
General Public
Follow along and spread the word
If any of this sparked something or you want to explore working together, please reach out.