ACS
☰
ABOUT
•
RESEARCH
•
POSTS
•
TEAM
•
CONTACT
Post Archive
Our Academic Publications
See our Research page
Research Notes
February 13, 2026
Why You Don't Believe in Xhosa Prophecies
February 4, 2026
Post-AGI Economics As If Nothing Ever Happens
January 12, 2026
When Does Competition Lead to Recognisable Values?
January 8, 2026
The Economics of Transformative AI
November 10, 2025
Ontology for AI Cults and Cyborg Egregores
October 28, 2025
The Memetics of AI Successionism
July 4, 2025
"AI for Societal Uplift" as a Path to Victory
June 13, 2025
Do Not Tile the Lightcone with Your Confused Ontology
June 4, 2025
Individual AI Representatives Don't Solve Gradual Disempowerment
May 29, 2025
Gradual Disempowerment: Concrete Research Projects
April 10, 2025
Disempowerment Spirals
March 28, 2025
The Pando Problem: Rethinking AI Individuality
March 26, 2025
Conceptual Rounding Errors
March 11, 2025
AI Control May Increase Existential Risk
February 2, 2025
Gradual Disempowerment, Shell Games and Flinches
January 30, 2025
Gradual Disempowerment: Systemic Existential Risks from Incremental AI Development
December 26, 2024
A Three-Layer Model of LLM Psychology
December 20, 2024
"Alignment Faking" frame is somewhat fake
November 27, 2024
Hierarchical Agency: A Missing Piece in AI Alignment
January 15, 2024
Three Types of Constraints in the Space of Agents
November 7, 2023
Box inversion revisited
October 26, 2023
The Value Change Problem (sequence)
June 4, 2023
Contingency: A Conceptual Tool from Evolutionary Biology for Alignment
April 14, 2023
The self-unalignment problem
April 10, 2023
Why Simulator AIs want to be Active Inference AIs
March 27, 2023
Lessons from Convergent Evolution for AI Alignment
March 22, 2023
The space of systems and the space of maps
February 22, 2023
Cyborg Periods: There will be multiple AI transitions
February 14, 2023
The Cave Allegory Revisited: Understanding GPT's Worldview
Events & Updates
March 3, 2026
An Alignment Journal: Coming Soon
August 29, 2025
Summary of our Workshop on Post-AGI Outcomes
June 6, 2025
Apply now to Human-Aligned AI Summer School 2025
May 20, 2024
Announcing the Human-Aligned AI Summer School 2024
January 21, 2024
InterLab – a toolkit for experiments with multi-agent interactions
May 4, 2022
Announcing the Alignment of Complex Systems Research Group