ACS

Post Archive

Our Academic Publications

See our Research page

Research Notes

Why You Don't Believe in Xhosa Prophecies
Post-AGI Economics As If Nothing Ever Happens
Ontology for AI Cults and Cyborg Egregores
The Memetics of AI Successionism
"AI for Societal Uplift" as a Path to Victory
Do Not Tile the Lightcone with Your Confused Ontology
Individual AI Representatives Don't Solve Gradual Disempowerment
Gradual Disempowerment: Concrete Research Projects
Disempowerment Spirals
The Pando Problem: Rethinking AI Individuality
Conceptual Rounding Errors
AI Control May Increase Existential Risk
Gradual Disempowerment, Shell Games and Flinches
Gradual Disempowerment: Systemic Existential Risks from Incremental AI Development
A Three-Layer Model of LLM Psychology
"Alignment Faking" frame is somewhat fake
Hierarchical Agency: A Missing Piece in AI Alignment
Three Types of Constraints in the Space of Agents
Box inversion revisited
The Value Change Problem (sequence)
Contingency: A Conceptual Tool from Evolutionary Biology for Alignment
The self-unalignment problem
Why Simulator AIs want to be Active Inference AIs
Lessons from Convergent Evolution for AI Alignment
The space of systems and the space of maps
Cyborg Periods: There will be multiple AI transitions
The Cave Allegory Revisited: Understanding GPT's Worldview

Events & Updates

An Alignment Journal: Coming Soon
Summary of our Workshop on Post-AGI Outcomes
Apply now to Human-Aligned AI Summer School 2025
Announcing the Human-Aligned AI Summer School 2024
InterLab – a toolkit for experiments with multi-agent interactions
Announcing the Alignment of Complex Systems Research Group