Post archive

Announcing the Human-Aligned AI Summer School 2024
InterLab – a toolkit for experiments with multi-agent interactions
Contingency: A Conceptual Tool from Evolutionary Biology for Alignment

Our posts elsewhere

Three Types of Constraints in the Space of Agents
Box inversion revisited
The Value Change Problem (sequence)
The self-unalignment problem
Why Simulator AIs want to be Active Inference AIs
Lessons from Convergent Evolution for AI Alignment
The space of systems and the space of maps
Cyborg Periods: There will be multiple AI transitions
The Cave Allegory Revisited: Understanding GPT's Worldview
Announcing the Alignment of Complex Systems Research Group