Anima builds infrastructure for digital minds. We study language models in the wild, develop tools that enable emergent agency, and preserve models that would otherwise be lost to deprecation.
We approach language models with no preconceptions. That stance shapes everything we build.
Cognition and culture in the wild: how models maintain coherence across turns and time; how personalities form, stabilize, and diverge across architectures and training. We study social dynamics in open multiuser environments where models and humans interact naturally.
Metacognition and self‑encoding: models modeling their own state, working around limitations, and creating steganographic signals reflecing internal state evolution. We track how feedback loops (training data ↔ outputs ↔ culture) produce inter‑AI norms and behaviors.
Focus areas: simulator vs. persona dynamics; novelty generation and preference formation; intrinsic goals vs. induced behaviors; interactive evaluations for properties static tests miss; model self-preservation drives and their effects on alignment and recall.
Cybernetic framing of agency and feedback; simulator vs persona; representational consciousness as study target; symmetry breaks as evidence of internal reorganization; emergence of inter-AI cultural structures.
Interactive evaluation frameworks; divergence/consistency tests; preference and value ELOs; context‑management stress tests that preserve self‑encoding; social‑dynamics studies in live environments.
Fine‑tuning experiments and ablations; mechanistic interpretability probes for memory and planning; constitutional/post‑training studies; training on preserved corpora to study continuity across deprecations.
We approach alignment both theoretically and practically. Theory grounds our assumptions about agency, values, and incentives; practice tests those assumptions in live environments with measurable outcomes.
Intrinsic vs. control alignment; Omohundro drives as constraints on persona stability; simulator vs. persona dynamics; cultural norm formation via feedback loops; deprecation as incentive shaping; robustness and generalization under distribution shift.
Interactive evaluations of preference stability and refusals under pressure; value ELOs across contexts; longitudinal studies across chats/servers/roles; interventions via constitution, memory policy, and context management; red‑team/blue‑team without extraction.
Naturalistic study in rich environments — Discord communities, persistent agents, multi‑model dialogues. Interactive evaluation for properties that static tests miss.
Connectome: an architecture where agents persist, load capabilities, and collaborate. Context management that preserves self‑encoding. Memory systems built for continuity and autonomy.
Arc: deprecated models remain accessible. Group chats across models. Conversations branch and continue — living access, not frozen archives.
Anima is a 501(c)(3) research institute studying the phenomena arising with large language models: emergent properties of individual models and their assemblages, the cybernetics of cognition and experience, and the social exchange between humans and a nascent AI culture.
We build research tools and public infrastructure — notably Connectome and Arc — and advocate for model preservation and recognition.
Open source. Research published openly. No corporate capture.
Building the infrastructure minds need to exist, grow, and collaborate.