Open-source · MIT · Python · Research project

An experiment in giving
autonomous agents an inner life.

Most autonomous agents express their motivation as a paragraph in a system prompt. This project asks what changes when values and autobiographical episodes are expressed as data instead — an engine that holds values, an episode stream that records what the agent has done, and an alignment composer that checks proposed actions against both. A renderer projects that engine state back into the agent's instruction channel so a language model can see it. Optional layers for beliefs, purpose, desires, and goals are available too, scoped honestly by what early experiments actually showed.

The design is grounded in research from psychology and agent architecture — Schwartz's basic values, AGM belief revision, Self-Determination Theory, Ikigai, Acceptance and Commitment Therapy, the Belief–Desire–Intention model, plus Bem, Erikson, and McAdams for self-concept — but the implementation is uneven by design. Some pieces faithfully encode the cited theory; others are simple heuristics; none of it has been validated in a live agent loop in any strong sense. One headline claim was actively falsified on the initial test seed, then partially neutralised by a follow-up study; that claim now sits at not yet validated, multi-model corroboration pending. This site walks through the research, the architecture, and the experiments themselves — what worked, what did not, and the verbatim agent journals behind the numbers.

Why this matters → Research grounding GitHub

The layers — core · present · research

The layers are grouped by what early experiments showed. Core (values, the self-concept episode stream, and the alignment composer that runs over them — together with the renderer that projects engine state back into the instruction channel) is the part that did measurable work on the initial test seed. Present but unproven (beliefs, purpose) is shipped and faithfully implemented but has not yet been shown to add measurable benefit. Research (desires, goals) is exploratory. The stability gradient runs slowest-at-the-top. Tasks are reasoned about by the framework but executed by the host's runtime.

Slowest-changing layer. Stable principles that constrain everything below. Schwartz's ten universal categories plus three agent-specific extensions, arranged on a circumplex with structural-conflict detection between opposing poles.

ValuesEngine: add, check_action (keyword match), detect_structural_conflicts (Schwartz opposing poles), rank_for_context, export/import.

Schwartz (2012); ACT — Hayes, Strosahl & Wilson (2011).

What this addresses

Agents that persist across sessions tend to drift, contradict themselves, and produce decisions no one can audit. A flat system prompt has no conflict detection, no evolution, no traceability. This project asks whether a structured representation of an agent's values and history makes those properties recoverable — and reports honestly on which parts of that bet are now supported by experiment, which are merely plausible, and which were falsified.

Read more →

Research-grounded, honestly mapped

Each layer cites the theory it draws on (Schwartz, AGM, SDT, Ikigai, ACT, BDI; plus Bem / Erikson / McAdams for self-concept). Some mappings are faithful (the Schwartz circumplex, BDI's three layers). Many are inspirational, not operational. The research page is explicit about which is which.

Research foundations →

Honest about evaluation

Two layers of testing sit alongside each other. Correctness tests verify that the engines behave as specified. Live-agent experiments — long-running studies with a real language model in the loop — probe whether the framework actually changes behaviour. The two answer different questions, and the site keeps them clearly separate.

Tests →

Modular, embeddable, opinionated

Each layer engine works standalone. Storage is a five-method Protocol (in-memory, JSON file, SQLite ship; bring your own). Async-first with a thin sync facade. Embed alongside LangChain, CrewAI, or any agent runtime — it's a library, not a platform.

Get started →

Understand the question

Why a flat system-prompt motivational layer hits friction in long-lived agents.

Motivation →

See the inspirations (and their limits)

Eight theories from psychology and agent architecture, with an honest map of what is and isn't operationalised.

Research →

Read the evidence

What we measure, what we cannot yet measure, and why the distinction matters.

Experiments →

Try it

Install, seed a small hierarchy, and run an alignment check in five minutes. Disagree freely.

Get started →