research that inspires us

Our ideas about organic alignment, coherence, and open-ended learning systems and how to go about building them have been built on a foundation of research from many sources. We share a selection of those sources below. We’ve bolded some of the ones that seemed most important to us.

Some Lighter Fare

Most of the research on the page is academic work, of the serious and sober variety. We believe this kind of work is important as it is the foundation of most of our own research. However, we also think that sometimes it’s more important to convey ideas in lighter and more accessible ways.

On the formation of functional wholes

Larger wholes formed of parts have their own independent existence, one that transcends the mechanics of the parts themselves. Think about our own bodies. Even though we have many different kinds of cells which all act independently, there can be no doubt that to the same degree the cells exist as meaningful agents, so do the humans made of those cells. Ant nests are as real as ants. Species are as real as organisms.

On multi-agent cooperation

The foundation of agents coming together to form a greater whole starts with agents being able to cooperate with others who they have not yet cohered with. Before agents come into sync enough that there can be complex webs of interdependence, there needs to be basic interdependence through trade of game theory. This is the seed of care that can develop over time into full alignment.

On consciousness, physics, and information theory

One of the underlying ideas that our research is based on is that you can model agents as having beliefs at many different scales. Cells have beliefs, organisms have beliefs, tribes and nests have beliefs. There are some very important question here: is there a distinction between being able to usefully model an agent as having beliefs, and the agent having them? What is required to have a sentient agent, and how do sentient agents vary? What are the fundamental thermodynamic limits on learning? And what’s the relationship between beliefs, experience, and matter?

On learning in new domains:

Cooperating with other agents and co-aligning with them is a very difficult problem, because it is inherently non-stationary. Because the environment is also made of learning agents, the environment is no longer inductive in any simple way. Actions impact future observations in unpredictable ways dependent on the changing beliefs of the other agents, in a way that does not predictably converge. This makes learning in non-stationary and open-ended environments crucial for alignment.

On expanding the limits of reinforcement learning

Reinforcement learning could also be called inductive learning: we reinforce policies that have worked before in order to do them again in similar situations. In order to succeed in learning in diverse and non-stationary domains, you have to learn how to learn.

On exploration and intrinsic motivation:

Learning how to learn makes exploration problems far more difficult. Instead of learning a strategy for exploration, the agent needs to learn how to explore the space of exploration strategies. This is a far harder requirement.

On Neural Cellular Automata:

NCAs are a very natural multicellular approach to learning that is very directly inspired by biological organisms. Also we think it’s just really cool.