Long-horizon agents — in games, simulators, and eventually robotics — need to maintain and update beliefs about an environment that outlasts any single rollout. We are interested in persistent state layers that survive resets, generalize across instantiations of the same world, and stay legible enough to debug. This is where memory, planning, and identity start to look like the same problem.
- Belief-state memory that updates with new observations
- Multi-scale temporal abstraction: frames, episodes, lifetimes
- Governed state transitions with audit trails
- Integration with reinforcement learning and planning loops
- Targeting embodied agents and open-ended environments