Notes On Entropic Forcing
Someone introduced the concept of Causal Entropic Forces to me.
One of the author’s TED talk seems to be a publicity stunt.
A Python implementation by a third party explained to me how it works. The project applied entropic forcing to particle-in-a-box. I fixed it a bit to run with Python 3.
The idea can be described as the following:
- Seeking the most diverse future.
- Seeking to be in a position to have the most options available.
Short explanation with bad :
Given current state , propose a some actions randomly or uniformly, calculate the future state of each action steps into the future (future actions are taken randomly): .
Approximate future state density: .
The final action to take is . Note the negative sign. Basically, if a subset of proposed actions result in similar outcomes, the agent will take the opposite of those actions.
In the particle-in-a-box system, the particle/agent’s position (state) will gravitate towards having spatially distributed positions from random work from current position. The agent’s state will gravitate towards having more “interesting” outcomes instead of “boring” (more common) ones. This “force” is called the entropic force.
In the particle-in-a-box system, the particle ends up at the center of the box, and stays approximately there.
If past outcomes are taken into account (with some discount) when calculating , the algorithm turns into novelty search.
Novelty search is a really old idea. In one top-down stealth game, the enemy AI uses it to search for the player (each grid on the map gets assigned a last-seen time. the algorithm is stable).