Hi,
We have been working on many different variants of AOgmaNeo that support Unsupervised Behavioral Learning (UBL).
For those that don't know, UBL is a sort of alternative to classic reinforcement learning (RL). It's a bit different of a paradigm, but instead of optimizing a reward function, it learns the dynamics of the environment and provides a kind if programmable interface to it. The main motivation behind UBL is that we want an agent that is easier to use with real-world robotics which may require a lot of hand-crafting. It also is able to handle instantaneously changing objectives, which regular RL cannot really (even with goal conditioning).
Currently, the best performing UBL branches are able to reproduce some of the results from the original RL version, but not yet all. There is still work to do!
As a result, it will still take a bit before UBL has a chance at making it into the master branch. If you wish to try it anyways, the latest experimental branches are "


