-
Notifications
You must be signed in to change notification settings - Fork 6
Roadmap
sofian edited this page Jul 5, 2013
·
11 revisions
Objective: Working implementation of RL & neural nets within the main architecture
- main abstract components: Agent, Environment, Observation, Action
- reinforcement learning module:
- neural network
- modular policy (e-greedy, softmax)
- modular reward function
- basic documentation
- examples
- README / LICENSE / INSTALL
- Easy switch AVR vs PC
Objective: Stabilization and documentation
- reorganize file structure
- document and comment all classes
- create more concise examples
- clean-up the build process:
- the objects (.o) should compile in sub-directories (build/)
- create a standard installation procedure (configure -> make -> install)
- allow saving/loading of models
- platform independent (look at Design Patterns for ideas)
- think about simulation vs hardware, for example to allow for offline pre-training
- extras:
- libmapper plugin
- arduino plugin
- possibility to stop learning (just act with some policy)
Objective: Integrate Behavior Trees
- Port libbehavior to a "static" version
- Implement a BehaviorTreeAgent
- Create an example and unit test
References:
Objective: Integrate one or many types of swarm algorithms / cellular automata / L-system (*) Not sure yet if we really need this... we'll see...
Objective: Stabilization and documentation
Objective: Augment the RL library
- implement discrete-states (ie. non-neural nets)
- implement continuous actions methods
- advanced methods in RL (to revisit)
- gaussian exploration (variant of softmax)
- ways to solve exploration vs exploitation:
- Directed exploration strategies
- limiting greedy exploration
- E3
- R-Max
- Actor-critic
- hierarchical RL
Objective: Stabilization and documentation
Genetic algorithms? Implementing Genetic Algorithms on Arduino Micro-Controllers
More advanced learning methods
Full-fledge agent-based framework