Directory Structure ------ . └── Search-based Planning └── Search_2D ├── bfs.py # breadth-first searching ├── dfs.py # depth-first searching ├── dijkstra.py # dijkstra's ├── a_star.py # A* ├── bidirectional_a_star.py # Bidirectional A* ├── ARAstar.py # Anytime Reparing A* ├── IDAstar.py # Iteratively Deepening A* ├── LRTAstar.py # Learning Real-time A* └── RTAAstar.py # Real-time Adaptive A* └── Search_3D ├── Astar3D.py # A*_3D ├── bidirectional_Astar3D.py # Bidirectional A*_3D └── LRT_Astar3D.py # Learning Real-time A*_3D └── gif # Animations └── Sampling-based Planning └── rrt_2D ├── rrt.py # rrt : goal-biased rrt └── rrt_star.py └── rrt_3D ├── rrt3D.py # rrt3D : goal-biased rrt3D └── rrtstar3D.py └── Stochastic Shortest Path ├── value_iteration.py # value iteration ├── policy_iteration.py # policy iteration ├── Q-value_iteration.py # Q-value iteration └── Q-policy_iteration.py # Q-policy iteration └── Model-free Control ├── Sarsa.py # SARSA : on-policy TD control └── Q-learning.py # Q-learning : off-policy TD control ## Animations ### DFS & BFS (Dijkstra) * Blue: starting state * Green: goal state
dfs bfs
### A* and A* Variants
astar biastar
arastar lrtastar
### Value/Policy/Q-value/Q-policy Iteration * Brown: losing states
value iteration value iteration
### SARSA(on-policy) & Q-learning(off-policy) * Brown: losing states
value iteration value iteration
## License MIT License