1
0
mirror of https://github.com/gryf/coach.git synced 2025-12-17 19:20:19 +01:00
Files
coach/docs/_sources/components/agents/index.rst.txt
Gal Leibovich 138ced23ba RL in Large Discrete Action Spaces - Wolpertinger Agent (#394)
* Currently this is specific to the case of discretizing a continuous action space. Can easily be adapted to other case by feeding the kNN otherwise, and removing the usage of a discretizing output action filter
2019-09-08 12:53:49 +03:00

49 lines
1.2 KiB
ReStructuredText

Agents
======
Coach supports many state-of-the-art reinforcement learning algorithms, which are separated into three main classes -
value optimization, policy optimization and imitation learning.
A detailed description of those algorithms can be found by navigating to each of the algorithm pages.
.. image:: /_static/img/algorithms.png
:width: 600px
:align: center
.. toctree::
:maxdepth: 1
:caption: Agents
policy_optimization/ac
policy_optimization/acer
imitation/bc
value_optimization/bs_dqn
value_optimization/categorical_dqn
imitation/cil
policy_optimization/cppo
policy_optimization/ddpg
other/dfp
value_optimization/double_dqn
value_optimization/dqn
value_optimization/dueling_dqn
value_optimization/mmc
value_optimization/n_step
value_optimization/naf
value_optimization/nec
value_optimization/pal
policy_optimization/pg
policy_optimization/ppo
value_optimization/rainbow
value_optimization/qr_dqn
policy_optimization/sac
policy_optimization/td3
policy_optimization/wolpertinger
.. autoclass:: rl_coach.base_parameters.AgentParameters
.. autoclass:: rl_coach.agents.agent.Agent
:members:
:inherited-members: