1
0
mirror of https://github.com/gryf/coach.git synced 2025-12-17 11:10:20 +01:00

RL in Large Discrete Action Spaces - Wolpertinger Agent (#394)

* Currently this is specific to the case of discretizing a continuous action space. Can easily be adapted to other case by feeding the kNN otherwise, and removing the usage of a discretizing output action filter
This commit is contained in:
Gal Leibovich
2019-09-08 12:53:49 +03:00
committed by GitHub
parent fc50398544
commit 138ced23ba
46 changed files with 1193 additions and 51 deletions

View File

@@ -206,6 +206,7 @@
| <a href="#T"><strong>T</strong></a>
| <a href="#U"><strong>U</strong></a>
| <a href="#V"><strong>V</strong></a>
| <a href="#W"><strong>W</strong></a>
</div>
<h2 id="A">A</h2>
@@ -956,6 +957,14 @@
</ul></td>
</tr></table>
<h2 id="W">W</h2>
<table style="width: 100%" class="indextable genindextable"><tr>
<td style="width: 33%; vertical-align: top;"><ul>
<li><a href="components/agents/policy_optimization/wolpertinger.html#rl_coach.agents.wolpertinger_agent.WolpertingerAlgorithmParameters">WolpertingerAlgorithmParameters (class in rl_coach.agents.wolpertinger_agent)</a>
</li>
</ul></td>
</tr></table>
</div>