1
0
mirror of https://github.com/gryf/coach.git synced 2025-12-18 11:40:18 +01:00
Files
coach/docs_raw/source/components/agents/value_optimization/mmc.rst
Itai Caspi 6d40ad1650 update of api docstrings across coach and tutorials [WIP] (#91)
* updating the documentation website
* adding the built docs
* update of api docstrings across coach and tutorials 0-2
* added some missing api documentation
* New Sphinx based documentation
2018-11-15 15:00:13 +02:00

1.2 KiB
Raw Blame History

Actions space: Discrete

References: Count-Based Exploration with Neural Density Models

Network Structure

/_static/img/design_imgs/dqn.png

Algorithm Description

Training the network

In MMC, targets are calculated as a mixture between Double DQN targets and full Monte Carlo samples (total discounted returns).

The DDQN targets are calculated in the same manner as in the DDQN agent:

yDDQNt=r(st,at)+γQ(st+1,argmaxaQ(st+1,a))

The Monte Carlo targets are calculated by summing up the discounted rewards across the entire episode:

yMCt=Tj=0γjr(st+j,at+j)

A mixing ratio $alpha$ is then used to get the final targets:

yt=(1α)⋅yDDQNt+αyMCt

Finally, the online network is trained using the current states as inputs, and the calculated targets. Once in every few thousand steps, copy the weights from the online network to the target network.