mirror of
https://github.com/gryf/coach.git
synced 2025-12-17 19:20:19 +01:00
bug-fix in dueling dqn
This commit is contained in:
@@ -47,7 +47,7 @@ class DuelingQHead(QHead):
|
||||
with tf.variable_scope("action_advantage"):
|
||||
action_advantage = self.dense_layer(512)(input_layer, activation=self.activation_function, name='fc1')
|
||||
action_advantage = self.dense_layer(self.num_actions)(action_advantage, name='fc2')
|
||||
action_advantage = action_advantage - tf.reduce_mean(action_advantage)
|
||||
action_advantage = action_advantage - tf.reduce_mean(action_advantage, axis=1)
|
||||
|
||||
# merge to state-action value function Q
|
||||
self.output = tf.add(state_value, action_advantage, name='output')
|
||||
|
||||
Reference in New Issue
Block a user