Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Batch Normalization combined with your DDPG implementation? #23

Open
ghost opened this issue Mar 7, 2019 · 0 comments
Open

Batch Normalization combined with your DDPG implementation? #23

ghost opened this issue Mar 7, 2019 · 0 comments

Comments

@ghost
Copy link

ghost commented Mar 7, 2019

Hi Morvan,

I am trying to implement your Batch Normalization tutorial on your DDPG algorithme tutorial, but i have a hard time understanding the bits?

one of my problems is:

`        self.a_loss = - tf.reduce_mean(q)  # maximize the q
        self.atrain = tf.train.AdamOptimizer(LR_A).minimize(self.a_loss, var_list=a_params)

        update_ops = tf.get_collection(tf.GraphKeys.UPDATE_OPS)		
        with tf.control_dependencies(update_ops):		
			with tf.control_dependencies(target_update):    # soft replacement happened at here
				self.q_target = self.R + ((GAMMA * (1- self.Done)) * (q_ * (1 - self.Done)))
				self.td_error = tf.losses.mean_squared_error(labels=self.q_target, predictions=q)
				self.ctrain = tf.train.AdamOptimizer(LR_C).minimize(self.td_error, var_list=c_params) `

Since you said you need to have that update_ops i imagned that it should look something like this, but this then won't include the atrain, if not this being incorrect of course?

furthermore if you could give some signs of directions on how to implement it on your ddpg implementation that would be nice,

Jan

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

0 participants