How to Perform Gradient Descent for DQN Loss Function

2 次查看(过去 30 天)
I'm writing the DQN from scratch, and I'm confused of the procedure of updating the evaluateNet from the gradient descent.
The standard DQN algorithm is to define two networks: . Train with minibatch, and update the with gradient descent step on
I define . When update the , I first make the , and then only update , which guarantee the . Then I update the . If I choose the feedforward train method as '', does [1] update the evalNet correctly via gradient descent?

回答(0 个)

Community Treasure Hunt

Find the treasures in MATLAB Central and discover how the community can help you!

Start Hunting!

Translated by