How to Perform Gradient Descent for DQN Loss Function
    2 views (last 30 days)
  
       Show older comments
    
I'm writing the DQN from scratch, and I'm confused of the procedure of updating the evaluateNet from the gradient descent.
The standard DQN algorithm is to define two networks:  
  . Train
. Train  with minibatch, and update the
 with minibatch, and update the  with gradient descent step on
 with gradient descent step on 
 
  . Train
. Train  with minibatch, and update the
 with minibatch, and update the  with gradient descent step on
 with gradient descent step on 
I define  . When update the
. When update the  , I first make the
, I first make the  , and then only update
, and then only update  , which guarantee the
, which guarantee the  . Then I update the
. Then I update the  . If I choose the feedforward train method as '
. If I choose the feedforward train method as ' ', does [1] update the evalNet correctly via gradient descent?
', does [1] update the evalNet correctly via gradient descent?
 . When update the
. When update the  , I first make the
, I first make the  , and then only update
, and then only update  , which guarantee the
, which guarantee the  . Then I update the
. Then I update the  . If I choose the feedforward train method as '
. If I choose the feedforward train method as ' ', does [1] update the evalNet correctly via gradient descent?
', does [1] update the evalNet correctly via gradient descent?0 Comments
Answers (0)
See Also
Categories
				Find more on Deep Learning Toolbox in Help Center and File Exchange
			
	Community Treasure Hunt
Find the treasures in MATLAB Central and discover how the community can help you!
Start Hunting!