I write the code following below but I got this error:
TypeError: backward() got an unexpected keyword argument 'retain_variables'
My code is:
def learn(self, batch_state, batch_next_state, batch_reward, batch_action):
outputs = self.model(batch_state).gather(1, batch_action.unsqueeze(1)).squeeze(1)
next_outputs = self.model(batch_next_state).detach().max(1)[0]
target = self.gamma*next_outputs + batch_reward
td_loss = F.smooth_l1_loss(outputs, target)
self.optimizer.zero_grad()
td_loss.backward(retain_variables = True)
self.optimizer.step()
I was having the same problem. This solution worked for me.
td_loss.backward(retain_graph = True)
It worked.
As a_guest mentions in the comments:
It should be retain_graph=True.
If you love us? You can donate to us via Paypal or buy me a coffee so we can maintain and grow! Thank you!
Donate Us With