MP2 - Question 3.1

MP2 - Question 3.1

par Thomas Brunet,
Nombre de réponses : 4
Hello,

For the first agent, the success criteria says "after training has stabilized around an optimal policy (value loss less than 1e-4)".  We were thus wondering if value loss = loss of the critic net, and if so should we really have our critic loss below 1e-4 ? Because it's not the case at all and we can't understand what would be wrong in our implementation. An assistant yesterday told us it was fine but was not so sure.

Thanks.
En réponse à Thomas Brunet

Re: MP2 - Question 3.1

par Skander Moalla,
Hello,

Yes the value loss refers to the critic network's loss and in the scope of question 3.1 (i.e. before you introduce reward stochasticity), this loss should converge when the agent has reached optimal performance.
Around the convergence the loss should be less than 1e-4 as described in the success criteria. (Much less even. In our runs it's 1e-10).
For this, you have to nail the details of the algorithm and in particular bootstrapping.

After question 3.1 this will not be the case anymore as you introduce stochasticity.
En réponse à Skander Moalla

Re: MP2 - Question 3.1

par Thomas Brunet,
Mh ok... The thing is that our loss is stuck around 1, while the rewards during evaluations every 20k steps do increase. 
When we talk about the loss is it well the formula used to update phi in the A2C algorithm ? So basically the advantage squared ?
Thanks.

En réponse à Thomas Brunet

Re: MP2 - Question 3.1

par Skander Moalla,

Yes the loss is the L2 distance between the estimated return and the value function. A common error is not handling autograd correctly when bootstrapping (line 6). Think of which V's should autograd move and which it shouldn't.