You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
hi @wanxinjin. I noticed that in the paper the learning rate for PDP, inverse KKT, and neural policy cloning methods in imitation learning was set to $\eta=10^{-4}$. But in scripts like "cartpole_inverseKKT.py", the parameter lr equals 1e-7. Why so?
The text was updated successfully, but these errors were encountered:
thanks. When useing different learning rates, I suppose it would be better to compare the imitation loss of these methods by changing the X label from 'iterations' to 'consumed time'.
hi @wanxinjin. I noticed that in the paper the learning rate for PDP, inverse KKT, and neural policy cloning methods in imitation learning was set to$\eta=10^{-4}$ . But in scripts like "cartpole_inverseKKT.py", the parameter lr equals 1e-7. Why so?
The text was updated successfully, but these errors were encountered: