[Question] Policy collapse #3267
Unanswered
AntonioClaudiossf
asked this question in
Q&A
Replies: 1 comment 1 reply
-
Thank you for posting this. It is a great question for our Discussions section, I'll move the post there for follow up by the team and others. In the meantime, you may want to consider the following:
Footnotes |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Question
Hi everyone,
I am currently working on a peg-insertion task using PPO in Isaac Lab. The training starts well, the agent improves and the reward increases, but after a certain point the performance suddenly collapses and becomes unstable.
I will attach the TensorBoard plots of the loss functions and the reward so you can better see what is happening. I have already tried tuning several PPO parameters, such as lowering the
kl_threshold
, changing thevalue_loss_scale
, adjusting theentropy_loss_scale
, modifying thelearning_rate
, and even testing different model size. However, the problem still occurs.The Loss during the train:
The rewards max, mean and min:
PPO parameters:
Model :
Has anyone faced similar issues with peg-insertion or other high-precision tasks? Do you have any ideas on why the training might be collapsing, or suggestions for further adjustments/debugging steps?
Any insights would be greatly appreciated.
Thanks in advance!
Beta Was this translation helpful? Give feedback.
All reactions