Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Question] Where is CVPO? #311

Open
3 tasks done
JiwenJ opened this issue Apr 4, 2024 · 1 comment
Open
3 tasks done

[Question] Where is CVPO? #311

JiwenJ opened this issue Apr 4, 2024 · 1 comment
Labels
question Further information is requested

Comments

@JiwenJ
Copy link

JiwenJ commented Apr 4, 2024

Required prerequisites

Questions

Hello, I believe there was a CVPO implementation in issue #57. I'm curious as to why it was removed.

@JiwenJ JiwenJ added the question Further information is requested label Apr 4, 2024
@Gaiejj
Copy link
Member

Gaiejj commented Apr 5, 2024

As previously discussed, we found that CVPO needs algorithmic adjustments to the environment's configuration (i.e., removing the randomness of the environment layer). We conducted detailed experiments and found that CVPO's performance on random layers was unsatisfactory.

Relevant experimental evidence has also been disclosed in the SafeRL research. You can refer to the paper published at ICLR 2024: Off-Policy Primal-Dual Safe Reinforcement Learning.

image

Anyway, we are currently trying to upload the requirements for customizing this algorithm's environment further to Safety-Gymnasium modifications to facilitate community use.

If you have any comments or ideas, feel free to discuss them further.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
question Further information is requested
Projects
None yet
Development

No branches or pull requests

2 participants