-
Notifications
You must be signed in to change notification settings - Fork 1.2k
Implement CQLPolicy and offline_cql example #506
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
Codecov ReportAll modified and coverable lines are covered by tests ✅
❗ Your organization needs to install the Codecov GitHub app to enable full functionality. Additional details and impacted files@@ Coverage Diff @@
## master #506 +/- ##
==========================================
+ Coverage 94.18% 94.42% +0.24%
==========================================
Files 62 63 +1
Lines 4127 4252 +125
==========================================
+ Hits 3887 4015 +128
+ Misses 240 237 -3
Flags with carried forward coverage won't be shown. Click here to find out more. ☔ View full report in Codecov by Sentry. |
make format
(required)make commit-checks
(required)This PR implements CQLPolicy, which could be used to train an offline agent in the environment of continuous action space. An experimental result 'halfcheetah-medium-v1' is provided, which is a d4rl environment (for Offline Reinforcement Learning).

Example usage is in the examples/offline/offline_bcq.py. Document modification and unit test for CQLPolicy are also completed.