这是indexloc提供的服务,不要输入任何密码
Skip to content

Support deterministic evaluation for onpolicy algorithms #354

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 7 commits into from
Apr 27, 2021

Conversation

ultmaster
Copy link
Contributor

  • I have marked all applicable categories:
    • exception-raising fix
    • algorithm implementation fix
    • documentation modification
    • new feature
  • If applicable, I have mentioned the relevant/related issue(s) Deterministic sampling for PGPolicy #353

Less important but also useful:

  • I have visited the source website
  • I have searched through the issue tracker for duplicates
  • I have mentioned version numbers, operating system and environment, where applicable:
    import tianshou, torch, numpy, sys
    print(tianshou.__version__, torch.__version__, numpy.__version__, sys.version, sys.platform)

@codecov-commenter
Copy link

codecov-commenter commented Apr 25, 2021

Codecov Report

Merging #354 (2faf4a0) into master (ff4d3cd) will increase coverage by 0.01%.
The diff coverage is 100.00%.

Impacted file tree graph

@@            Coverage Diff             @@
##           master     #354      +/-   ##
==========================================
+ Coverage   94.46%   94.48%   +0.01%     
==========================================
  Files          53       53              
  Lines        3413     3424      +11     
==========================================
+ Hits         3224     3235      +11     
  Misses        189      189              
Flag Coverage Δ
unittests 94.48% <100.00%> (+0.01%) ⬆️

Flags with carried forward coverage won't be shown. Click here to find out more.

Impacted Files Coverage Δ
tianshou/policy/modelfree/a2c.py 96.87% <ø> (ø)
tianshou/policy/modelfree/npg.py 98.85% <ø> (ø)
tianshou/policy/modelfree/ppo.py 90.90% <ø> (ø)
tianshou/policy/modelfree/trpo.py 93.33% <ø> (ø)
tianshou/policy/base.py 79.56% <100.00%> (+0.93%) ⬆️
tianshou/policy/imitation/base.py 100.00% <100.00%> (ø)
tianshou/policy/modelfree/pg.py 96.42% <100.00%> (+0.42%) ⬆️

Continue to review full report at Codecov.

Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update ff4d3cd...2faf4a0. Read the comment docs.

@Trinkle23897 Trinkle23897 requested review from ChenDRAG and danagi April 25, 2021 12:50
@Trinkle23897 Trinkle23897 changed the title Support deterministic evaluation for PG Support deterministic evaluation for onpolicy algorithms Apr 25, 2021
@Trinkle23897 Trinkle23897 linked an issue Apr 26, 2021 that may be closed by this pull request
8 tasks
@Trinkle23897 Trinkle23897 merged commit f4e05d5 into thu-ml:master Apr 27, 2021
BFAnas pushed a commit to BFAnas/tianshou that referenced this pull request May 5, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

Deterministic sampling for PGPolicy
5 participants