Releases: JuliaReinforcementLearning/ReinforcementLearning.jl
Releases · JuliaReinforcementLearning/ReinforcementLearning.jl
ReinforcementLearningCore-v0.15.0
ReinforcementLearningCore ReinforcementLearningCore-v0.15.0
Diff since ReinforcementLearningCore-v0.14.0
Merged pull requests:
- Bump RLZoo to v0.8 (#1031) (@jeremiahpslewis)
- Fix RLZoo version (#1032) (@jeremiahpslewis)
- Drop devmode, prepare RL.jl v0.11 for release (#1035) (@jeremiahpslewis)
- Update docs script for new 'limited' RL.jl release (#1038) (@jeremiahpslewis)
- Tabular Approximator fixes (pre v0.11 changes) (#1040) (@jeremiahpslewis)
- Swap RLZoo for RLFarm in CI, drop RLExperiments (#1041) (@jeremiahpslewis)
- Buildkite tweaks for monorepo (#1042) (@jeremiahpslewis)
- Drop archived projects (#1043) (@jeremiahpslewis)
- Simplify Experiment code after dropping RLExperiment (#1044) (@jeremiahpslewis)
- Fix code coverage scope so it ignores test dir (#1045) (@jeremiahpslewis)
- Fix reset and stop conditions (#1046) (@jeremiahpslewis)
- Drop Functors and use Flux.@layer (#1048) (@jeremiahpslewis)
- Fix naming consistency and add missing hook tests (#1049) (@jeremiahpslewis)
- Add SARS tdlearning back to lib (#1050) (@jeremiahpslewis)
- Update FluxModelApproximator references to FluxApproximator (#1051) (@jeremiahpslewis)
- Epsilon Speedy Explorer (#1052) (@jeremiahpslewis)
- Add TotalRewardPerEpisodeLastN hook (#1053) (@jeremiahpslewis)
- Fix abstract_learner for multiplayer games (#1054) (@jeremiahpslewis)
- Update versions (#1055) (@jeremiahpslewis)
- Update Docs for v0.11 release (#1056) (@jeremiahpslewis)
- Update Katex version, fix vulnerability (#1058) (@jeremiahpslewis)
Closed issues:
ReinforcementLearningBase-v0.13.0
ReinforcementLearningBase ReinforcementLearningBase-v0.13.0
Diff since ReinforcementLearningBase-v0.12.2
Merged pull requests:
- Fix offline agent test (#1025) (@joelreymont)
- Fix spell check CI errors (#1027) (@joelreymont)
- GPU Code Migration Part 2.1 (#1029) (@jeremiahpslewis)
- Bump RLZoo to v0.8 (#1031) (@jeremiahpslewis)
- Fix RLZoo version (#1032) (@jeremiahpslewis)
- Drop devmode, prepare RL.jl v0.11 for release (#1035) (@jeremiahpslewis)
- Update docs script for new 'limited' RL.jl release (#1038) (@jeremiahpslewis)
- Tabular Approximator fixes (pre v0.11 changes) (#1040) (@jeremiahpslewis)
- Swap RLZoo for RLFarm in CI, drop RLExperiments (#1041) (@jeremiahpslewis)
- Buildkite tweaks for monorepo (#1042) (@jeremiahpslewis)
- Drop archived projects (#1043) (@jeremiahpslewis)
- Simplify Experiment code after dropping RLExperiment (#1044) (@jeremiahpslewis)
- Fix code coverage scope so it ignores test dir (#1045) (@jeremiahpslewis)
- Fix reset and stop conditions (#1046) (@jeremiahpslewis)
- Drop Functors and use Flux.@layer (#1048) (@jeremiahpslewis)
- Fix naming consistency and add missing hook tests (#1049) (@jeremiahpslewis)
- Add SARS tdlearning back to lib (#1050) (@jeremiahpslewis)
- Update FluxModelApproximator references to FluxApproximator (#1051) (@jeremiahpslewis)
- Epsilon Speedy Explorer (#1052) (@jeremiahpslewis)
- Add TotalRewardPerEpisodeLastN hook (#1053) (@jeremiahpslewis)
- Fix abstract_learner for multiplayer games (#1054) (@jeremiahpslewis)
- Update versions (#1055) (@jeremiahpslewis)
- Update Docs for v0.11 release (#1056) (@jeremiahpslewis)
- Update Katex version, fix vulnerability (#1058) (@jeremiahpslewis)
Closed issues:
ReinforcementLearningZoo-v0.9.0
ReinforcementLearningZoo ReinforcementLearningZoo-v0.9.0
Merged pull requests:
- Fix deprecations (#10) (@femtocleaner[bot])
- implement epsilon-greedy policy with parametric type (#12) (@jbrea)
- improve docs (#13) (@jbrea)
- refactor policies (#15) (@jbrea)
- Add ReinforcementLearningBase as dependent (#16) (@jbrea)
- fix examples (#18) (@jbrea)
- refactor existing components (#26) (@findmyway)
- Prioritized dqn (#29) (@findmyway)
- add double dqn (#30) (@findmyway)
- add rainbow (#31) (@findmyway)
- use new api in ReinforcementLearningEnvironments.jl (#33) (@findmyway)
- bugfix and api simplification (#34) (@findmyway)
- Switch Tracker.jl to Zygote.jl (#37) (@findmyway)
- Support both Knet and Flux(with Zygote) (#38) (@findmyway)
- add docs (#39) (@findmyway)
- export AbstractActionSelector and add more comments (#42) (@findmyway)
- Refactor buffer (#45) (@findmyway)
- fix example in doc && update examples (#46) (@findmyway)
- fix a performance bug in rainbow (#47) (@findmyway)
- update dependencies (#48) (@findmyway)
- update dependencies and docs (#49) (@findmyway)
- update benchmark for circular_array_buffer (#50) (@findmyway)
- Install TagBot as a GitHub Action (#53) (@JuliaTagBot)
- clean up code (#54) (@findmyway)
- add compat (#55) (@findmyway)
- CompatHelper: add new compat entry for "Reexport" at version "0.2" (#56) (@github-actions[bot])
- add documentation stage in travis (#57) (@findmyway)
- Add doc in travis (#58) (@findmyway)
- Fix link in docs/src/index.md (#60) (@amanbh)
- Update doc (#61) (@findmyway)
- Update README.md & website link (#70) (@findmyway)
- Update dependency (#78) (@findmyway)
- MassInstallAction: Install the CompatHelper workflow on this repository (#99) (@findmyway)
- CompatHelper: bump compat for "Reexport" to "1.0" (#172) (@github-actions[bot])
- update dependency (#177) (@findmyway)
- Add Dockerfile (#187) (@findmyway)
- Update readme (#188) (@findmyway)
- docs: add findmyway as a contributor (#189) (@allcontributors[bot])
- docs: add drozzy as a contributor (#195) (@allcontributors[bot])
- docs: add rcnlee as a contributor (#199) (@allcontributors[bot])
- docs: add norci as a contributor (#200) (@allcontributors[bot])
- docs: add xiruizhao as a contributor (#203) (@allcontributors[bot])
- docs: add metab0t as a contributor (#204) (@allcontributors[bot])
- docs: add albheim as a contributor (#207) (@allcontributors[bot])
- docs: add michelangelo21 as a contributor (#214) (@allcontributors[bot])
- docs: add pilgrimygy as a contributor (#216) (@allcontributors[bot])
- docs: add Mobius1D as a contributor (#218) (@allcontributors[bot])
- docs: add ilancoulon as a contributor (#224) (@allcontributors[bot])
- docs: add pilgrimygy as a contributor (#230) (@allcontributors[bot])
- docs: add JinraeKim as a contributor (#243) (@allcontributors[bot])
- Prepare v0.9 (#252) (@findmyway)
- docs: add luigiannelli as a contributor (#254) (@allcontributors[bot])
- docs: add JBoerma as a contributor (#255) (@allcontributors[bot])
- CompatHelper: bump compat for "ReinforcementLearningEnvironments" to "0.5" (#260) (@github-actions[bot])
- Fix inconsitencies in wrappers (#263) (@albheim)
- setup CI for each subpackages (#264) (@findmyway)
- Fix atari experiments (#265) (@Mobius1D)
- Add timeperstep hook to qrdqn to fix test error (#266) (@albheim)
- Update Flux version (#267) (@findmyway)
- Setup docs generation pipeline (#269) (@findmyway)
- Misc doc related fixes (#270) (@findmyway)
- Update README.md (#271) (@findmyway)
- docs: add JinraeKim as a contributor (#272) (@allcontributors[bot])
- Improve docs GitHub action (#273) (@findmyway)
- Fix docs pipeline (#275) (@findmyway)
- update readme (#276) (@findmyway)
- CompatHelper: add new compat entry for "UnicodePlots" at version "1.3" for package ReinforcementLearningCore (#277) (@github-actions[bot])
- CompatHelper: bump compat for "Distributions" to "0.25" for package ReinforcementLearningCore (#278) (@github-actions[bot])
- CompatHelper: bump compat for "Distributions" to "0.25" for package ReinforcementLearningZoo (#279) (@github-actions[bot])
- docs: add plu70n as a contributor (#282) (@allcontributors[bot])
- Fix bug in CI (#283) (@findmyway)
- Use Weave.jl to generate RLExperiments (#284) (@findmyway)
- QRDQN experiment reproducibility fix (#294) (@ashwani-rathee)
- Add Manifest.toml (#295) (@findmyway)
- docs: add ashwani-rathee as a contributor (#296) (@allcontributors[bot])
- Add basic doc structure (#300) (@findmyway)
- Update guide (#302) (@findmyway)
- Update experiments (#303) (@findmyway)
- fix figs (#304) (@findmyway)
- Fix some simple experiments (#308) (@findmyway)
- add plotting for cartpole and mountaincar with Plots.jl (#309) (@jamblejoe)
- Remove GR in RLEnvs (#310) (@findmyway)
- docs: add jamblejoe as a contributor (#311) (@allcontributors[bot])
- Add compat of [email protected] in ReinforcementLearningExperiments (#312) (@findmyway)
- Add example of SimplexSpace (#313) (@findmyway)
- Improve tutorial (#314) (@findmyway)
- Fix Atari related experiments (#315) (@findmyway)
- CompatHelper: add new compat entry for "ImageTransformations" at version "0.8" for package ReinforcementLearningExperiments (#317) (@github-actions[bot])
- CompatHelper: add new compat entry for "ArcadeLearningEnvironment" at version "0.2" for package ReinforcementLearningExperiments (#318) (@github-actions[bot])
- CompatHelper: add new compat entry for "CUDA" at version "3" for package ReinforcementLearningExperiments (#319) (@github-actions[bot])
- update tips (#321) (@findmyway)
- CompatHelper: bump compat for "GPUArrays" to "7" for package ReinforcementLearningCore (#322) (@github-actions[bot])
- docs: add albheim as a contributor for doc (#323) (@allcontributors[bot])
- Fix broken test (#325) (@Mobius1D)
- Add a warning in docstring of state (#327) (@findmyway)
- Update doc string of PrioritizedDQNLearner (#329) (@findmyway)
- Expand DDPG to multi action spaces (#330) (@Mobius1D)
- CompatHelper: bump compat for "StructArrays" to "0.6" for package ReinforcementLearningZoo (#331) (@github-actions[bot])
- fix 332 (#333) (@findmyway)
- correct spelling in FAQ (#334) (@ultradian)
- docs: add ultradian as a contributor for doc (#335) (@allcontributors[bot])
- fix typo (#338) (@findmyway)
- docs: add eltociear as a contributor for doc (#339) (@allcontributors[bot])
- CompatHelper: bump compat for "FillArrays" to "0.12" for package ReinforcementLearningCore (#340) (@github-actions[bot])
- Add copyto function (#345) (@pilgrimygy)
- add Base.:(==) and Base.hash for AbstractEnv and test nash_conv on KuhnPokerEnv (#348) (@peterchen96)
- Fix legal_actions_mask indexing error in CircularSLART (#350) (@findmyway)
- bump version of RLCore (#351) (@findmyway)
- bump version of RLBae (#352) (@findmyway)
- add LICENSE in RLBase (#353) (@findmyway)
- bump version of RLZoo (#355) (@findmyway)
- docs: add 00krishna as a contributor for bug (#356) (@allcontributors[bot])
- Add the tuning entropy component (#365) (@pilgrimygy)
- Make general components (#370) (@pilgrimygy)
- add weighted_softmax_explorer in the
explorers.jl
(#382) (@peterchen96) - Supplement functions in ReservoirTrajectory and BehaviorCloningPolicy (#390) (@peterchen96)
- Update Flux version (#391) (@findmyway)
- AddSequentialEnv (#394) (@findmyway)
- Throw error in MultiAgentManager if it is applied to a SIMULTANEOUS env (#395) (@findmyway)
- docs: add pkienscherf as a contributor for bug (#396) (@allcontributors[bot])
- Implementation of NFSP and NFSP_KuhnPoker experiment (#402) (@peterchen96)
- Updated RLDatasets.jl (#403) (@Mobius1D)
- Gym d4rl extension (#405) (@Mobius1D)
- updates as per GridWorlds v0.5.0 (#406) (@Sid-Bhatia-0)
- Reduce allocations and update docstring for GaussianNetwork (#414) (@albheim)
- Fix a bug (#415) (@pilgrimygy)
- Expand to d4rl-pybullet (#416) (@Mobius1D)
- docs: add pilgrimygy as a contributor for bug (#417) (@allcontributors[bot])
- Fix 418 (#420) (@findmyway)
- docs: add Krastanov as a contributor for doc (#422) (@allcontributors[bot])
- Make SAC policy use internal rng (#423) (@albheim)
- Add wrapped_env[!] to access env inside wrappers (#426) (@albheim)
- add stock trading env (#428) (@findmyway)
- Add Atari datasets released by Google Research (#429) (@Mobius1D)
- add kwargs to plot(env::) (#430) (@jamblejoe)
- Unify parameter names (#437) (@pilgrimygy)
- docs: add albheim as a contributor for maintenance (#438) (@allcontributors[bot])
- correct
nfsp
implementation (#439) (@peterchen96) - update the nfsp experiment's parameters (#440) (@peterchen96)
- Tiny text typo (#441) (@Nthman)
- docs: add LaarsOman as a contributor for doc (#442) (@allcontributors[bot])
- Add pre-train step; VAE component; CRR and PLAS algorithms (#443) (@pilgrimygy)
- add MADDPG algorithm (#444) (@peterchen96)
- add_Graph_Shortest_Path (#445) (@findmyway)
- try to fix bugs of ActionTransformedEnv (#447) (@peterchen96)
- Update report (#448) (@pilgrimygy)
- Summer ospp project 210370190 mid-term report (#449) (@peterchen96)
- add maxdepth kwarg to remove print_tree() deprecation warning (#450) (@burmecia)
- docs: add burmecia as a contributor for code (#451) (@allcontributors[bot])
- RL unplugged implementation with tests (#452) (@Mobius1D)
- Update report (#453) (@Mobius1D)
- disable notebook generation (#454) (@johnnychen94)
- Revert "Update report" (#456) (@findmyway)
- Update report (#457) (@pilgrimygy)
- fix installation docs (#458) (@Mobius1D)
- docs: add peterchen96 as a contributor for code, doc (#459) (@allcontributors[bot])
- Create LICENSE (#461) (@Mobius1D)
- Add docs (#462) (@Mobius1D)
- Fix make.jl (#463) (@Mobius1D)
- Delete LICENSE (#465) (@Mobius1D)
- fix CI (#466) (@findmyway)
- Fix RLDatasets.jl documentation (#467) (@Mobius1D)
- update report (#468) (@peterchen96)
- Fix ci (#469) (@Mobius1D)
- Update maddpg and the report (#470) (@peterchen96)
- Report (#474) (@pilgrimygy)
- Control whether
run
displays description of experiment (#477) (@Shuhua...
ReinforcementLearningExperiments-v0.4.0
ReinforcementLearningExperiments ReinforcementLearningExperiments-v0.4.0
Merged pull requests:
- Fix deprecations (#10) (@femtocleaner[bot])
- implement epsilon-greedy policy with parametric type (#12) (@jbrea)
- improve docs (#13) (@jbrea)
- refactor policies (#15) (@jbrea)
- Add ReinforcementLearningBase as dependent (#16) (@jbrea)
- fix examples (#18) (@jbrea)
- refactor existing components (#26) (@findmyway)
- Prioritized dqn (#29) (@findmyway)
- add double dqn (#30) (@findmyway)
- add rainbow (#31) (@findmyway)
- use new api in ReinforcementLearningEnvironments.jl (#33) (@findmyway)
- bugfix and api simplification (#34) (@findmyway)
- Switch Tracker.jl to Zygote.jl (#37) (@findmyway)
- Support both Knet and Flux(with Zygote) (#38) (@findmyway)
- add docs (#39) (@findmyway)
- export AbstractActionSelector and add more comments (#42) (@findmyway)
- Refactor buffer (#45) (@findmyway)
- fix example in doc && update examples (#46) (@findmyway)
- fix a performance bug in rainbow (#47) (@findmyway)
- update dependencies (#48) (@findmyway)
- update dependencies and docs (#49) (@findmyway)
- update benchmark for circular_array_buffer (#50) (@findmyway)
- Install TagBot as a GitHub Action (#53) (@JuliaTagBot)
- clean up code (#54) (@findmyway)
- add compat (#55) (@findmyway)
- CompatHelper: add new compat entry for "Reexport" at version "0.2" (#56) (@github-actions[bot])
- add documentation stage in travis (#57) (@findmyway)
- Add doc in travis (#58) (@findmyway)
- Fix link in docs/src/index.md (#60) (@amanbh)
- Update doc (#61) (@findmyway)
- Update README.md & website link (#70) (@findmyway)
- Update dependency (#78) (@findmyway)
- MassInstallAction: Install the CompatHelper workflow on this repository (#99) (@findmyway)
- CompatHelper: bump compat for "Reexport" to "1.0" (#172) (@github-actions[bot])
- update dependency (#177) (@findmyway)
- Add Dockerfile (#187) (@findmyway)
- Update readme (#188) (@findmyway)
- docs: add findmyway as a contributor (#189) (@allcontributors[bot])
- docs: add drozzy as a contributor (#195) (@allcontributors[bot])
- docs: add rcnlee as a contributor (#199) (@allcontributors[bot])
- docs: add norci as a contributor (#200) (@allcontributors[bot])
- docs: add xiruizhao as a contributor (#203) (@allcontributors[bot])
- docs: add metab0t as a contributor (#204) (@allcontributors[bot])
- docs: add albheim as a contributor (#207) (@allcontributors[bot])
- docs: add michelangelo21 as a contributor (#214) (@allcontributors[bot])
- docs: add pilgrimygy as a contributor (#216) (@allcontributors[bot])
- docs: add Mobius1D as a contributor (#218) (@allcontributors[bot])
- docs: add ilancoulon as a contributor (#224) (@allcontributors[bot])
- docs: add pilgrimygy as a contributor (#230) (@allcontributors[bot])
- docs: add JinraeKim as a contributor (#243) (@allcontributors[bot])
- Prepare v0.9 (#252) (@findmyway)
- docs: add luigiannelli as a contributor (#254) (@allcontributors[bot])
- docs: add JBoerma as a contributor (#255) (@allcontributors[bot])
- CompatHelper: bump compat for "ReinforcementLearningEnvironments" to "0.5" (#260) (@github-actions[bot])
- Fix inconsitencies in wrappers (#263) (@albheim)
- setup CI for each subpackages (#264) (@findmyway)
- Fix atari experiments (#265) (@Mobius1D)
- Add timeperstep hook to qrdqn to fix test error (#266) (@albheim)
- Update Flux version (#267) (@findmyway)
- Setup docs generation pipeline (#269) (@findmyway)
- Misc doc related fixes (#270) (@findmyway)
- Update README.md (#271) (@findmyway)
- docs: add JinraeKim as a contributor (#272) (@allcontributors[bot])
- Improve docs GitHub action (#273) (@findmyway)
- Fix docs pipeline (#275) (@findmyway)
- update readme (#276) (@findmyway)
- CompatHelper: add new compat entry for "UnicodePlots" at version "1.3" for package ReinforcementLearningCore (#277) (@github-actions[bot])
- CompatHelper: bump compat for "Distributions" to "0.25" for package ReinforcementLearningCore (#278) (@github-actions[bot])
- CompatHelper: bump compat for "Distributions" to "0.25" for package ReinforcementLearningZoo (#279) (@github-actions[bot])
- docs: add plu70n as a contributor (#282) (@allcontributors[bot])
- Fix bug in CI (#283) (@findmyway)
- Use Weave.jl to generate RLExperiments (#284) (@findmyway)
- QRDQN experiment reproducibility fix (#294) (@ashwani-rathee)
- Add Manifest.toml (#295) (@findmyway)
- docs: add ashwani-rathee as a contributor (#296) (@allcontributors[bot])
- Add basic doc structure (#300) (@findmyway)
- Update guide (#302) (@findmyway)
- Update experiments (#303) (@findmyway)
- fix figs (#304) (@findmyway)
- Fix some simple experiments (#308) (@findmyway)
- add plotting for cartpole and mountaincar with Plots.jl (#309) (@jamblejoe)
- Remove GR in RLEnvs (#310) (@findmyway)
- docs: add jamblejoe as a contributor (#311) (@allcontributors[bot])
- Add compat of [email protected] in ReinforcementLearningExperiments (#312) (@findmyway)
- Add example of SimplexSpace (#313) (@findmyway)
- Improve tutorial (#314) (@findmyway)
- Fix Atari related experiments (#315) (@findmyway)
- CompatHelper: add new compat entry for "ImageTransformations" at version "0.8" for package ReinforcementLearningExperiments (#317) (@github-actions[bot])
- CompatHelper: add new compat entry for "ArcadeLearningEnvironment" at version "0.2" for package ReinforcementLearningExperiments (#318) (@github-actions[bot])
- CompatHelper: add new compat entry for "CUDA" at version "3" for package ReinforcementLearningExperiments (#319) (@github-actions[bot])
- update tips (#321) (@findmyway)
- CompatHelper: bump compat for "GPUArrays" to "7" for package ReinforcementLearningCore (#322) (@github-actions[bot])
- docs: add albheim as a contributor for doc (#323) (@allcontributors[bot])
- Fix broken test (#325) (@Mobius1D)
- Add a warning in docstring of state (#327) (@findmyway)
- Update doc string of PrioritizedDQNLearner (#329) (@findmyway)
- Expand DDPG to multi action spaces (#330) (@Mobius1D)
- CompatHelper: bump compat for "StructArrays" to "0.6" for package ReinforcementLearningZoo (#331) (@github-actions[bot])
- fix 332 (#333) (@findmyway)
- correct spelling in FAQ (#334) (@ultradian)
- docs: add ultradian as a contributor for doc (#335) (@allcontributors[bot])
- fix typo (#338) (@findmyway)
- docs: add eltociear as a contributor for doc (#339) (@allcontributors[bot])
- CompatHelper: bump compat for "FillArrays" to "0.12" for package ReinforcementLearningCore (#340) (@github-actions[bot])
- Add copyto function (#345) (@pilgrimygy)
- add Base.:(==) and Base.hash for AbstractEnv and test nash_conv on KuhnPokerEnv (#348) (@peterchen96)
- Fix legal_actions_mask indexing error in CircularSLART (#350) (@findmyway)
- bump version of RLCore (#351) (@findmyway)
- bump version of RLBae (#352) (@findmyway)
- add LICENSE in RLBase (#353) (@findmyway)
- bump version of RLZoo (#355) (@findmyway)
- docs: add 00krishna as a contributor for bug (#356) (@allcontributors[bot])
- Add the tuning entropy component (#365) (@pilgrimygy)
- Make general components (#370) (@pilgrimygy)
- add weighted_softmax_explorer in the
explorers.jl
(#382) (@peterchen96) - Supplement functions in ReservoirTrajectory and BehaviorCloningPolicy (#390) (@peterchen96)
- Update Flux version (#391) (@findmyway)
- AddSequentialEnv (#394) (@findmyway)
- Throw error in MultiAgentManager if it is applied to a SIMULTANEOUS env (#395) (@findmyway)
- docs: add pkienscherf as a contributor for bug (#396) (@allcontributors[bot])
- Implementation of NFSP and NFSP_KuhnPoker experiment (#402) (@peterchen96)
- Updated RLDatasets.jl (#403) (@Mobius1D)
- Gym d4rl extension (#405) (@Mobius1D)
- updates as per GridWorlds v0.5.0 (#406) (@Sid-Bhatia-0)
- Reduce allocations and update docstring for GaussianNetwork (#414) (@albheim)
- Fix a bug (#415) (@pilgrimygy)
- Expand to d4rl-pybullet (#416) (@Mobius1D)
- docs: add pilgrimygy as a contributor for bug (#417) (@allcontributors[bot])
- Fix 418 (#420) (@findmyway)
- docs: add Krastanov as a contributor for doc (#422) (@allcontributors[bot])
- Make SAC policy use internal rng (#423) (@albheim)
- Add wrapped_env[!] to access env inside wrappers (#426) (@albheim)
- add stock trading env (#428) (@findmyway)
- Add Atari datasets released by Google Research (#429) (@Mobius1D)
- add kwargs to plot(env::) (#430) (@jamblejoe)
- Unify parameter names (#437) (@pilgrimygy)
- docs: add albheim as a contributor for maintenance (#438) (@allcontributors[bot])
- correct
nfsp
implementation (#439) (@peterchen96) - update the nfsp experiment's parameters (#440) (@peterchen96)
- Tiny text typo (#441) (@Nthman)
- docs: add LaarsOman as a contributor for doc (#442) (@allcontributors[bot])
- Add pre-train step; VAE component; CRR and PLAS algorithms (#443) (@pilgrimygy)
- add MADDPG algorithm (#444) (@peterchen96)
- add_Graph_Shortest_Path (#445) (@findmyway)
- try to fix bugs of ActionTransformedEnv (#447) (@peterchen96)
- Update report (#448) (@pilgrimygy)
- Summer ospp project 210370190 mid-term report (#449) (@peterchen96)
- add maxdepth kwarg to remove print_tree() deprecation warning (#450) (@burmecia)
- docs: add burmecia as a contributor for code (#451) (@allcontributors[bot])
- RL unplugged implementation with tests (#452) (@Mobius1D)
- Update report (#453) (@Mobius1D)
- disable notebook generation (#454) (@johnnychen94)
- Revert "Update report" (#456) (@findmyway)
- Update report (#457) (@pilgrimygy)
- fix installation docs (#458) (@Mobius1D)
- docs: add peterchen96 as a contributor for code, doc (#459) (@allcontributors[bot])
- Create LICENSE (#461) (@Mobius1D)
- Add docs (#462) (@Mobius1D)
- Fix make.jl (#463) (@Mobius1D)
- Delete LICENSE (#465) (@Mobius1D)
- fix CI (#466) (@findmyway)
- Fix RLDatasets.jl documentation (#467) (@Mobius1D)
- update report (#468) (@peterchen96)
- Fix ci (#469) (@Mobius1D)
- Update maddpg and the report (#470) (@peterchen96)
- Report (#474) (@pilgrimygy)
- Control whether
run
displays description of experiment...
ReinforcementLearningEnvironments-v0.8.8
ReinforcementLearningEnvironments ReinforcementLearningEnvironments-v0.8.8
Diff since ReinforcementLearningEnvironments-v0.8.7
Merged pull requests:
- Fix offline agent test (#1025) (@joelreymont)
- Fix spell check CI errors (#1027) (@joelreymont)
- GPU Code Migration Part 2.1 (#1029) (@jeremiahpslewis)
Closed issues:
ReinforcementLearningCore-v0.14.0
ReinforcementLearningCore ReinforcementLearningCore-v0.14.0
Diff since ReinforcementLearningCore-v0.13.1
Merged pull requests:
- Fix offline agent test (#1025) (@joelreymont)
- Fix spell check CI errors (#1027) (@joelreymont)
- GPU Code Migration Part 2.1 (#1029) (@jeremiahpslewis)
Closed issues:
ReinforcementLearningEnvironments-v0.8.7
ReinforcementLearningEnvironments ReinforcementLearningEnvironments-v0.8.7
ReinforcementLearningEnvironments-v0.8.6
ReinforcementLearningEnvironments ReinforcementLearningEnvironments-v0.8.6
Merged pull requests:
- Fix deprecations (#10) (@femtocleaner[bot])
- implement epsilon-greedy policy with parametric type (#12) (@jbrea)
- improve docs (#13) (@jbrea)
- refactor policies (#15) (@jbrea)
- Add ReinforcementLearningBase as dependent (#16) (@jbrea)
- fix examples (#18) (@jbrea)
- refactor existing components (#26) (@findmyway)
- Prioritized dqn (#29) (@findmyway)
- add double dqn (#30) (@findmyway)
- add rainbow (#31) (@findmyway)
- use new api in ReinforcementLearningEnvironments.jl (#33) (@findmyway)
- bugfix and api simplification (#34) (@findmyway)
- Switch Tracker.jl to Zygote.jl (#37) (@findmyway)
- Support both Knet and Flux(with Zygote) (#38) (@findmyway)
- add docs (#39) (@findmyway)
- export AbstractActionSelector and add more comments (#42) (@findmyway)
- Refactor buffer (#45) (@findmyway)
- fix example in doc && update examples (#46) (@findmyway)
- fix a performance bug in rainbow (#47) (@findmyway)
- update dependencies (#48) (@findmyway)
- update dependencies and docs (#49) (@findmyway)
- update benchmark for circular_array_buffer (#50) (@findmyway)
- Install TagBot as a GitHub Action (#53) (@JuliaTagBot)
- clean up code (#54) (@findmyway)
- add compat (#55) (@findmyway)
- CompatHelper: add new compat entry for "Reexport" at version "0.2" (#56) (@github-actions[bot])
- add documentation stage in travis (#57) (@findmyway)
- Add doc in travis (#58) (@findmyway)
- Fix link in docs/src/index.md (#60) (@amanbh)
- Update doc (#61) (@findmyway)
- Update README.md & website link (#70) (@findmyway)
- Update dependency (#78) (@findmyway)
- MassInstallAction: Install the CompatHelper workflow on this repository (#99) (@findmyway)
- CompatHelper: bump compat for "Reexport" to "1.0" (#172) (@github-actions[bot])
- update dependency (#177) (@findmyway)
- Add Dockerfile (#187) (@findmyway)
- Update readme (#188) (@findmyway)
- docs: add findmyway as a contributor (#189) (@allcontributors[bot])
- docs: add drozzy as a contributor (#195) (@allcontributors[bot])
- docs: add rcnlee as a contributor (#199) (@allcontributors[bot])
- docs: add norci as a contributor (#200) (@allcontributors[bot])
- docs: add xiruizhao as a contributor (#203) (@allcontributors[bot])
- docs: add metab0t as a contributor (#204) (@allcontributors[bot])
- docs: add albheim as a contributor (#207) (@allcontributors[bot])
- docs: add michelangelo21 as a contributor (#214) (@allcontributors[bot])
- docs: add pilgrimygy as a contributor (#216) (@allcontributors[bot])
- docs: add Mobius1D as a contributor (#218) (@allcontributors[bot])
- docs: add ilancoulon as a contributor (#224) (@allcontributors[bot])
- docs: add pilgrimygy as a contributor (#230) (@allcontributors[bot])
- docs: add JinraeKim as a contributor (#243) (@allcontributors[bot])
- Prepare v0.9 (#252) (@findmyway)
- docs: add luigiannelli as a contributor (#254) (@allcontributors[bot])
- docs: add JBoerma as a contributor (#255) (@allcontributors[bot])
- CompatHelper: bump compat for "ReinforcementLearningEnvironments" to "0.5" (#260) (@github-actions[bot])
- Fix inconsitencies in wrappers (#263) (@albheim)
- setup CI for each subpackages (#264) (@findmyway)
- Fix atari experiments (#265) (@Mobius1D)
- Add timeperstep hook to qrdqn to fix test error (#266) (@albheim)
- Update Flux version (#267) (@findmyway)
- Setup docs generation pipeline (#269) (@findmyway)
- Misc doc related fixes (#270) (@findmyway)
- Update README.md (#271) (@findmyway)
- docs: add JinraeKim as a contributor (#272) (@allcontributors[bot])
- Improve docs GitHub action (#273) (@findmyway)
- Fix docs pipeline (#275) (@findmyway)
- update readme (#276) (@findmyway)
- CompatHelper: add new compat entry for "UnicodePlots" at version "1.3" for package ReinforcementLearningCore (#277) (@github-actions[bot])
- CompatHelper: bump compat for "Distributions" to "0.25" for package ReinforcementLearningCore (#278) (@github-actions[bot])
- CompatHelper: bump compat for "Distributions" to "0.25" for package ReinforcementLearningZoo (#279) (@github-actions[bot])
- docs: add plu70n as a contributor (#282) (@allcontributors[bot])
- Fix bug in CI (#283) (@findmyway)
- Use Weave.jl to generate RLExperiments (#284) (@findmyway)
- QRDQN experiment reproducibility fix (#294) (@ashwani-rathee)
- Add Manifest.toml (#295) (@findmyway)
- docs: add ashwani-rathee as a contributor (#296) (@allcontributors[bot])
- Add basic doc structure (#300) (@findmyway)
- Update guide (#302) (@findmyway)
- Update experiments (#303) (@findmyway)
- fix figs (#304) (@findmyway)
- Fix some simple experiments (#308) (@findmyway)
- add plotting for cartpole and mountaincar with Plots.jl (#309) (@jamblejoe)
- Remove GR in RLEnvs (#310) (@findmyway)
- docs: add jamblejoe as a contributor (#311) (@allcontributors[bot])
- Add compat of [email protected] in ReinforcementLearningExperiments (#312) (@findmyway)
- Add example of SimplexSpace (#313) (@findmyway)
- Improve tutorial (#314) (@findmyway)
- Fix Atari related experiments (#315) (@findmyway)
- CompatHelper: add new compat entry for "ImageTransformations" at version "0.8" for package ReinforcementLearningExperiments (#317) (@github-actions[bot])
- CompatHelper: add new compat entry for "ArcadeLearningEnvironment" at version "0.2" for package ReinforcementLearningExperiments (#318) (@github-actions[bot])
- CompatHelper: add new compat entry for "CUDA" at version "3" for package ReinforcementLearningExperiments (#319) (@github-actions[bot])
- update tips (#321) (@findmyway)
- CompatHelper: bump compat for "GPUArrays" to "7" for package ReinforcementLearningCore (#322) (@github-actions[bot])
- docs: add albheim as a contributor for doc (#323) (@allcontributors[bot])
- Fix broken test (#325) (@Mobius1D)
- Add a warning in docstring of state (#327) (@findmyway)
- Update doc string of PrioritizedDQNLearner (#329) (@findmyway)
- Expand DDPG to multi action spaces (#330) (@Mobius1D)
- CompatHelper: bump compat for "StructArrays" to "0.6" for package ReinforcementLearningZoo (#331) (@github-actions[bot])
- fix 332 (#333) (@findmyway)
- correct spelling in FAQ (#334) (@ultradian)
- docs: add ultradian as a contributor for doc (#335) (@allcontributors[bot])
- fix typo (#338) (@findmyway)
- docs: add eltociear as a contributor for doc (#339) (@allcontributors[bot])
- CompatHelper: bump compat for "FillArrays" to "0.12" for package ReinforcementLearningCore (#340) (@github-actions[bot])
- Add copyto function (#345) (@pilgrimygy)
- add Base.:(==) and Base.hash for AbstractEnv and test nash_conv on KuhnPokerEnv (#348) (@peterchen96)
- Fix legal_actions_mask indexing error in CircularSLART (#350) (@findmyway)
- bump version of RLCore (#351) (@findmyway)
- bump version of RLBae (#352) (@findmyway)
- add LICENSE in RLBase (#353) (@findmyway)
- bump version of RLZoo (#355) (@findmyway)
- docs: add 00krishna as a contributor for bug (#356) (@allcontributors[bot])
- Add the tuning entropy component (#365) (@pilgrimygy)
- Make general components (#370) (@pilgrimygy)
- add weighted_softmax_explorer in the
explorers.jl
(#382) (@peterchen96) - Supplement functions in ReservoirTrajectory and BehaviorCloningPolicy (#390) (@peterchen96)
- Update Flux version (#391) (@findmyway)
- AddSequentialEnv (#394) (@findmyway)
- Throw error in MultiAgentManager if it is applied to a SIMULTANEOUS env (#395) (@findmyway)
- docs: add pkienscherf as a contributor for bug (#396) (@allcontributors[bot])
- Implementation of NFSP and NFSP_KuhnPoker experiment (#402) (@peterchen96)
- Updated RLDatasets.jl (#403) (@Mobius1D)
- Gym d4rl extension (#405) (@Mobius1D)
- updates as per GridWorlds v0.5.0 (#406) (@Sid-Bhatia-0)
- Reduce allocations and update docstring for GaussianNetwork (#414) (@albheim)
- Fix a bug (#415) (@pilgrimygy)
- Expand to d4rl-pybullet (#416) (@Mobius1D)
- docs: add pilgrimygy as a contributor for bug (#417) (@allcontributors[bot])
- Fix 418 (#420) (@findmyway)
- docs: add Krastanov as a contributor for doc (#422) (@allcontributors[bot])
- Make SAC policy use internal rng (#423) (@albheim)
- Add wrapped_env[!] to access env inside wrappers (#426) (@albheim)
- add stock trading env (#428) (@findmyway)
- Add Atari datasets released by Google Research (#429) (@Mobius1D)
- add kwargs to plot(env::) (#430) (@jamblejoe)
- Unify parameter names (#437) (@pilgrimygy)
- docs: add albheim as a contributor for maintenance (#438) (@allcontributors[bot])
- correct
nfsp
implementation (#439) (@peterchen96) - update the nfsp experiment's parameters (#440) (@peterchen96)
- Tiny text typo (#441) (@Nthman)
- docs: add LaarsOman as a contributor for doc (#442) (@allcontributors[bot])
- Add pre-train step; VAE component; CRR and PLAS algorithms (#443) (@pilgrimygy)
- add MADDPG algorithm (#444) (@peterchen96)
- add_Graph_Shortest_Path (#445) (@findmyway)
- try to fix bugs of ActionTransformedEnv (#447) (@peterchen96)
- Update report (#448) (@pilgrimygy)
- Summer ospp project 210370190 mid-term report (#449) (@peterchen96)
- add maxdepth kwarg to remove print_tree() deprecation warning (#450) (@burmecia)
- docs: add burmecia as a contributor for code (#451) (@allcontributors[bot])
- RL unplugged implementation with tests (#452) (@Mobius1D)
- Update report (#453) (@Mobius1D)
- disable notebook generation (#454) (@johnnychen94)
- Revert "Update report" (#456) (@findmyway)
- Update report (#457) (@pilgrimygy)
- fix installation docs (#458) (@Mobius1D)
- docs: add peterchen96 as a contributor for code, doc (#459) (@allcontributors[bot])
- Create LICENSE (#461) (@Mobius1D)
- Add docs (#462) (@Mobius1D)
- Fix make.jl (#463) (@Mobius1D)
- Delete LICENSE (#465) (@Mobius1D)
- fix CI (#466) (@findmyway)
- Fix RLDatasets.jl documentation (#467) (@Mobius1D)
- update report (#468) (@peterchen96)
- Fix ci (#469) (@Mobius1D)
- Update maddpg and the report (#470) (@peterchen96)
- Report (#474) (@pilgrimygy)
- Control whether
run
displays description of experime...
ReinforcementLearningCore-v0.13.1
ReinforcementLearningCore ReinforcementLearningCore-v0.13.1
Merged pull requests:
- Fix deprecations (#10) (@femtocleaner[bot])
- implement epsilon-greedy policy with parametric type (#12) (@jbrea)
- improve docs (#13) (@jbrea)
- refactor policies (#15) (@jbrea)
- Add ReinforcementLearningBase as dependent (#16) (@jbrea)
- fix examples (#18) (@jbrea)
- refactor existing components (#26) (@findmyway)
- Prioritized dqn (#29) (@findmyway)
- add double dqn (#30) (@findmyway)
- add rainbow (#31) (@findmyway)
- use new api in ReinforcementLearningEnvironments.jl (#33) (@findmyway)
- bugfix and api simplification (#34) (@findmyway)
- Switch Tracker.jl to Zygote.jl (#37) (@findmyway)
- Support both Knet and Flux(with Zygote) (#38) (@findmyway)
- add docs (#39) (@findmyway)
- export AbstractActionSelector and add more comments (#42) (@findmyway)
- Refactor buffer (#45) (@findmyway)
- fix example in doc && update examples (#46) (@findmyway)
- fix a performance bug in rainbow (#47) (@findmyway)
- update dependencies (#48) (@findmyway)
- update dependencies and docs (#49) (@findmyway)
- update benchmark for circular_array_buffer (#50) (@findmyway)
- Install TagBot as a GitHub Action (#53) (@JuliaTagBot)
- clean up code (#54) (@findmyway)
- add compat (#55) (@findmyway)
- CompatHelper: add new compat entry for "Reexport" at version "0.2" (#56) (@github-actions[bot])
- add documentation stage in travis (#57) (@findmyway)
- Add doc in travis (#58) (@findmyway)
- Fix link in docs/src/index.md (#60) (@amanbh)
- Update doc (#61) (@findmyway)
- Update README.md & website link (#70) (@findmyway)
- Update dependency (#78) (@findmyway)
- MassInstallAction: Install the CompatHelper workflow on this repository (#99) (@findmyway)
- CompatHelper: bump compat for "Reexport" to "1.0" (#172) (@github-actions[bot])
- update dependency (#177) (@findmyway)
- Add Dockerfile (#187) (@findmyway)
- Update readme (#188) (@findmyway)
- docs: add findmyway as a contributor (#189) (@allcontributors[bot])
- docs: add drozzy as a contributor (#195) (@allcontributors[bot])
- docs: add rcnlee as a contributor (#199) (@allcontributors[bot])
- docs: add norci as a contributor (#200) (@allcontributors[bot])
- docs: add xiruizhao as a contributor (#203) (@allcontributors[bot])
- docs: add metab0t as a contributor (#204) (@allcontributors[bot])
- docs: add albheim as a contributor (#207) (@allcontributors[bot])
- docs: add michelangelo21 as a contributor (#214) (@allcontributors[bot])
- docs: add pilgrimygy as a contributor (#216) (@allcontributors[bot])
- docs: add Mobius1D as a contributor (#218) (@allcontributors[bot])
- docs: add ilancoulon as a contributor (#224) (@allcontributors[bot])
- docs: add pilgrimygy as a contributor (#230) (@allcontributors[bot])
- docs: add JinraeKim as a contributor (#243) (@allcontributors[bot])
- Prepare v0.9 (#252) (@findmyway)
- docs: add luigiannelli as a contributor (#254) (@allcontributors[bot])
- docs: add JBoerma as a contributor (#255) (@allcontributors[bot])
- CompatHelper: bump compat for "ReinforcementLearningEnvironments" to "0.5" (#260) (@github-actions[bot])
- Fix inconsitencies in wrappers (#263) (@albheim)
- setup CI for each subpackages (#264) (@findmyway)
- Fix atari experiments (#265) (@Mobius1D)
- Add timeperstep hook to qrdqn to fix test error (#266) (@albheim)
- Update Flux version (#267) (@findmyway)
- Setup docs generation pipeline (#269) (@findmyway)
- Misc doc related fixes (#270) (@findmyway)
- Update README.md (#271) (@findmyway)
- docs: add JinraeKim as a contributor (#272) (@allcontributors[bot])
- Improve docs GitHub action (#273) (@findmyway)
- Fix docs pipeline (#275) (@findmyway)
- update readme (#276) (@findmyway)
- CompatHelper: add new compat entry for "UnicodePlots" at version "1.3" for package ReinforcementLearningCore (#277) (@github-actions[bot])
- CompatHelper: bump compat for "Distributions" to "0.25" for package ReinforcementLearningCore (#278) (@github-actions[bot])
- CompatHelper: bump compat for "Distributions" to "0.25" for package ReinforcementLearningZoo (#279) (@github-actions[bot])
- docs: add plu70n as a contributor (#282) (@allcontributors[bot])
- Fix bug in CI (#283) (@findmyway)
- Use Weave.jl to generate RLExperiments (#284) (@findmyway)
- QRDQN experiment reproducibility fix (#294) (@ashwani-rathee)
- Add Manifest.toml (#295) (@findmyway)
- docs: add ashwani-rathee as a contributor (#296) (@allcontributors[bot])
- Add basic doc structure (#300) (@findmyway)
- Update guide (#302) (@findmyway)
- Update experiments (#303) (@findmyway)
- fix figs (#304) (@findmyway)
- Fix some simple experiments (#308) (@findmyway)
- add plotting for cartpole and mountaincar with Plots.jl (#309) (@jamblejoe)
- Remove GR in RLEnvs (#310) (@findmyway)
- docs: add jamblejoe as a contributor (#311) (@allcontributors[bot])
- Add compat of [email protected] in ReinforcementLearningExperiments (#312) (@findmyway)
- Add example of SimplexSpace (#313) (@findmyway)
- Improve tutorial (#314) (@findmyway)
- Fix Atari related experiments (#315) (@findmyway)
- CompatHelper: add new compat entry for "ImageTransformations" at version "0.8" for package ReinforcementLearningExperiments (#317) (@github-actions[bot])
- CompatHelper: add new compat entry for "ArcadeLearningEnvironment" at version "0.2" for package ReinforcementLearningExperiments (#318) (@github-actions[bot])
- CompatHelper: add new compat entry for "CUDA" at version "3" for package ReinforcementLearningExperiments (#319) (@github-actions[bot])
- update tips (#321) (@findmyway)
- CompatHelper: bump compat for "GPUArrays" to "7" for package ReinforcementLearningCore (#322) (@github-actions[bot])
- docs: add albheim as a contributor for doc (#323) (@allcontributors[bot])
- Fix broken test (#325) (@Mobius1D)
- Add a warning in docstring of state (#327) (@findmyway)
- Update doc string of PrioritizedDQNLearner (#329) (@findmyway)
- Expand DDPG to multi action spaces (#330) (@Mobius1D)
- CompatHelper: bump compat for "StructArrays" to "0.6" for package ReinforcementLearningZoo (#331) (@github-actions[bot])
- fix 332 (#333) (@findmyway)
- correct spelling in FAQ (#334) (@ultradian)
- docs: add ultradian as a contributor for doc (#335) (@allcontributors[bot])
- fix typo (#338) (@findmyway)
- docs: add eltociear as a contributor for doc (#339) (@allcontributors[bot])
- CompatHelper: bump compat for "FillArrays" to "0.12" for package ReinforcementLearningCore (#340) (@github-actions[bot])
- Add copyto function (#345) (@pilgrimygy)
- add Base.:(==) and Base.hash for AbstractEnv and test nash_conv on KuhnPokerEnv (#348) (@peterchen96)
- Fix legal_actions_mask indexing error in CircularSLART (#350) (@findmyway)
- bump version of RLCore (#351) (@findmyway)
- bump version of RLBae (#352) (@findmyway)
- add LICENSE in RLBase (#353) (@findmyway)
- bump version of RLZoo (#355) (@findmyway)
- docs: add 00krishna as a contributor for bug (#356) (@allcontributors[bot])
- Add the tuning entropy component (#365) (@pilgrimygy)
- Make general components (#370) (@pilgrimygy)
- add weighted_softmax_explorer in the
explorers.jl
(#382) (@peterchen96) - Supplement functions in ReservoirTrajectory and BehaviorCloningPolicy (#390) (@peterchen96)
- Update Flux version (#391) (@findmyway)
- AddSequentialEnv (#394) (@findmyway)
- Throw error in MultiAgentManager if it is applied to a SIMULTANEOUS env (#395) (@findmyway)
- docs: add pkienscherf as a contributor for bug (#396) (@allcontributors[bot])
- Implementation of NFSP and NFSP_KuhnPoker experiment (#402) (@peterchen96)
- Updated RLDatasets.jl (#403) (@Mobius1D)
- Gym d4rl extension (#405) (@Mobius1D)
- updates as per GridWorlds v0.5.0 (#406) (@Sid-Bhatia-0)
- Reduce allocations and update docstring for GaussianNetwork (#414) (@albheim)
- Fix a bug (#415) (@pilgrimygy)
- Expand to d4rl-pybullet (#416) (@Mobius1D)
- docs: add pilgrimygy as a contributor for bug (#417) (@allcontributors[bot])
- Fix 418 (#420) (@findmyway)
- docs: add Krastanov as a contributor for doc (#422) (@allcontributors[bot])
- Make SAC policy use internal rng (#423) (@albheim)
- Add wrapped_env[!] to access env inside wrappers (#426) (@albheim)
- add stock trading env (#428) (@findmyway)
- Add Atari datasets released by Google Research (#429) (@Mobius1D)
- add kwargs to plot(env::) (#430) (@jamblejoe)
- Unify parameter names (#437) (@pilgrimygy)
- docs: add albheim as a contributor for maintenance (#438) (@allcontributors[bot])
- correct
nfsp
implementation (#439) (@peterchen96) - update the nfsp experiment's parameters (#440) (@peterchen96)
- Tiny text typo (#441) (@Nthman)
- docs: add LaarsOman as a contributor for doc (#442) (@allcontributors[bot])
- Add pre-train step; VAE component; CRR and PLAS algorithms (#443) (@pilgrimygy)
- add MADDPG algorithm (#444) (@peterchen96)
- add_Graph_Shortest_Path (#445) (@findmyway)
- try to fix bugs of ActionTransformedEnv (#447) (@peterchen96)
- Update report (#448) (@pilgrimygy)
- Summer ospp project 210370190 mid-term report (#449) (@peterchen96)
- add maxdepth kwarg to remove print_tree() deprecation warning (#450) (@burmecia)
- docs: add burmecia as a contributor for code (#451) (@allcontributors[bot])
- RL unplugged implementation with tests (#452) (@Mobius1D)
- Update report (#453) (@Mobius1D)
- disable notebook generation (#454) (@johnnychen94)
- Revert "Update report" (#456) (@findmyway)
- Update report (#457) (@pilgrimygy)
- fix installation docs (#458) (@Mobius1D)
- docs: add peterchen96 as a contributor for code, doc (#459) (@allcontributors[bot])
- Create LICENSE (#461) (@Mobius1D)
- Add docs (#462) (@Mobius1D)
- Fix make.jl (#463) (@Mobius1D)
- Delete LICENSE (#465) (@Mobius1D)
- fix CI (#466) (@findmyway)
- Fix RLDatasets.jl documentation (#467) (@Mobius1D)
- update report (#468) (@peterchen96)
- Fix ci (#469) (@Mobius1D)
- Update maddpg and the report (#470) (@peterchen96)
- Report (#474) (@pilgrimygy)
- Control whether
run
displays description of experiment (#477) (@shu...
ReinforcementLearningBase-v0.12.2
ReinforcementLearningBase ReinforcementLearningBase-v0.12.2
Merged pull requests:
- Fix deprecations (#10) (@femtocleaner[bot])
- implement epsilon-greedy policy with parametric type (#12) (@jbrea)
- improve docs (#13) (@jbrea)
- refactor policies (#15) (@jbrea)
- Add ReinforcementLearningBase as dependent (#16) (@jbrea)
- fix examples (#18) (@jbrea)
- refactor existing components (#26) (@findmyway)
- Prioritized dqn (#29) (@findmyway)
- add double dqn (#30) (@findmyway)
- add rainbow (#31) (@findmyway)
- use new api in ReinforcementLearningEnvironments.jl (#33) (@findmyway)
- bugfix and api simplification (#34) (@findmyway)
- Switch Tracker.jl to Zygote.jl (#37) (@findmyway)
- Support both Knet and Flux(with Zygote) (#38) (@findmyway)
- add docs (#39) (@findmyway)
- export AbstractActionSelector and add more comments (#42) (@findmyway)
- Refactor buffer (#45) (@findmyway)
- fix example in doc && update examples (#46) (@findmyway)
- fix a performance bug in rainbow (#47) (@findmyway)
- update dependencies (#48) (@findmyway)
- update dependencies and docs (#49) (@findmyway)
- update benchmark for circular_array_buffer (#50) (@findmyway)
- Install TagBot as a GitHub Action (#53) (@JuliaTagBot)
- clean up code (#54) (@findmyway)
- add compat (#55) (@findmyway)
- CompatHelper: add new compat entry for "Reexport" at version "0.2" (#56) (@github-actions[bot])
- add documentation stage in travis (#57) (@findmyway)
- Add doc in travis (#58) (@findmyway)
- Fix link in docs/src/index.md (#60) (@amanbh)
- Update doc (#61) (@findmyway)
- Update README.md & website link (#70) (@findmyway)
- Update dependency (#78) (@findmyway)
- MassInstallAction: Install the CompatHelper workflow on this repository (#99) (@findmyway)
- CompatHelper: bump compat for "Reexport" to "1.0" (#172) (@github-actions[bot])
- update dependency (#177) (@findmyway)
- Add Dockerfile (#187) (@findmyway)
- Update readme (#188) (@findmyway)
- docs: add findmyway as a contributor (#189) (@allcontributors[bot])
- docs: add drozzy as a contributor (#195) (@allcontributors[bot])
- docs: add rcnlee as a contributor (#199) (@allcontributors[bot])
- docs: add norci as a contributor (#200) (@allcontributors[bot])
- docs: add xiruizhao as a contributor (#203) (@allcontributors[bot])
- docs: add metab0t as a contributor (#204) (@allcontributors[bot])
- docs: add albheim as a contributor (#207) (@allcontributors[bot])
- docs: add michelangelo21 as a contributor (#214) (@allcontributors[bot])
- docs: add pilgrimygy as a contributor (#216) (@allcontributors[bot])
- docs: add Mobius1D as a contributor (#218) (@allcontributors[bot])
- docs: add ilancoulon as a contributor (#224) (@allcontributors[bot])
- docs: add pilgrimygy as a contributor (#230) (@allcontributors[bot])
- docs: add JinraeKim as a contributor (#243) (@allcontributors[bot])
- Prepare v0.9 (#252) (@findmyway)
- docs: add luigiannelli as a contributor (#254) (@allcontributors[bot])
- docs: add JBoerma as a contributor (#255) (@allcontributors[bot])
- CompatHelper: bump compat for "ReinforcementLearningEnvironments" to "0.5" (#260) (@github-actions[bot])
- Fix inconsitencies in wrappers (#263) (@albheim)
- setup CI for each subpackages (#264) (@findmyway)
- Fix atari experiments (#265) (@Mobius1D)
- Add timeperstep hook to qrdqn to fix test error (#266) (@albheim)
- Update Flux version (#267) (@findmyway)
- Setup docs generation pipeline (#269) (@findmyway)
- Misc doc related fixes (#270) (@findmyway)
- Update README.md (#271) (@findmyway)
- docs: add JinraeKim as a contributor (#272) (@allcontributors[bot])
- Improve docs GitHub action (#273) (@findmyway)
- Fix docs pipeline (#275) (@findmyway)
- update readme (#276) (@findmyway)
- CompatHelper: add new compat entry for "UnicodePlots" at version "1.3" for package ReinforcementLearningCore (#277) (@github-actions[bot])
- CompatHelper: bump compat for "Distributions" to "0.25" for package ReinforcementLearningCore (#278) (@github-actions[bot])
- CompatHelper: bump compat for "Distributions" to "0.25" for package ReinforcementLearningZoo (#279) (@github-actions[bot])
- docs: add plu70n as a contributor (#282) (@allcontributors[bot])
- Fix bug in CI (#283) (@findmyway)
- Use Weave.jl to generate RLExperiments (#284) (@findmyway)
- QRDQN experiment reproducibility fix (#294) (@ashwani-rathee)
- Add Manifest.toml (#295) (@findmyway)
- docs: add ashwani-rathee as a contributor (#296) (@allcontributors[bot])
- Add basic doc structure (#300) (@findmyway)
- Update guide (#302) (@findmyway)
- Update experiments (#303) (@findmyway)
- fix figs (#304) (@findmyway)
- Fix some simple experiments (#308) (@findmyway)
- add plotting for cartpole and mountaincar with Plots.jl (#309) (@jamblejoe)
- Remove GR in RLEnvs (#310) (@findmyway)
- docs: add jamblejoe as a contributor (#311) (@allcontributors[bot])
- Add compat of [email protected] in ReinforcementLearningExperiments (#312) (@findmyway)
- Add example of SimplexSpace (#313) (@findmyway)
- Improve tutorial (#314) (@findmyway)
- Fix Atari related experiments (#315) (@findmyway)
- CompatHelper: add new compat entry for "ImageTransformations" at version "0.8" for package ReinforcementLearningExperiments (#317) (@github-actions[bot])
- CompatHelper: add new compat entry for "ArcadeLearningEnvironment" at version "0.2" for package ReinforcementLearningExperiments (#318) (@github-actions[bot])
- CompatHelper: add new compat entry for "CUDA" at version "3" for package ReinforcementLearningExperiments (#319) (@github-actions[bot])
- update tips (#321) (@findmyway)
- CompatHelper: bump compat for "GPUArrays" to "7" for package ReinforcementLearningCore (#322) (@github-actions[bot])
- docs: add albheim as a contributor for doc (#323) (@allcontributors[bot])
- Fix broken test (#325) (@Mobius1D)
- Add a warning in docstring of state (#327) (@findmyway)
- Update doc string of PrioritizedDQNLearner (#329) (@findmyway)
- Expand DDPG to multi action spaces (#330) (@Mobius1D)
- CompatHelper: bump compat for "StructArrays" to "0.6" for package ReinforcementLearningZoo (#331) (@github-actions[bot])
- fix 332 (#333) (@findmyway)
- correct spelling in FAQ (#334) (@ultradian)
- docs: add ultradian as a contributor for doc (#335) (@allcontributors[bot])
- fix typo (#338) (@findmyway)
- docs: add eltociear as a contributor for doc (#339) (@allcontributors[bot])
- CompatHelper: bump compat for "FillArrays" to "0.12" for package ReinforcementLearningCore (#340) (@github-actions[bot])
- Add copyto function (#345) (@pilgrimygy)
- add Base.:(==) and Base.hash for AbstractEnv and test nash_conv on KuhnPokerEnv (#348) (@peterchen96)
- Fix legal_actions_mask indexing error in CircularSLART (#350) (@findmyway)
- bump version of RLCore (#351) (@findmyway)
- bump version of RLBae (#352) (@findmyway)
- add LICENSE in RLBase (#353) (@findmyway)
- bump version of RLZoo (#355) (@findmyway)
- docs: add 00krishna as a contributor for bug (#356) (@allcontributors[bot])
- Add the tuning entropy component (#365) (@pilgrimygy)
- Make general components (#370) (@pilgrimygy)
- add weighted_softmax_explorer in the
explorers.jl
(#382) (@peterchen96) - Supplement functions in ReservoirTrajectory and BehaviorCloningPolicy (#390) (@peterchen96)
- Update Flux version (#391) (@findmyway)
- AddSequentialEnv (#394) (@findmyway)
- Throw error in MultiAgentManager if it is applied to a SIMULTANEOUS env (#395) (@findmyway)
- docs: add pkienscherf as a contributor for bug (#396) (@allcontributors[bot])
- Implementation of NFSP and NFSP_KuhnPoker experiment (#402) (@peterchen96)
- Updated RLDatasets.jl (#403) (@Mobius1D)
- Gym d4rl extension (#405) (@Mobius1D)
- updates as per GridWorlds v0.5.0 (#406) (@Sid-Bhatia-0)
- Reduce allocations and update docstring for GaussianNetwork (#414) (@albheim)
- Fix a bug (#415) (@pilgrimygy)
- Expand to d4rl-pybullet (#416) (@Mobius1D)
- docs: add pilgrimygy as a contributor for bug (#417) (@allcontributors[bot])
- Fix 418 (#420) (@findmyway)
- docs: add Krastanov as a contributor for doc (#422) (@allcontributors[bot])
- Make SAC policy use internal rng (#423) (@albheim)
- Add wrapped_env[!] to access env inside wrappers (#426) (@albheim)
- add stock trading env (#428) (@findmyway)
- Add Atari datasets released by Google Research (#429) (@Mobius1D)
- add kwargs to plot(env::) (#430) (@jamblejoe)
- Unify parameter names (#437) (@pilgrimygy)
- docs: add albheim as a contributor for maintenance (#438) (@allcontributors[bot])
- correct
nfsp
implementation (#439) (@peterchen96) - update the nfsp experiment's parameters (#440) (@peterchen96)
- Tiny text typo (#441) (@Nthman)
- docs: add LaarsOman as a contributor for doc (#442) (@allcontributors[bot])
- Add pre-train step; VAE component; CRR and PLAS algorithms (#443) (@pilgrimygy)
- add MADDPG algorithm (#444) (@peterchen96)
- add_Graph_Shortest_Path (#445) (@findmyway)
- try to fix bugs of ActionTransformedEnv (#447) (@peterchen96)
- Update report (#448) (@pilgrimygy)
- Summer ospp project 210370190 mid-term report (#449) (@peterchen96)
- add maxdepth kwarg to remove print_tree() deprecation warning (#450) (@burmecia)
- docs: add burmecia as a contributor for code (#451) (@allcontributors[bot])
- RL unplugged implementation with tests (#452) (@Mobius1D)
- Update report (#453) (@Mobius1D)
- disable notebook generation (#454) (@johnnychen94)
- Revert "Update report" (#456) (@findmyway)
- Update report (#457) (@pilgrimygy)
- fix installation docs (#458) (@Mobius1D)
- docs: add peterchen96 as a contributor for code, doc (#459) (@allcontributors[bot])
- Create LICENSE (#461) (@Mobius1D)
- Add docs (#462) (@Mobius1D)
- Fix make.jl (#463) (@Mobius1D)
- Delete LICENSE (#465) (@Mobius1D)
- fix CI (#466) (@findmyway)
- Fix RLDatasets.jl documentation (#467) (@Mobius1D)
- update report (#468) (@peterchen96)
- Fix ci (#469) (@Mobius1D)
- Update maddpg and the report (#470) (@peterchen96)
- Report (#474) (@pilgrimygy)
- Control whether
run
displays description of experiment (#477) (@shu...