1
0
mirror of https://github.com/gryf/coach.git synced 2025-12-17 11:10:20 +01:00
Commit Graph

28 Commits

Author SHA1 Message Date
Gal Leibovich
138ced23ba RL in Large Discrete Action Spaces - Wolpertinger Agent (#394)
* Currently this is specific to the case of discretizing a continuous action space. Can easily be adapted to other case by feeding the kNN otherwise, and removing the usage of a discretizing output action filter
2019-09-08 12:53:49 +03:00
Gal Novik
92460736bc Updated tutorial and docs (#386)
Improved getting started tutorial, and updated docs to point to version 1.0.0
2019-08-05 16:46:15 +03:00
Gal Leibovich
19ad2d60a7 Batch RL Tutorial (#372) 2019-07-14 18:43:48 +03:00
Gal Leibovich
7eb884c5b2 TD3 (#338) 2019-06-16 11:11:21 +03:00
anabwan
ddffac8570 fixed release version (#333)
* fixed release version

* update docs
2019-05-28 11:11:15 +03:00
anabwan
342b7184bc Enabling Coach Documentation to be run even when environments are not installed (#326) 2019-05-27 10:46:07 +03:00
guyk1971
74db141d5e SAC algorithm (#282)
* SAC algorithm

* SAC - updates to agent (learn_from_batch), sac_head and sac_q_head to fix problem in gradient calculation. Now SAC agents is able to train.
gym_environment - fixing an error in access to gym.spaces

* Soft Actor Critic - code cleanup

* code cleanup

* V-head initialization fix

* SAC benchmarks

* SAC Documentation

* typo fix

* documentation fixes

* documentation and version update

* README typo
2019-05-01 18:37:49 +03:00
shadiendrawis
2b5d1dabe6 ACER algorithm (#184)
* initial ACER commit

* Code cleanup + several fixes

* Q-retrace bug fix + small clean-ups

* added documentation for acer

* ACER benchmarks

* update benchmarks table

* Add nightly running of golden and trace tests. (#202)

Resolves #200

* comment out nightly trace tests until values reset.

* remove redundant observe ignore (#168)

* ensure nightly test env containers exist. (#205)

Also bump integration test timeout

* wxPython removal (#207)

Replacing wxPython with Python's Tkinter.
Also removing the option to choose multiple files as it is unused and causes errors, and fixing the load file/directory spinner.

* Create CONTRIBUTING.md (#210)

* Create CONTRIBUTING.md.  Resolves #188

* run nightly golden tests sequentially. (#217)

Should reduce resource requirements and potential CPU contention but increases
overall execution time.

* tests: added new setup configuration + test args (#211)

- added utils for future tests and conftest
- added test args

* new docs build

* golden test update
2019-02-20 23:52:34 +02:00
Zach Dwiel
cd812b0d25 more clear names for methods of Space (#181)
* rename Space.val_matches_space_definition -> contains; Space.is_point_in_space_shape -> valid_index
* rename valid_index -> is_valid_index
2019-01-14 15:02:53 -05:00
Gal Leibovich
f12857a8c7 Docs changes - fixing blogpost links, removing importing all exploration policies (#139)
* updated docs

* removing imports for all exploration policies in __init__ + setting the right blog-post link

* small cleanups
2018-12-05 16:16:16 -05:00
Gal Novik
1e618647ab adding .nojekyll file for github pages to function properly 2018-11-27 22:35:16 +02:00
Gal Novik
7e3aca22eb Documentation fix 2018-11-27 22:32:46 +02:00
Balaji Subramaniam
d06197f663 Add documentation on distributed Coach. (#158)
* Added documentation on distributed Coach.
2018-11-27 12:26:15 +02:00
Itai Caspi
6d40ad1650 update of api docstrings across coach and tutorials [WIP] (#91)
* updating the documentation website
* adding the built docs
* update of api docstrings across coach and tutorials 0-2
* added some missing api documentation
* New Sphinx based documentation
2018-11-15 15:00:13 +02:00
Gal Leibovich
8f99409387 updating algorithms.png for README 2018-08-16 16:46:26 +03:00
Gal Novik
19ca5c24b1 pre-release 0.10.0 2018-08-13 17:11:34 +03:00
itaicaspi-intel
5d5562bf62 moving the docs to github 2018-04-23 09:14:20 +03:00
Itai Caspi
125c7ee38d Release 0.9
Main changes are detailed below:

New features -
* CARLA 0.7 simulator integration
* Human control of the game play
* Recording of human game play and storing / loading the replay buffer
* Behavioral cloning agent and presets
* Golden tests for several presets
* Selecting between deep / shallow image embedders
* Rendering through pygame (with some boost in performance)

API changes -
* Improved environment wrapper API
* Added an evaluate flag to allow convenient evaluation of existing checkpoints
* Improve frameskip definition in Gym

Bug fixes -
* Fixed loading of checkpoints for agents with more than one network
* Fixed the N Step Q learning agent python3 compatibility
2017-12-19 19:27:16 +02:00
Zach Dwiel
7bdba396d2 Update add_env.md 2017-11-14 17:57:55 +02:00
Itai Caspi
fd103a7b69 updated the algorithms diagram with QR-DQN 2017-11-01 15:24:54 +02:00
Itai Caspi
a8bce9828c new feature - implementation of Quantile Regression DQN (https://arxiv.org/pdf/1710.10044v1.pdf)
API change - Distributional DQN renamed to Categorical DQN
2017-11-01 15:09:07 +02:00
Gal Leibovich
eb0b57d7fa Updating PPO references per issue #11 2017-10-24 16:57:44 +03:00
Itai Caspi
a1656c2ae6 fixed docs color for mobile 2017-10-23 11:46:27 +03:00
Gal Novik
6009b73eb6 fixed some documentation typos 2017-10-22 22:21:45 +03:00
Gal Leibovich
cc9580a949 updated docs with links to github + a few more words on Dashboard functionality 2017-10-22 16:33:49 +03:00
Itai Caspi
00fca9b6e0 updated the paper links in the docs and restyled the theme 2017-10-19 17:16:12 +03:00
Gal Leibovich
8c708820a9 docs update + removing unused code from parallel_actor 2017-10-19 17:07:30 +03:00
Gal Leibovich
1d4c3455e7 coach v0.8.0 2017-10-19 13:10:15 +03:00