-
Notifications
You must be signed in to change notification settings - Fork 5.9k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[RLlib] Docs do-over (new API stack): Add scaling guide rst page. #49528
[RLlib] Docs do-over (new API stack): Add scaling guide rst page. #49528
Conversation
Signed-off-by: sven1977 <[email protected]>
Signed-off-by: sven1977 <[email protected]>
Signed-off-by: sven1977 <[email protected]>
Signed-off-by: sven1977 <[email protected]>
Signed-off-by: sven1977 <[email protected]>
Signed-off-by: sven1977 <[email protected]>
Signed-off-by: sven1977 <[email protected]>
…redo_scaling_guide
Signed-off-by: sven1977 <[email protected]>
…redo_scaling_guide
Signed-off-by: sven1977 <[email protected]>
Signed-off-by: sven1977 <[email protected]>
Signed-off-by: sven1977 <[email protected]>
Signed-off-by: sven1977 <[email protected]>
Signed-off-by: sven1977 <[email protected]>
doc/source/rllib/rllib-training.rst
Outdated
Configuring RLlib Algorithms | ||
---------------------------- | ||
|
||
You can configure RLlib algorithms in a modular fashion by working with so-called |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
You can configure RLlib algorithms in a modular fashion by working with so-called | |
You can configure RLlib algorithms in a modular fashion by working with |
doc/source/rllib/rllib-training.rst
Outdated
|
||
You can configure RLlib algorithms in a modular fashion by working with so-called | ||
`AlgorithmConfig` objects. | ||
In essence, you first create a `config = AlgorithmConfig()` object and then call methods |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
In essence, you first create a `config = AlgorithmConfig()` object and then call methods | |
First create a `config = AlgorithmConfig()` object and then call methods |
doc/source/rllib/rllib-training.rst
Outdated
In essence, you first create a `config = AlgorithmConfig()` object and then call methods | ||
on it to set the desired configuration options. | ||
Each RLlib algorithm has its own config class that inherits from `AlgorithmConfig`. | ||
For instance, to create a `PPO` algorithm, you start with a `PPOConfig` object, to work |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
For instance, to create a `PPO` algorithm, you start with a `PPOConfig` object, to work | |
For instance, to create a `PPO` algorithm, start with a `PPOConfig` object, to work |
doc/source/rllib/rllib-training.rst
Outdated
on it to set the desired configuration options. | ||
Each RLlib algorithm has its own config class that inherits from `AlgorithmConfig`. | ||
For instance, to create a `PPO` algorithm, you start with a `PPOConfig` object, to work | ||
with a `DQN` algorithm, you start with a `DQNConfig` object, etc. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
with a `DQN` algorithm, you start with a `DQNConfig` object, etc. | |
with a `DQN` algorithm, start with a `DQNConfig` object, etc. |
doc/source/rllib/rllib-training.rst
Outdated
|
||
.. note:: | ||
|
||
Each algorithm has its specific settings, but most configuration options are shared. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Each algorithm has its specific settings, but most configuration options are shared. | |
Each algorithm has its specific settings, but most configuration options are shared. |
doc/source/rllib/scaling-guide.rst
Outdated
single local :py:class:`~ray.rllib.core.learner.learner.Learner` instance (``num_learners=0``), | ||
depends on whether you have a GPU available or not. | ||
If exactly one GPU is available, you should run these two algorithms with ``num_learners=0, num_gpus_per_learner=1``, | ||
if no GPU is available, you should set ``num_learners=1, num_gpus_per_learner=0``. For > 1 GPUs available, you should |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
if no GPU is available, you should set ``num_learners=1, num_gpus_per_learner=0``. For > 1 GPUs available, you should | |
if no GPU is available, set ``num_learners=1, num_gpus_per_learner=0``. For > 1 GPUs available, you should |
Signed-off-by: sven1977 <[email protected]>
Signed-off-by: sven1977 <[email protected]>
Co-authored-by: angelinalg <[email protected]> Signed-off-by: Sven Mika <[email protected]>
Signed-off-by: sven1977 <[email protected]>
…cs_redo_scaling_guide
Signed-off-by: sven1977 <[email protected]>
…_redo_scaling_guide Signed-off-by: sven1977 <[email protected]> # Conflicts: # doc/source/rllib/user-guides.rst
Signed-off-by: sven1977 <[email protected]>
Docs do-over (new API stack):
Why are these changes needed?
Related issue number
Checks
git commit -s
) in this PR.scripts/format.sh
to lint the changes in this PR.method in Tune, I've added it in
doc/source/tune/api/
under thecorresponding
.rst
file.