mirror of
https://github.com/saymrwulf/stable-baselines3.git
synced 2026-05-14 20:58:03 +00:00
Update doc
This commit is contained in:
parent
b1322ff5d6
commit
3a5b743760
3 changed files with 3 additions and 3 deletions
|
|
@ -9,7 +9,7 @@ Try it online with Colab Notebooks!
|
|||
All the following examples can be executed online using Google colab |colab|
|
||||
notebooks:
|
||||
|
||||
- `Full Tutorial <https://github.com/araffin/rl-tutorial-jnrr19>`_
|
||||
- `Full Tutorial <https://github.com/araffin/rl-tutorial-jnrr19/tree/sb3>`_
|
||||
- `All Notebooks <https://github.com/Stable-Baselines-Team/rl-colab-notebooks/tree/sb3>`_
|
||||
- `Getting Started`_
|
||||
- `Training, Saving, Loading`_
|
||||
|
|
|
|||
|
|
@ -15,14 +15,14 @@ General advice when using Reinforcement Learning
|
|||
TL;DR
|
||||
-----
|
||||
|
||||
1. Read about RL and Stable Baselines
|
||||
1. Read about RL and Stable Baselines3
|
||||
2. Do quantitative experiments and hyperparameter tuning if needed
|
||||
3. Evaluate the performance using a separate test environment
|
||||
4. For better performance, increase the training budget
|
||||
|
||||
|
||||
Like any other subject, if you want to work with RL, you should first read about it (we have a dedicated `resource page <rl.html>`_ to get you started)
|
||||
to understand what you are using. We also recommend you read Stable Baselines (SB) documentation and do the `tutorial <https://github.com/araffin/rl-tutorial-jnrr19>`_.
|
||||
to understand what you are using. We also recommend you read Stable Baselines3 (SB3) documentation and do the `tutorial <https://github.com/araffin/rl-tutorial-jnrr19/tree/sb3>`_.
|
||||
It covers basic usage and guide you towards more advanced concepts of the library (e.g. callbacks and wrappers).
|
||||
|
||||
Reinforcement Learning differs from other machine learning methods in several ways. The data used to train the agent is collected
|
||||
|
|
|
|||
Loading…
Reference in a new issue