Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Addressing paper feedbacks regarding the documentation #265

Merged
merged 2 commits into from
Jun 11, 2024
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
6 changes: 4 additions & 2 deletions docs/Documentation/docsCoverage.md
Original file line number Diff line number Diff line change
Expand Up @@ -15,10 +15,12 @@ The **Coverage Environment** is our second training environment, building on the
::: warning Warning
The DSSE project requires Python version 3.10.5 or higher.

The installation of GDAL (requirement for using opendrift) may need the installation of the following packages
To install [GDAL](https://gdal.org/) (a requirement for using OpenDrift), you may need to install the following packages:
```sh
sudo apt-get install -y libgdal-dev gdal-bin
sudo apt install -y libgdal-dev gdal-bin
```

For Windows, Microsoft Visual C++ 14.0 or greater is required for building.
:::

::: tip Tip
Expand Down
15 changes: 15 additions & 0 deletions docs/QuickStart/quickStart.md
Original file line number Diff line number Diff line change
Expand Up @@ -63,6 +63,21 @@ while not done:
```
:::

::: details About the agent policy
The "random_policy" function is designed to abstract the concept of a model or function that chooses actions within the environment's action space. In the example below, it samples a random action from the action space and returns a dictionary mapping agents to the actions they should perform next, based on the given observations and number of agents.

```python
def random_policy(obs, agents):
actions = {}
for agent in agents:
actions[agent] = env.action_space(agent).sample()
return actions

actions = random_policy(observations, env.get_agents())
```
Users can adapt this function by incorporating a trained model, such as one implemented with RLlib. For a demonstration, refer to the [Algorithms](../Documentation/docsAlgorithms) section, which shows how to use a model to select actions based on the received observations.
:::

## Coverage Environment

### Install
Expand Down