-
Notifications
You must be signed in to change notification settings - Fork 40
/
common.py
333 lines (255 loc) · 10.5 KB
/
common.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
# Copyright (c) Meta Platforms, Inc. and affiliates.
#
# This source code is licensed under the license found in the
# LICENSE file in the root directory of this source tree.
#
from __future__ import annotations
import importlib
import warnings
from enum import Enum
from pathlib import Path
from typing import Any, Callable, Dict, List, Optional
from tensordict import TensorDictBase
from torchrl.data import Composite
from torchrl.envs import EnvBase, RewardSum, Transform
from benchmarl.utils import _read_yaml_config, DEVICE_TYPING
def _type_check_task_config(
environemnt_name: str,
task_name: str,
config: Dict[str, Any],
warn_on_missing_dataclass: bool = True,
):
task_config_class = _get_task_config_class(environemnt_name, task_name)
if task_config_class is not None:
return task_config_class(**config).__dict__
else:
if warn_on_missing_dataclass:
warnings.warn(
"TaskConfig python dataclass not foud, task is being loaded without type checks"
)
return config
def _get_task_config_class(environemnt_name: str, task_name: str):
try:
module = importlib.import_module(
f"{'.'.join(__name__.split('.')[:-1])}.{environemnt_name}.{task_name}"
)
return module.TaskConfig
except ModuleNotFoundError:
return None
class Task(Enum):
"""Task.
Tasks are enums, one enum for each environment.
Each enum member has a config attribute that is a dictionary which can be loaded from .yaml
files. You can also access and modify this attribute directly.
Each new environment should inherit from Task and instantiate its members as
TASK_1 = None
TASK_2 = None
...
Tasks configs are loaded from benchmarl/conf/environments
"""
def __new__(cls, *args, **kwargs):
value = len(cls.__members__) + 1
obj = object.__new__(cls)
obj._value_ = value
return obj
def __init__(self, config: Dict[str, Any]):
self.config = config
def update_config(self, config: Dict[str, Any]) -> Task:
"""
Updates the task config
Args:
config (dictionary): The config to update in the task
Returns: The updated task
"""
if self.config is None:
self.config = config
else:
self.config.update(config)
return self
def get_env_fun(
self,
num_envs: int,
continuous_actions: bool,
seed: Optional[int],
device: DEVICE_TYPING,
) -> Callable[[], EnvBase]:
"""
This function is used to obtain a TorchRL object from the enum Task.
Args:
num_envs (int): The number of envs that should be in the batch_size of the returned env.
In vectorized envs, this can be used to set the number of batched environments.
If your environment is not vectorized, you can just ignore this, and it will be
wrapped in a :class:`torchrl.envs.SerialEnv` with num_envs automatically.
continuous_actions (bool): Whether your environment should have continuous or discrete actions.
If your environment does not support both, ignore this and refer to the supports_x_actions methods.
seed (optional, int): The seed of your env
device (str): the device of your env, you can pass this to any torchrl env constructor
Returns: a function that takes no arguments and returns a :class:`torchrl.envs.EnvBase` object
"""
raise NotImplementedError
def supports_continuous_actions(self) -> bool:
"""
Return true if your task supports continuous actions.
If true, self.get_env_fun might be called with continuous_actions=True
"""
raise NotImplementedError
def supports_discrete_actions(self) -> bool:
"""
Return true if your task supports discrete actions.
If true, self.get_env_fun might be called with continuous_actions=False
"""
raise NotImplementedError
def max_steps(self, env: EnvBase) -> int:
"""
The maximum number of steps allowed in an evaluation rollout.
Args:
env (EnvBase): An environment created via self.get_env_fun
"""
raise NotImplementedError
def has_render(self, env: EnvBase) -> bool:
"""
If env.render() should be called on the environment
Args:
env (EnvBase): An environment created via self.get_env_fun
"""
raise NotImplementedError
def group_map(self, env: EnvBase) -> Dict[str, List[str]]:
"""
The group_map mapping agents groups to agent names.
This should be reelected in the TensorDicts coming from the environment where
agent data is supposed to be stacked according to this.
Args:
env (EnvBase): An environment created via self.get_env_fun
"""
raise NotImplementedError
def observation_spec(self, env: EnvBase) -> Composite:
"""
A spec for the observation.
Must be a Composite with as many entries as needed nested under the ``group_name`` key.
Args:
env (EnvBase): An environment created via self.get_env_fun
Examples:
>>> print(task.observation_spec(env))
Composite(
agents: Composite(
observation: Composite(
image: UnboundedDiscreteTensorSpec(
shape=torch.Size([8, 88, 88, 3]),
space=ContinuousBox(
low=Tensor(shape=torch.Size([8, 88, 88, 3]), device=cpu, dtype=torch.int64, contiguous=True),
high=Tensor(shape=torch.Size([8, 88, 88, 3]), device=cpu, dtype=torch.int64, contiguous=True)),
device=cpu,
dtype=torch.uint8,
domain=discrete),
array: Unbounded(
shape=torch.Size([8, 3]),
space=None,
device=cpu,
dtype=torch.float32,
domain=continuous), device=cpu, shape=torch.Size([8])), device=cpu, shape=torch.Size([8])), device=cpu, shape=torch.Size([]))
"""
raise NotImplementedError
def info_spec(self, env: EnvBase) -> Optional[Composite]:
"""
A spec for the info.
If provided, must be a Composite with one (group_name, "info") entry per group (this entry can be composite).
Args:
env (EnvBase): An environment created via self.get_env_fun
"""
raise NotImplementedError
def state_spec(self, env: EnvBase) -> Optional[Composite]:
"""
A spec for the state.
If provided, must be a Composite with one entry.
Args:
env (EnvBase): An environment created via self.get_env_fun
"""
raise NotImplementedError
def action_spec(self, env: EnvBase) -> Composite:
"""
A spec for the action.
If provided, must be a Composite with one (group_name, "action") entry per group.
Args:
env (EnvBase): An environment created via self.get_env_fun
"""
raise NotImplementedError
def action_mask_spec(self, env: EnvBase) -> Optional[Composite]:
"""
A spec for the action mask.
If provided, must be a Composite with one (group_name, "action_mask") entry per group.
Args:
env (EnvBase): An environment created via self.get_env_fun
"""
raise NotImplementedError
@staticmethod
def env_name() -> str:
"""
The name of the environment in the benchmarl/conf/task folder
"""
raise NotImplementedError
@staticmethod
def log_info(batch: TensorDictBase) -> Dict[str, float]:
"""
Return a str->float dict with extra items to log.
This function has access to the collected batch and is optional.
Args:
batch (TensorDictBase): the batch obtained from collection.
"""
return {}
def get_reward_sum_transform(self, env: EnvBase) -> Transform:
"""
Returns the RewardSum transform for the environment
Args:
env (EnvBase): An environment created via self.get_env_fun
"""
if "_reset" in env.reset_keys:
reset_keys = ["_reset"] * len(self.group_map(env).keys())
else:
reset_keys = env.reset_keys
return RewardSum(reset_keys=reset_keys)
def get_env_transforms(self, env: EnvBase) -> List[Transform]:
"""
Returns a list of :class:`torchrl.envs.Transform` to be applied to the env.
Args:
env (EnvBase): An environment created via self.get_env_fun
"""
return []
def get_replay_buffer_transforms(self, env: EnvBase) -> List[Transform]:
"""
Returns a list of :class:`torchrl.envs.Transform` to be applied to the :class:`torchrl.data.ReplayBuffer`.
Args:
env (EnvBase): An environment created via self.get_env_fun
"""
return []
@staticmethod
def render_callback(experiment, env: EnvBase, data: TensorDictBase):
try:
return env.render(mode="rgb_array")
except TypeError:
return env.render()
def __repr__(self):
cls_name = self.__class__.__name__
return f"{cls_name}.{self.name}: (config={self.config})"
def __str__(self):
return self.__repr__()
@staticmethod
def _load_from_yaml(name: str) -> Dict[str, Any]:
yaml_path = Path(__file__).parent.parent / "conf" / "task" / f"{name}.yaml"
return _read_yaml_config(str(yaml_path.resolve()))
def get_from_yaml(self, path: Optional[str] = None) -> Task:
"""
Load the task configuration from yaml
Args:
path (str, optional): The full path of the yaml file to load from. If None, it will default to
benchmarl/conf/task/self.env_name()/self.name
Returns: the task with the loaded config
"""
environment_name = self.env_name()
task_name = self.name.lower()
full_name = str(Path(environment_name) / Path(task_name))
if path is None:
config = Task._load_from_yaml(full_name)
else:
config = _read_yaml_config(path)
config = _type_check_task_config(environment_name, task_name, config)
return self.update_config(config)