You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I would like to run an experiment with independent learning on a heterogeneous setting where agents get individual and different rewards.
The default setting for group_map for MeltingPot envs is ALL_IN_ONE_GROUP.
When I switch it to ONE_GROUP_PER_AGENT (as suggested by TorchRL's documentation), an error occurs from MeltingPotTask.get_replay_buffer_transforms() as the method creates a transformation with all keys and the transformation is applied to each player's TensorDict where only that player's keys exists. For example, a transformation with keys of "player_0" and "player_1" is being applied to a TensorDict that contains only a key for "player_0".
For the moment, I disabled it.
I think this issue could be solved by making the replay buffer transformations only if the key exists, similar to ExcludeTransform.
The text was updated successfully, but these errors were encountered:
I would like to run an experiment with independent learning on a heterogeneous setting where agents get individual and different rewards.
The default setting for group_map for MeltingPot envs is ALL_IN_ONE_GROUP.
When I switch it to ONE_GROUP_PER_AGENT (as suggested by TorchRL's documentation), an error occurs from MeltingPotTask.get_replay_buffer_transforms() as the method creates a transformation with all keys and the transformation is applied to each player's TensorDict where only that player's keys exists. For example, a transformation with keys of "player_0" and "player_1" is being applied to a TensorDict that contains only a key for "player_0".
For the moment, I disabled it.
I think this issue could be solved by making the replay buffer transformations only if the key exists, similar to ExcludeTransform.
The text was updated successfully, but these errors were encountered: