diff --git a/notebooks/train_and_export_onnx_example_continuous.ipynb b/notebooks/train_and_export_onnx_example_continuous.ipynb index 91782918..007129d0 100644 --- a/notebooks/train_and_export_onnx_example_continuous.ipynb +++ b/notebooks/train_and_export_onnx_example_continuous.ipynb @@ -42,7 +42,7 @@ "import yaml\n", "import torch\n", "import matplotlib.pyplot as plt\n", - "import gym as gymnasium\n", + "import gymnasium as gym\n", "from IPython import display\n", "import numpy as np\n", "import onnx\n", diff --git a/notebooks/train_and_export_onnx_example_discrete.ipynb b/notebooks/train_and_export_onnx_example_discrete.ipynb index f5bddf75..c4783ee5 100644 --- a/notebooks/train_and_export_onnx_example_discrete.ipynb +++ b/notebooks/train_and_export_onnx_example_discrete.ipynb @@ -41,7 +41,7 @@ "import yaml\n", "import torch\n", "import matplotlib.pyplot as plt\n", - "import gym as gymnasium\n", + "import gymnasium as gym\n", "from IPython import display\n", "import numpy as np\n", "import onnx\n", diff --git a/notebooks/train_and_export_onnx_example_lstm_continuous.ipynb b/notebooks/train_and_export_onnx_example_lstm_continuous.ipynb index 97b5e029..4e4738e4 100644 --- a/notebooks/train_and_export_onnx_example_lstm_continuous.ipynb +++ b/notebooks/train_and_export_onnx_example_lstm_continuous.ipynb @@ -46,7 +46,7 @@ "import yaml\n", "import torch\n", "import matplotlib.pyplot as plt\n", - "import gym as gymnasium\n", + "import gymnasium as gym\n", "from IPython import display\n", "import numpy as np\n", "import onnx\n", diff --git a/rl_games/algos_torch/players.py b/rl_games/algos_torch/players.py index 7219cf54..e657dadb 100644 --- a/rl_games/algos_torch/players.py +++ b/rl_games/algos_torch/players.py @@ -2,7 +2,7 @@ from rl_games.algos_torch import torch_ext from rl_games.algos_torch.running_mean_std import RunningMeanStd from rl_games.common.tr_helpers import unsqueeze_obs -import gym as gymnasium +import gymnasium as gym import torch from torch import nn import numpy as np diff --git a/rl_games/algos_torch/sac_agent.py b/rl_games/algos_torch/sac_agent.py index dad8de0c..ba73cc9f 100644 --- a/rl_games/algos_torch/sac_agent.py +++ b/rl_games/algos_torch/sac_agent.py @@ -4,7 +4,7 @@ from rl_games.common import schedulers from rl_games.common import experience from rl_games.common.a2c_common import print_statistics - +from rl_games.common.env_configurations import patch_env_info from rl_games.interfaces.base_algorithm import BaseAlgorithm from torch.utils.tensorboard import SummaryWriter from datetime import datetime @@ -103,7 +103,7 @@ def base_init(self, base_name, config): self.env_info = config.get('env_info') if self.env_info is None: self.vec_env = vecenv.create_vec_env(self.env_name, self.num_actors, **self.env_config) - self.env_info = self.vec_env.get_env_info() + self.env_info = patch_env_info(self.vec_env.get_env_info()) self._device = config.get('device', 'cuda:0') diff --git a/rl_games/common/a2c_common.py b/rl_games/common/a2c_common.py index 64b71abd..e0daa608 100644 --- a/rl_games/common/a2c_common.py +++ b/rl_games/common/a2c_common.py @@ -12,9 +12,10 @@ from rl_games.common.diagnostics import DefaultDiagnostics, PpoDiagnostics from rl_games.algos_torch import model_builder from rl_games.interfaces.base_algorithm import BaseAlgorithm +from rl_games.common.env_configurations import patch_env_info import numpy as np import time -import gym as gymnasium +import gymnasium as gym from datetime import datetime from tensorboardX import SummaryWriter @@ -127,7 +128,7 @@ def __init__(self, base_name, params): self.env_info = config.get('env_info') if self.env_info is None: self.vec_env = vecenv.create_vec_env(self.env_name, self.num_actors, **self.env_config) - self.env_info = self.vec_env.get_env_info() + self.env_info = patch_env_info(self.vec_env.get_env_info()) else: self.vec_env = config.get('vec_env', None) diff --git a/rl_games/common/env_configurations.py b/rl_games/common/env_configurations.py index e71af5e6..b443bdb5 100644 --- a/rl_games/common/env_configurations.py +++ b/rl_games/common/env_configurations.py @@ -5,7 +5,7 @@ from rl_games.envs.brax import create_brax_env from rl_games.envs.envpool import create_envpool from rl_games.envs.cule import create_cule -import gym as gymnasium +import gymnasium as gym from gym.wrappers import FlattenObservation, FilterObservation import numpy as np import math @@ -109,10 +109,10 @@ def create_dm_control_env(**kwargs): return env def create_super_mario_env(name='SuperMarioBros-v1'): - import gym as gymnasium + import gymnasium as gym from nes_py.wrappers import JoypadSpace from gym_super_mario_bros.actions import SIMPLE_MOVEMENT, COMPLEX_MOVEMENT - import gym as gymnasium_super_mario_bros + import gymnasium as gym_super_mario_bros env = gym_super_mario_bros.make(name) env = JoypadSpace(env, SIMPLE_MOVEMENT) @@ -121,11 +121,11 @@ def create_super_mario_env(name='SuperMarioBros-v1'): return env def create_super_mario_env_stage1(name='SuperMarioBrosRandomStage1-v1'): - import gym as gymnasium + import gymnasium as gym from nes_py.wrappers import JoypadSpace from gym_super_mario_bros.actions import SIMPLE_MOVEMENT, COMPLEX_MOVEMENT - import gym as gymnasium_super_mario_bros + import gymnasium as gym_super_mario_bros stage_names = [ 'SuperMarioBros-1-1-v1', 'SuperMarioBros-1-2-v1', @@ -143,13 +143,13 @@ def create_super_mario_env_stage1(name='SuperMarioBrosRandomStage1-v1'): return env def create_quadrupped_env(): - import gym as gymnasium + import gymnasium as gym import roboschool import quadruppedEnv return wrappers.FrameStack(wrappers.MaxAndSkipEnv(gym.make('QuadruppedWalk-v1'), 4, False), 2, True) def create_roboschool_env(name): - import gym as gymnasium + import gymnasium as gym import roboschool return gym.make(name) @@ -203,8 +203,8 @@ def create_test_env(name, **kwargs): return env def create_minigrid_env(name, **kwargs): - import gym as gymnasium_minigrid - import gym as gymnasium_minigrid.wrappers + import gym_minigrid + import gym_minigrid.wrappers state_bonus = kwargs.pop('state_bonus', False) @@ -442,7 +442,7 @@ def get_env_info(env): if hasattr(env, "value_size"): result_shapes['value_size'] = env.value_size print(result_shapes) - return result_shapes + return patch_env_info(result_shapes) def get_obs_and_action_spaces_from_config(config): env_config = config.get('env_config', {}) @@ -453,9 +453,8 @@ def get_obs_and_action_spaces_from_config(config): def patch_env_info(env_info): - import gymnas env_info['observation_space'] = gymnasium_utils.convert_space(env_info['observation_space'] ) - env_info['action_space'] = gymnasium_utils.convert_space(env_info['observation_space'] ) + env_info['action_space'] = gymnasium_utils.convert_space(env_info['action_space'] ) if 'state_space' in env_info: env_info['state_space'] = gymnasium_utils.convert_space(env_info['state_space'] ) return env_info diff --git a/rl_games/common/experience.py b/rl_games/common/experience.py index 0ac7681e..dbebac27 100644 --- a/rl_games/common/experience.py +++ b/rl_games/common/experience.py @@ -1,6 +1,6 @@ import numpy as np import random -import gym as gymnasium +import gymnasium as gym import torch from rl_games.common.segment_tree import SumSegmentTree, MinSegmentTree import torch diff --git a/rl_games/common/player.py b/rl_games/common/player.py index bc5f3930..7f98c40d 100644 --- a/rl_games/common/player.py +++ b/rl_games/common/player.py @@ -2,7 +2,7 @@ import shutil import threading import time -import gym as gymnasium +import gymnasium as gym import numpy as np import torch import copy @@ -11,7 +11,7 @@ from rl_games.common import vecenv from rl_games.common import env_configurations from rl_games.algos_torch import model_builder - +from rl_games.common.env_configurations import patch_env_info class BasePlayer(object): @@ -32,11 +32,11 @@ def __init__(self, params): print('[BasePlayer] Creating vecenv: ', self.env_name) self.env = vecenv.create_vec_env( self.env_name, self.config['num_actors'], **self.env_config) - self.env_info = self.env.get_env_info() + self.env_info = patch_env_info(self.env.get_env_info()) else: print('[BasePlayer] Creating regular env: ', self.env_name) self.env = self.create_env() - self.env_info = env_configurations.get_env_info(self.env) + self.env_info = patch_env_info(env_configurations.get_env_info(self.env)) else: self.env = config.get('vec_env') diff --git a/rl_games/common/vecenv.py b/rl_games/common/vecenv.py index b07248f2..9f5586f1 100644 --- a/rl_games/common/vecenv.py +++ b/rl_games/common/vecenv.py @@ -2,7 +2,7 @@ from rl_games.common.env_configurations import configurations from rl_games.common.tr_helpers import dicts_to_dict_with_arrays import numpy as np -import gym as gymnasium +import gymnasium as gym import random from time import sleep import torch diff --git a/rl_games/common/wrappers.py b/rl_games/common/wrappers.py index d96b3971..2a9be40c 100644 --- a/rl_games/common/wrappers.py +++ b/rl_games/common/wrappers.py @@ -5,7 +5,7 @@ os.environ.setdefault('PATH', '') from collections import deque -import gym as gymnasium +import gymnasium as gym from gym import spaces from copy import copy diff --git a/rl_games/envs/brax.py b/rl_games/envs/brax.py index ab00d29a..fe7c51c6 100644 --- a/rl_games/envs/brax.py +++ b/rl_games/envs/brax.py @@ -1,5 +1,5 @@ from rl_games.common.ivecenv import IVecEnv -import gym as gymnasium +import gymnasium as gym import numpy as np import torch.utils.dlpack as tpack diff --git a/rl_games/envs/cule.py b/rl_games/envs/cule.py index 9452b853..fc066fb6 100644 --- a/rl_games/envs/cule.py +++ b/rl_games/envs/cule.py @@ -1,5 +1,5 @@ from rl_games.common.ivecenv import IVecEnv -import gym as gymnasium +import gymnasium as gym import torch import numpy as np diff --git a/rl_games/envs/diambra/diambra.py b/rl_games/envs/diambra/diambra.py index 5e0ee74b..94f38185 100644 --- a/rl_games/envs/diambra/diambra.py +++ b/rl_games/envs/diambra/diambra.py @@ -1,4 +1,4 @@ -import gym as gymnasium +import gymnasium as gym import numpy as np import os import random diff --git a/rl_games/envs/envpool.py b/rl_games/envs/envpool.py index dc61323e..67f45651 100644 --- a/rl_games/envs/envpool.py +++ b/rl_games/envs/envpool.py @@ -1,5 +1,5 @@ from rl_games.common.ivecenv import IVecEnv -import gym as gymnasium +import gymnasium as gym import numpy as np diff --git a/rl_games/envs/multiwalker.py b/rl_games/envs/multiwalker.py index 7e74f2e2..d3161939 100644 --- a/rl_games/envs/multiwalker.py +++ b/rl_games/envs/multiwalker.py @@ -1,4 +1,4 @@ -import gym as gymnasium +import gymnasium as gym import numpy as np from pettingzoo.sisl import multiwalker_v6 import yaml diff --git a/rl_games/envs/slimevolley_selfplay.py b/rl_games/envs/slimevolley_selfplay.py index daa79196..e61163c1 100644 --- a/rl_games/envs/slimevolley_selfplay.py +++ b/rl_games/envs/slimevolley_selfplay.py @@ -1,8 +1,9 @@ -import gym as gymnasium +import gymnasium as gym import numpy as np import slimevolleygym import yaml from rl_games.torch_runner import Runner +from rl_games.common.env_configurations import patch_env_info import os class SlimeVolleySelfplay(gym.Env): @@ -32,7 +33,7 @@ def create_agent(self, config='rl_games/configs/ma/ppo_slime_self_play.yaml'): config = yaml.safe_load(stream) runner = Runner() from rl_games.common.env_configurations import get_env_info - config['params']['config']['env_info'] = get_env_info(self) + config['params']['config']['env_info'] = patch_env_info(get_env_info(self)) runner.load(config) config = runner.get_prebuilt_config() diff --git a/rl_games/envs/smac_env.py b/rl_games/envs/smac_env.py index ad1686f3..4e90aca8 100644 --- a/rl_games/envs/smac_env.py +++ b/rl_games/envs/smac_env.py @@ -1,4 +1,4 @@ -import gym as gymnasium +import gymnasium as gym import numpy as np from smac.env import StarCraft2Env from smac.env import MultiAgentEnv diff --git a/rl_games/envs/smac_v2_env.py b/rl_games/envs/smac_v2_env.py index daed73e0..059013d3 100644 --- a/rl_games/envs/smac_v2_env.py +++ b/rl_games/envs/smac_v2_env.py @@ -1,4 +1,4 @@ -import gym as gymnasium +import gymnasium as gym import numpy as np import yaml from smacv2.env import StarCraft2Env diff --git a/rl_games/envs/test/__init__.py b/rl_games/envs/test/__init__.py index 9f140618..6ac308d6 100644 --- a/rl_games/envs/test/__init__.py +++ b/rl_games/envs/test/__init__.py @@ -1,4 +1,4 @@ -import gym as gymnasium +import gymnasium as gym gym.envs.register( id='TestRnnEnv-v0', diff --git a/rl_games/envs/test/example_env.py b/rl_games/envs/test/example_env.py index 4904bad9..c9f8023e 100644 --- a/rl_games/envs/test/example_env.py +++ b/rl_games/envs/test/example_env.py @@ -1,4 +1,4 @@ -import gym as gymnasium +import gymnasium as gym import numpy as np diff --git a/rl_games/envs/test/rnn_env.py b/rl_games/envs/test/rnn_env.py index 6924493d..7e0c427d 100644 --- a/rl_games/envs/test/rnn_env.py +++ b/rl_games/envs/test/rnn_env.py @@ -1,4 +1,4 @@ -import gym as gymnasium +import gymnasium as gym import numpy as np diff --git a/rl_games/envs/test/test_asymmetric_env.py b/rl_games/envs/test/test_asymmetric_env.py index 6c5ebd85..577ab27c 100644 --- a/rl_games/envs/test/test_asymmetric_env.py +++ b/rl_games/envs/test/test_asymmetric_env.py @@ -1,4 +1,4 @@ -import gym as gymnasium +import gymnasium as gym import numpy as np from rl_games.common.wrappers import MaskVelocityWrapper