diff --git a/reinforcement_learning/__init__.py b/reinforcement_learning/__init__.py new file mode 100644 index 00000000..e69de29b diff --git a/reinforcement_learning/stat_wrapper.py b/reinforcement_learning/stat_wrapper.py index 0a8f09fd..febcf865 100644 --- a/reinforcement_learning/stat_wrapper.py +++ b/reinforcement_learning/stat_wrapper.py @@ -25,7 +25,6 @@ def observation(self, agent_id, agent_obs): """Called before observations are returned from the environment Use this to define custom featurizers. Changing the space itself requires you to define the observation space again (i.e. Gym.spaces.Dict(gym.spaces....))""" - # NOTE: nmmo minigame obs is mapping proxy, which doesn't work with pufferlib flatten return agent_obs def action(self, agent_id, agent_atn):