# Copyright (c) 2019-2020, RTE (https://www.rte-france.com)
# See AUTHORS.txt
# This Source Code Form is subject to the terms of the Mozilla Public License, version 2.0.
# If a copy of the Mozilla Public License, version 2.0 was not distributed with this file,
# you can obtain one at http://mozilla.org/MPL/2.0/.
# SPDX-License-Identifier: MPL-2.0
# This file is part of Grid2Op, Grid2Op a testbed platform to model sequential decision making in power systems.
from abc import ABC, abstractmethod
from grid2op.Space import RandomObject
[docs]class BaseAgent(RandomObject, ABC):
This class represents the base class of an BaseAgent. All bot / controller / agent used in the Grid2Op simulator
should derived from this class.
To work properly, it is advise to create BaseAgent after the :class:`grid2op.Environment` has been created and reuse
the :attr:`grid2op.Environment.Environment.action_space` to build the BaseAgent.
It represent the action space ie a tool that can serve to create valid action. Note that a valid action can
be illegal or ambiguous, and so lead to a "game over" or to a error. But at least it will have a proper size.
def __init__(self, action_space):
self.action_space = copy.deepcopy(action_space)
[docs] def reset(self, obs):
This method is called at the beginning of a new episode.
It is implemented by agents to reset their internal state if needed.
The first observation corresponding to the initial state of the environment.
[docs] def seed(self, seed):
This function is used to guarantee that the "pseudo random numbers" generated and used by the agent instance
will be deterministic.
This guarantee, if the recommendation in :func:`BaseAgent.act` are followed that the agent will produce the same
set of actions if it faces the same observations in the same order. This is particularly important for
You can override this function with the method of your choosing, but if you do so, don't forget to call
The seed used
a tuple of seed used
return super().seed(seed), self.action_space.seed(seed)
def act(self, observation, reward, done=False):
This is the main method of an BaseAgent. Given the current observation and the current reward (ie the reward
that the environment send to the agent after the previous action has been implemented).
In order to be reproducible, and to make proper use of the
:func:`BaseAgent.seed` capabilities, you must absolutely NOT use the `random` python module (which will not
be seeded) nor the `np.random` module and avoid any other "sources" of pseudo random numbers.
You can adapt your code the following way. Instead of using `np.random` use `self.space_prng`.
For example, if you wanted to write
`np.random.randint(1,5)` replace it by `self.space_prng.randint(1,5)`. It is the same for `np.random.normal()`
replaced by `self.space_prng.normal()`.
You have an example of such usage in :func:`RandomAgent.my_act`.
If you really need other sources of randomness (for example if you use tensorflow or torch) we strongly
recommend you to overload the :func:`BaseAgent.seed` accordingly. In that
The current observation of the :class:`grid2op.Environment.Environment`
The current reward. This is the reward obtained by the previous action
Whether the episode has ended or not. Used to maintain gym compatibility
The action chosen by the bot / controler / agent.