# Copyright (c) 2019-2020, RTE (https://www.rte-france.com)
# See AUTHORS.txt
# This Source Code Form is subject to the terms of the Mozilla Public License, version 2.0.
# If a copy of the Mozilla Public License, version 2.0 was not distributed with this file,
# you can obtain one at http://mozilla.org/MPL/2.0/.
# SPDX-License-Identifier: MPL-2.0
# This file is part of Grid2Op, Grid2Op a testbed platform to model sequential decision making in power systems.
from grid2op.Reward.baseReward import BaseReward
from grid2op.dtypes import dt_float
[docs]class GameplayReward(BaseReward):
"""
This rewards is strictly computed based on the Game status.
It yields a negative reward in case of game over.
A half negative reward on rules infringement.
Otherwise the reward is positive.
Examples
---------
You can use this reward in any environment with:
.. code-block:: python
import grid2op
from grid2op.Reward import GameplayReward
# then you create your environment with it:
NAME_OF_THE_ENVIRONMENT = "l2rpn_case14_sandbox"
env = grid2op.make(NAME_OF_THE_ENVIRONMENT,reward_class=GameplayReward)
# and do a step with a "do nothing" action
obs = env.reset()
obs, reward, done, info = env.step(env.action_space())
# the reward is computed with the GameplayReward class
"""
[docs] def __init__(self, logger=None):
BaseReward.__init__(self, logger=logger)
self.reward_min = dt_float(-1.0)
self.reward_max = dt_float(1.0)
[docs] def __call__(self, action, env, has_error, is_done, is_illegal, is_ambiguous):
if has_error:
return self.reward_min
elif is_illegal or is_ambiguous:
# Did not respect the rules
return self.reward_min / dt_float(2.0)
else:
# Keep playing or finished episode
return self.reward_max