Source code for lagom.envs.wrappers.clip_reward

import numpy as np

from gym import RewardWrapper


[docs]class ClipReward(RewardWrapper): r""""Clip reward to [min, max]. """ def __init__(self, env, min_r, max_r): super().__init__(env) self.min_r = min_r self.max_r = max_r def reward(self, reward): return np.clip(reward, self.min_r, self.max_r)