Skip to content

Commit 9c01f52

Browse files
committed
tweaked difficulty of Plane
1 parent 070d515 commit 9c01f52

File tree

1 file changed

+3
-5
lines changed

1 file changed

+3
-5
lines changed

src/target_gym/plane/env.py

Lines changed: 3 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,5 @@
11
from typing import Tuple
22

3-
# Optional: jax imports (only used in jax env)
4-
import chex
53
import jax
64
import jax.numpy as jnp
75
from flax import struct
@@ -104,8 +102,8 @@ class PlaneParams(EnvParams):
104102
max_steps_in_episode: int = 10_000
105103
min_alt: float = 0.0
106104
max_alt: float = 40_000.0 / 3.281
107-
target_altitude_range: Tuple[float, float] = (5_000.0, 5_000.0)
108-
initial_altitude_range: Tuple[float, float] = (3_000.0, 5_000.0)
105+
target_altitude_range: Tuple[float, float] = (3_000.0, 8_000.0)
106+
initial_altitude_range: Tuple[float, float] = (3_000.0, 8_000.0)
109107
initial_z_dot: float = 0.0
110108
initial_x_dot: float = 200.0
111109
initial_theta_dot: float = 0.0
@@ -159,7 +157,7 @@ def compute_reward(state: PlaneState, params: PlaneParams, xp=jnp):
159157
reward = xp.where(
160158
done_alt,
161159
-1.0 * params.max_steps_in_episode,
162-
((max_alt_diff - xp.abs(state.target_altitude - state.z)) / max_alt_diff) ** 2,
160+
((max_alt_diff - xp.abs(state.target_altitude - state.z)) / max_alt_diff) ** 10,
163161
)
164162
return reward
165163

0 commit comments

Comments
 (0)