Skip to content

Commit 9268934

Browse files
authored
[RLlib] LINT: Enable ruff imports for rllib/algorithms and rllib/utils (#56734)
1 parent 1d57f9a commit 9268934

File tree

98 files changed

+352
-328
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

98 files changed

+352
-328
lines changed

pyproject.toml

Lines changed: 0 additions & 16 deletions
Original file line numberDiff line numberDiff line change
@@ -65,22 +65,6 @@ afterray = ["psutil", "setproctitle"]
6565
"python/ray/__init__.py" = ["I"]
6666
"python/ray/dag/__init__.py" = ["I"]
6767
"python/ray/air/__init__.py" = ["I"]
68-
# "rllib/__init__.py" = ["I"]
69-
# "rllib/benchmarks/*" = ["I"]
70-
# "rllib/connectors/*" = ["I"]
71-
# "rllib/evaluation/*" = ["I"]
72-
# "rllib/models/*" = ["I"]
73-
"rllib/utils/*" = ["I"]
74-
"rllib/algorithms/*" = ["I"]
75-
# "rllib/core/*" = ["I"]
76-
# "rllib/examples/*" = ["I"]
77-
# "rllib/offline/*" = ["I"]
78-
# "rllib/tests/*" = ["I"]
79-
# "rllib/callbacks/*" = ["I"]
80-
# "rllib/env/*" = ["I"]
81-
# "rllib/execution/*" = ["I"]
82-
# "rllib/policy/*" = ["I"]
83-
# "rllib/tuned_examples/*" = ["I"]
8468
"release/*" = ["I"]
8569

8670
# TODO(matthewdeng): Remove this line

rllib/algorithms/__init__.py

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -6,15 +6,14 @@
66
from ray.rllib.algorithms.dqn.dqn import DQN, DQNConfig
77
from ray.rllib.algorithms.impala.impala import (
88
IMPALA,
9-
IMPALAConfig,
109
Impala,
10+
IMPALAConfig,
1111
ImpalaConfig,
1212
)
1313
from ray.rllib.algorithms.marwil.marwil import MARWIL, MARWILConfig
1414
from ray.rllib.algorithms.ppo.ppo import PPO, PPOConfig
1515
from ray.rllib.algorithms.sac.sac import SAC, SACConfig
1616

17-
1817
__all__ = [
1918
"Algorithm",
2019
"AlgorithmConfig",

rllib/algorithms/algorithm.py

Lines changed: 12 additions & 13 deletions
Original file line numberDiff line numberDiff line change
@@ -13,6 +13,7 @@
1313
from collections import defaultdict
1414
from datetime import datetime
1515
from typing import (
16+
TYPE_CHECKING,
1617
Any,
1718
Callable,
1819
Collection,
@@ -23,7 +24,6 @@
2324
Set,
2425
Tuple,
2526
Type,
26-
TYPE_CHECKING,
2727
Union,
2828
)
2929

@@ -47,9 +47,9 @@
4747
from ray.rllib.algorithms.utils import (
4848
AggregatorActor,
4949
_get_env_runner_bundles,
50-
_get_offline_eval_runner_bundles,
5150
_get_learner_bundles,
5251
_get_main_process_bundle,
52+
_get_offline_eval_runner_bundles,
5353
)
5454
from ray.rllib.callbacks.utils import make_callback
5555
from ray.rllib.connectors.agent.obs_preproc import ObsPreprocessorConnector
@@ -84,30 +84,30 @@
8484
from ray.rllib.execution.rollout_ops import synchronous_parallel_sample
8585
from ray.rllib.offline import get_dataset_and_shards
8686
from ray.rllib.offline.estimators import (
87-
OffPolicyEstimator,
88-
ImportanceSampling,
89-
WeightedImportanceSampling,
9087
DirectMethod,
9188
DoublyRobust,
89+
ImportanceSampling,
90+
OffPolicyEstimator,
91+
WeightedImportanceSampling,
9292
)
9393
from ray.rllib.offline.offline_evaluator import OfflineEvaluator
9494
from ray.rllib.policy.policy import Policy, PolicySpec
9595
from ray.rllib.policy.sample_batch import DEFAULT_POLICY_ID, SampleBatch
96-
from ray.rllib.utils import deep_update, FilterManager, force_list
96+
from ray.rllib.utils import FilterManager, deep_update, force_list
9797
from ray.rllib.utils.actor_manager import FaultTolerantActorManager
9898
from ray.rllib.utils.annotations import (
9999
DeveloperAPI,
100100
ExperimentalAPI,
101101
OldAPIStack,
102-
override,
103102
OverrideToImplementCustomLogic,
104103
OverrideToImplementCustomLogic_CallToSuperRecommended,
105104
PublicAPI,
105+
override,
106106
)
107107
from ray.rllib.utils.checkpoints import (
108-
Checkpointable,
109108
CHECKPOINT_VERSION,
110109
CHECKPOINT_VERSION_LEARNER_AND_ENV_RUNNER,
110+
Checkpointable,
111111
get_checkpoint_info,
112112
try_import_msgpack,
113113
)
@@ -134,9 +134,9 @@
134134
NUM_AGENT_STEPS_TRAINED,
135135
NUM_AGENT_STEPS_TRAINED_LIFETIME,
136136
NUM_ENV_STEPS_SAMPLED,
137+
NUM_ENV_STEPS_SAMPLED_FOR_EVALUATION_THIS_ITER,
137138
NUM_ENV_STEPS_SAMPLED_LIFETIME,
138139
NUM_ENV_STEPS_SAMPLED_THIS_ITER,
139-
NUM_ENV_STEPS_SAMPLED_FOR_EVALUATION_THIS_ITER,
140140
NUM_ENV_STEPS_TRAINED,
141141
NUM_ENV_STEPS_TRAINED_LIFETIME,
142142
NUM_EPISODES,
@@ -147,13 +147,13 @@
147147
RESTORE_ENV_RUNNERS_TIMER,
148148
RESTORE_EVAL_ENV_RUNNERS_TIMER,
149149
RESTORE_OFFLINE_EVAL_RUNNERS_TIMER,
150+
STEPS_TRAINED_THIS_ITER_COUNTER,
150151
SYNCH_ENV_CONNECTOR_STATES_TIMER,
151152
SYNCH_EVAL_ENV_CONNECTOR_STATES_TIMER,
152153
SYNCH_WORKER_WEIGHTS_TIMER,
153154
TIMERS,
154155
TRAINING_ITERATION_TIMER,
155156
TRAINING_STEP_TIMER,
156-
STEPS_TRAINED_THIS_ITER_COUNTER,
157157
)
158158
from ray.rllib.utils.metrics.learner_info import LEARNER_INFO
159159
from ray.rllib.utils.metrics.metrics_logger import MetricsLogger
@@ -164,7 +164,7 @@
164164
)
165165
from ray.rllib.utils.replay_buffers import MultiAgentReplayBuffer, ReplayBuffer
166166
from ray.rllib.utils.runners.runner_group import RunnerGroup
167-
from ray.rllib.utils.serialization import deserialize_type, NOT_SERIALIZABLE
167+
from ray.rllib.utils.serialization import NOT_SERIALIZABLE, deserialize_type
168168
from ray.rllib.utils.spaces import space_utils
169169
from ray.rllib.utils.typing import (
170170
AgentConnectorDataType,
@@ -191,8 +191,7 @@
191191
from ray.tune.execution.placement_groups import PlacementGroupFactory
192192
from ray.tune.experiment.trial import ExportFormat
193193
from ray.tune.logger import Logger, UnifiedLogger
194-
from ray.tune.registry import ENV_CREATOR, _global_registry
195-
from ray.tune.registry import get_trainable_cls
194+
from ray.tune.registry import ENV_CREATOR, _global_registry, get_trainable_cls
196195
from ray.tune.resources import Resources
197196
from ray.tune.result import TRAINING_ITERATION
198197
from ray.tune.trainable import Trainable

rllib/algorithms/algorithm_config.py

Lines changed: 9 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -1,10 +1,11 @@
11
import copy
22
import dataclasses
3-
from enum import Enum
43
import logging
54
import math
65
import sys
6+
from enum import Enum
77
from typing import (
8+
TYPE_CHECKING,
89
Any,
910
Callable,
1011
Collection,
@@ -13,16 +14,20 @@
1314
Optional,
1415
Tuple,
1516
Type,
16-
TYPE_CHECKING,
1717
Union,
1818
)
19-
from typing_extensions import Self
2019

2120
import gymnasium as gym
2221
import tree
2322
from packaging import version
23+
from typing_extensions import Self
2424

2525
import ray
26+
from ray._common.deprecation import (
27+
DEPRECATED_VALUE,
28+
Deprecated,
29+
deprecation_warning,
30+
)
2631
from ray.rllib.callbacks.callbacks import RLlibCallback
2732
from ray.rllib.connectors.connector_v2 import ConnectorV2
2833
from ray.rllib.core import DEFAULT_MODULE_ID
@@ -34,7 +39,7 @@
3439
from ray.rllib.core.rl_module.default_model_config import DefaultModelConfig
3540
from ray.rllib.core.rl_module.multi_rl_module import MultiRLModuleSpec
3641
from ray.rllib.core.rl_module.rl_module import RLModuleSpec
37-
from ray.rllib.env import INPUT_ENV_SPACES, INPUT_ENV_SINGLE_SPACES
42+
from ray.rllib.env import INPUT_ENV_SINGLE_SPACES, INPUT_ENV_SPACES
3843
from ray.rllib.env.multi_agent_env import MultiAgentEnv
3944
from ray.rllib.env.wrappers.atari_wrappers import is_atari
4045
from ray.rllib.evaluation.collectors.sample_collector import SampleCollector
@@ -49,11 +54,6 @@
4954
OldAPIStack,
5055
OverrideToImplementCustomLogic_CallToSuperRecommended,
5156
)
52-
from ray._common.deprecation import (
53-
DEPRECATED_VALUE,
54-
Deprecated,
55-
deprecation_warning,
56-
)
5757
from ray.rllib.utils.framework import try_import_tf, try_import_torch
5858
from ray.rllib.utils.from_config import NotProvided, from_config
5959
from ray.rllib.utils.schedules.scheduler import Scheduler
@@ -84,7 +84,6 @@
8484
from ray.util import log_once
8585
from ray.util.placement_group import PlacementGroup
8686

87-
8887
if TYPE_CHECKING:
8988
from ray.rllib.algorithms.algorithm import Algorithm
9089
from ray.rllib.core.learner import Learner

rllib/algorithms/appo/appo.py

Lines changed: 4 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -10,23 +10,24 @@
1010
https://arxiv.org/pdf/1912.00167
1111
"""
1212

13+
import logging
1314
from typing import Optional, Type
15+
1416
from typing_extensions import Self
15-
import logging
1617

18+
from ray._common.deprecation import DEPRECATED_VALUE, deprecation_warning
1719
from ray.rllib.algorithms.algorithm_config import AlgorithmConfig, NotProvided
1820
from ray.rllib.algorithms.impala.impala import IMPALA, IMPALAConfig
1921
from ray.rllib.core.rl_module.rl_module import RLModuleSpec
2022
from ray.rllib.policy.policy import Policy
2123
from ray.rllib.utils.annotations import override
22-
from ray._common.deprecation import DEPRECATED_VALUE, deprecation_warning
2324
from ray.rllib.utils.metrics import (
2425
LAST_TARGET_UPDATE_TS,
26+
LEARNER_STATS_KEY,
2527
NUM_AGENT_STEPS_SAMPLED,
2628
NUM_ENV_STEPS_SAMPLED,
2729
NUM_TARGET_UPDATES,
2830
)
29-
from ray.rllib.utils.metrics import LEARNER_STATS_KEY
3031

3132
logger = logging.getLogger(__name__)
3233

rllib/algorithms/appo/appo_tf_policy.py

Lines changed: 11 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -5,37 +5,37 @@
55
Keep in sync with changes to VTraceTFPolicy.
66
"""
77

8-
import numpy as np
98
import logging
10-
import gymnasium as gym
119
from typing import Dict, List, Optional, Type, Union
1210

11+
import gymnasium as gym
12+
import numpy as np
13+
1314
from ray.rllib.algorithms.appo.utils import make_appo_models
1415
from ray.rllib.algorithms.impala import vtrace_tf as vtrace
1516
from ray.rllib.algorithms.impala.impala_tf_policy import (
16-
_make_time_major,
1717
VTraceClipGradients,
1818
VTraceOptimizer,
19+
_make_time_major,
1920
)
2021
from ray.rllib.evaluation.postprocessing import (
22+
Postprocessing,
2123
compute_bootstrap_value,
2224
compute_gae_for_sample_batch,
23-
Postprocessing,
2425
)
25-
from ray.rllib.models.tf.tf_action_dist import Categorical
26-
from ray.rllib.policy.sample_batch import SampleBatch
26+
from ray.rllib.models.modelv2 import ModelV2
27+
from ray.rllib.models.tf.tf_action_dist import Categorical, TFActionDistribution
2728
from ray.rllib.policy.dynamic_tf_policy_v2 import DynamicTFPolicyV2
2829
from ray.rllib.policy.eager_tf_policy_v2 import EagerTFPolicyV2
30+
from ray.rllib.policy.sample_batch import SampleBatch
2931
from ray.rllib.policy.tf_mixins import (
3032
EntropyCoeffSchedule,
31-
LearningRateSchedule,
32-
KLCoeffMixin,
33-
ValueNetworkMixin,
3433
GradStatsMixin,
34+
KLCoeffMixin,
35+
LearningRateSchedule,
3536
TargetNetworkMixin,
37+
ValueNetworkMixin,
3638
)
37-
from ray.rllib.models.modelv2 import ModelV2
38-
from ray.rllib.models.tf.tf_action_dist import TFActionDistribution
3939
from ray.rllib.utils.annotations import (
4040
override,
4141
)

rllib/algorithms/appo/appo_torch_policy.py

Lines changed: 9 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -5,37 +5,38 @@
55
Keep in sync with changes to VTraceTFPolicy.
66
"""
77

8-
import gymnasium as gym
9-
import numpy as np
108
import logging
119
from typing import Any, Dict, List, Optional, Type, Union
1210

11+
import gymnasium as gym
12+
import numpy as np
13+
1314
import ray
14-
from ray.rllib.algorithms.appo.utils import make_appo_models
1515
import ray.rllib.algorithms.impala.vtrace_torch as vtrace
16+
from ray.rllib.algorithms.appo.utils import make_appo_models
1617
from ray.rllib.algorithms.impala.impala_torch_policy import (
17-
make_time_major,
1818
VTraceOptimizer,
19+
make_time_major,
1920
)
2021
from ray.rllib.evaluation.postprocessing import (
22+
Postprocessing,
2123
compute_bootstrap_value,
2224
compute_gae_for_sample_batch,
23-
Postprocessing,
2425
)
2526
from ray.rllib.models.action_dist import ActionDistribution
2627
from ray.rllib.models.modelv2 import ModelV2
2728
from ray.rllib.models.torch.torch_action_dist import (
28-
TorchDistributionWrapper,
2929
TorchCategorical,
30+
TorchDistributionWrapper,
3031
)
3132
from ray.rllib.models.torch.torch_modelv2 import TorchModelV2
3233
from ray.rllib.policy.sample_batch import SampleBatch
3334
from ray.rllib.policy.torch_mixins import (
3435
EntropyCoeffSchedule,
35-
LearningRateSchedule,
3636
KLCoeffMixin,
37-
ValueNetworkMixin,
37+
LearningRateSchedule,
3838
TargetNetworkMixin,
39+
ValueNetworkMixin,
3940
)
4041
from ray.rllib.policy.torch_policy_v2 import TorchPolicyV2
4142
from ray.rllib.utils.annotations import override

rllib/algorithms/appo/default_appo_rl_module.py

Lines changed: 2 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -8,12 +8,11 @@
88
TARGET_NETWORK_ACTION_DIST_INPUTS,
99
TargetNetworkAPI,
1010
)
11-
from ray.rllib.utils.typing import NetworkType
12-
1311
from ray.rllib.utils.annotations import (
14-
override,
1512
OverrideToImplementCustomLogic_CallToSuperRecommended,
13+
override,
1614
)
15+
from ray.rllib.utils.typing import NetworkType
1716
from ray.util.annotations import DeveloperAPI
1817

1918

rllib/algorithms/appo/tests/test_appo.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -11,9 +11,9 @@
1111
NUM_ENV_STEPS_SAMPLED_LIFETIME,
1212
)
1313
from ray.rllib.utils.test_utils import (
14+
check_compute_single_action,
1415
check_train_results,
1516
check_train_results_new_api_stack,
16-
check_compute_single_action,
1717
)
1818

1919

rllib/algorithms/appo/tests/test_appo_learner.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
import unittest
2-
import numpy as np
32

3+
import numpy as np
44
import tree # pip install dm_tree
55

66
import ray
@@ -13,7 +13,6 @@
1313
from ray.rllib.utils.metrics import LEARNER_RESULTS
1414
from ray.rllib.utils.torch_utils import convert_to_torch_tensor
1515

16-
1716
frag_length = 50
1817

1918
FAKE_BATCH = {
@@ -119,7 +118,8 @@ def test_kl_coeff_changes(self):
119118

120119

121120
if __name__ == "__main__":
122-
import pytest
123121
import sys
124122

123+
import pytest
124+
125125
sys.exit(pytest.main(["-v", __file__]))

0 commit comments

Comments
 (0)