Skip to content

Commit aec17f1

Browse files
[pre-commit.ci] auto fixes from pre-commit.com hooks
for more information, see https://pre-commit.ci
1 parent 61a8423 commit aec17f1

File tree

5 files changed

+12
-17
lines changed

5 files changed

+12
-17
lines changed

torch_points3d/core/instantiator.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -43,7 +43,7 @@ def litmodel(self, cfg: DictConfig) -> "PointCloudBaseModule":
4343

4444
def model(self, cfg: DictConfig) -> "PointCloudBaseModel":
4545
return self.instantiate(cfg, self)
46-
46+
4747
def backbone(self, cfg: DictConfig):
4848
return self.instantiate(cfg)
4949

Lines changed: 3 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,3 @@
1-
from typing import Any, Dict, Optional, Tuple, Union
2-
31
import torch.nn as nn
42

53
from torch_points3d.core.instantiator import Instantiator
@@ -12,7 +10,7 @@ def __init__(self, instantiator: Instantiator):
1210
self.instantiator = instantiator
1311

1412
def set_input(self, data):
15-
raise(NotImplementedError("set_input needs to be defined!"))
16-
13+
raise (NotImplementedError("set_input needs to be defined!"))
14+
1715
def forward(self):
18-
raise(NotImplementedError("forward needs to be defined!"))
16+
raise (NotImplementedError("forward needs to be defined!"))

torch_points3d/models/segmentation/base_model.py

Lines changed: 4 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,3 @@
1-
from typing import Any, Dict, Optional, Tuple, Union
21
from omegaconf import DictConfig
32

43
import torch.nn as nn
@@ -11,10 +10,10 @@
1110
class SegmentationBaseModel(PointCloudBaseModel):
1211
def __init__(self, instantiator: Instantiator, num_classes: int, backbone: DictConfig):
1312
super().__init__(instantiator)
14-
13+
1514
print(backbone)
16-
self.backbone = self.instantiator.backbone(backbone)
17-
15+
self.backbone = self.instantiator.backbone(backbone)
16+
1817
self.head = nn.Sequential(nn.Linear(self.backbone.output_nc, num_classes))
1918

2019
def set_input(self, data):
@@ -24,7 +23,7 @@ def set_input(self, data):
2423
self.labels = data.y
2524
else:
2625
self.labels = None
27-
26+
2827
def forward(self):
2928
features = self.backbone(self.input).x
3029
logits = self.head(features)

torch_points3d/tasks/base_model.py

Lines changed: 2 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -1,11 +1,9 @@
11
from typing import Any, Dict, Optional, Tuple, Union
22

33
import pytorch_lightning as pl
4-
import torch
54
from pytorch_lightning.utilities import rank_zero_info
65
from pytorch_lightning.utilities.exceptions import MisconfigurationException
76
from omegaconf import DictConfig
8-
import torch.nn.functional as F
97

108
from torch_points3d.core.instantiator import Instantiator
119
from torch_points3d.core.config import OptimizerConfig, SchedulerConfig
@@ -17,7 +15,7 @@ def __init__(
1715
model: DictConfig,
1816
optimizer: OptimizerConfig,
1917
instantiator: Instantiator,
20-
scheduler: SchedulerConfig = None, # scheduler shouldn't be required
18+
scheduler: SchedulerConfig = None, # scheduler shouldn't be required
2119
):
2220
super().__init__()
2321
# some optimizers/schedulers need parameters only known dynamically
@@ -31,7 +29,7 @@ def __init__(
3129
def _init_model(self, model_cfg):
3230
print(model_cfg)
3331
self.model = self.instantiator.model(model_cfg)
34-
32+
3533
def configure_optimizers(self) -> Dict:
3634
"""Prepare optimizer and scheduler"""
3735
optims = {}

torch_points3d/trainer.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -30,12 +30,12 @@ def __init__(
3030
"The instantiator did not return a DataModule instance." " Hydra hint: is `dataset._target_` defined?`"
3131
)
3232
self.data_module.setup("fit")
33-
33+
3434
self.litmodel: PointCloudBaseModule = instantiator.litmodel(model)
3535
print(self.litmodel)
3636
self.trainer = instantiator.trainer(
3737
trainer,
38-
logger=None, # eventually add logger config back in
38+
logger=None, # eventually add logger config back in
3939
)
4040

4141
def train(self):

0 commit comments

Comments
 (0)