From e4688ae754faeb2b93cf5664bedbae0ff3ea9b62 Mon Sep 17 00:00:00 2001 From: Arnaud Gelas Date: Fri, 15 Jan 2021 19:57:40 +0100 Subject: [PATCH] Fix isort failures in utilities (#5530) Remove from skipped module in pyproject.toml and fix failures on: - pytorch_lightning/utilities/*.py --- pyproject.toml | 4 --- pytorch_lightning/utilities/__init__.py | 28 ++++++++----------- pytorch_lightning/utilities/argparse.py | 3 +- pytorch_lightning/utilities/data.py | 3 +- pytorch_lightning/utilities/debugging.py | 2 +- .../utilities/device_dtype_mixin.py | 2 +- pytorch_lightning/utilities/device_parser.py | 3 +- pytorch_lightning/utilities/model_helpers.py | 2 +- pytorch_lightning/utilities/parsing.py | 2 +- 9 files changed, 21 insertions(+), 28 deletions(-) diff --git a/pyproject.toml b/pyproject.toml index a5a521960d..2715976c8f 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -65,10 +65,6 @@ skip_glob = [ # todo "pytorch_lightning/tuner/*", - - - # todo - "pytorch_lightning/utilities/*", ] profile = "black" line_length = 120 diff --git a/pytorch_lightning/utilities/__init__.py b/pytorch_lightning/utilities/__init__.py index adcf996c66..0a5ed04eb7 100644 --- a/pytorch_lightning/utilities/__init__.py +++ b/pytorch_lightning/utilities/__init__.py @@ -22,31 +22,25 @@ from pytorch_lightning.utilities.distributed import ( # noqa: F401 rank_zero_only, rank_zero_warn, ) -from pytorch_lightning.utilities.enums import ( # noqa: F401 - LightningEnum, - AMPType, - DistributedType, - DeviceType, -) +from pytorch_lightning.utilities.enums import AMPType, DeviceType, DistributedType, LightningEnum # noqa: F401 from pytorch_lightning.utilities.imports import ( # noqa: F401 _APEX_AVAILABLE, - _NATIVE_AMP_AVAILABLE, - _XLA_AVAILABLE, - _OMEGACONF_AVAILABLE, - _HYDRA_AVAILABLE, - _HOROVOD_AVAILABLE, - _TORCHTEXT_AVAILABLE, - _FAIRSCALE_AVAILABLE, - _RPC_AVAILABLE, - _GROUP_AVAILABLE, - _FAIRSCALE_PIPE_AVAILABLE, _BOLTS_AVAILABLE, + _FAIRSCALE_AVAILABLE, + _FAIRSCALE_PIPE_AVAILABLE, + _GROUP_AVAILABLE, + _HOROVOD_AVAILABLE, + _HYDRA_AVAILABLE, _module_available, + _NATIVE_AMP_AVAILABLE, + _OMEGACONF_AVAILABLE, + _RPC_AVAILABLE, + _TORCHTEXT_AVAILABLE, + _XLA_AVAILABLE, ) from pytorch_lightning.utilities.parsing import AttributeDict, flatten_dict, is_picklable # noqa: F401 from pytorch_lightning.utilities.xla_device import XLADeviceUtils # noqa: F401 - _TPU_AVAILABLE = XLADeviceUtils.tpu_device_exists() FLOAT16_EPSILON = numpy.finfo(numpy.float16).eps diff --git a/pytorch_lightning/utilities/argparse.py b/pytorch_lightning/utilities/argparse.py index 70d36e9dcc..36a0739f1d 100644 --- a/pytorch_lightning/utilities/argparse.py +++ b/pytorch_lightning/utilities/argparse.py @@ -15,7 +15,8 @@ import inspect import os from argparse import ArgumentParser, Namespace from contextlib import suppress -from typing import Dict, Union, List, Tuple, Any +from typing import Any, Dict, List, Tuple, Union + from pytorch_lightning.utilities import parsing diff --git a/pytorch_lightning/utilities/data.py b/pytorch_lightning/utilities/data.py index 1b4907ab8c..aee260e627 100644 --- a/pytorch_lightning/utilities/data.py +++ b/pytorch_lightning/utilities/data.py @@ -13,11 +13,12 @@ # limitations under the License. from distutils.version import LooseVersion +from typing import Union + import torch from torch.utils.data import DataLoader, IterableDataset from pytorch_lightning.utilities import rank_zero_warn -from typing import Union def has_iterable_dataset(dataloader: DataLoader): diff --git a/pytorch_lightning/utilities/debugging.py b/pytorch_lightning/utilities/debugging.py index ffccda3e6f..c80f68213d 100644 --- a/pytorch_lightning/utilities/debugging.py +++ b/pytorch_lightning/utilities/debugging.py @@ -16,7 +16,7 @@ import os import time from collections import Counter from functools import wraps -from typing import Callable, Any, Optional +from typing import Any, Callable, Optional def enabled_only(fn: Callable): diff --git a/pytorch_lightning/utilities/device_dtype_mixin.py b/pytorch_lightning/utilities/device_dtype_mixin.py index cf8e83648b..79182cd086 100644 --- a/pytorch_lightning/utilities/device_dtype_mixin.py +++ b/pytorch_lightning/utilities/device_dtype_mixin.py @@ -12,7 +12,7 @@ # See the License for the specific language governing permissions and # limitations under the License. -from typing import Union, Optional +from typing import Optional, Union import torch from torch.nn import Module diff --git a/pytorch_lightning/utilities/device_parser.py b/pytorch_lightning/utilities/device_parser.py index b28369a69f..650a5cf600 100644 --- a/pytorch_lightning/utilities/device_parser.py +++ b/pytorch_lightning/utilities/device_parser.py @@ -11,8 +11,9 @@ # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. +from typing import Any, List, MutableSequence, Optional, Union + import torch -from typing import Union, Any, List, Optional, MutableSequence from pytorch_lightning.utilities import _TPU_AVAILABLE from pytorch_lightning.utilities.exceptions import MisconfigurationException diff --git a/pytorch_lightning/utilities/model_helpers.py b/pytorch_lightning/utilities/model_helpers.py index 993d9e11e1..87bd9e6c45 100644 --- a/pytorch_lightning/utilities/model_helpers.py +++ b/pytorch_lightning/utilities/model_helpers.py @@ -14,8 +14,8 @@ from typing import Union -from pytorch_lightning.core.lightning import LightningModule from pytorch_lightning.core.datamodule import LightningDataModule +from pytorch_lightning.core.lightning import LightningModule def is_overridden(method_name: str, model: Union[LightningModule, LightningDataModule]) -> bool: diff --git a/pytorch_lightning/utilities/parsing.py b/pytorch_lightning/utilities/parsing.py index b207320c25..521dd52005 100644 --- a/pytorch_lightning/utilities/parsing.py +++ b/pytorch_lightning/utilities/parsing.py @@ -15,7 +15,7 @@ import inspect import pickle from argparse import Namespace -from typing import Dict, Union, Tuple +from typing import Dict, Tuple, Union from pytorch_lightning.utilities import rank_zero_warn