2020-07-25 18:38:51 +00:00
|
|
|
# Copyright The PyTorch Lightning team.
|
|
|
|
#
|
|
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
# you may not use this file except in compliance with the License.
|
|
|
|
# You may obtain a copy of the License at
|
|
|
|
#
|
|
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
#
|
|
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
# See the License for the specific language governing permissions and
|
|
|
|
# limitations under the License.
|
|
|
|
|
2020-11-13 15:05:54 +00:00
|
|
|
"""Trainer to automate the training."""
|
|
|
|
|
2019-07-09 00:11:20 +00:00
|
|
|
import os
|
2020-07-24 15:42:15 +00:00
|
|
|
import warnings
|
2020-09-12 03:33:09 +00:00
|
|
|
from typing import Dict, Iterable, List, Optional, Union
|
2019-07-09 00:11:20 +00:00
|
|
|
|
2019-03-31 01:45:16 +00:00
|
|
|
import torch
|
2020-02-23 02:23:30 +00:00
|
|
|
from torch.utils.data import DataLoader
|
2019-07-09 00:11:20 +00:00
|
|
|
|
2020-10-19 20:20:17 +00:00
|
|
|
from pytorch_lightning.callbacks import Callback, ModelCheckpoint
|
2020-07-24 15:42:15 +00:00
|
|
|
from pytorch_lightning.core.datamodule import LightningDataModule
|
2020-03-24 18:55:27 +00:00
|
|
|
from pytorch_lightning.core.lightning import LightningModule
|
2020-11-02 20:51:43 +00:00
|
|
|
from pytorch_lightning.core.memory import ModelSummary
|
|
|
|
from pytorch_lightning.core.step_result import Result, EvalResult
|
2020-02-23 02:23:30 +00:00
|
|
|
from pytorch_lightning.loggers import LightningLoggerBase
|
2020-09-12 03:33:09 +00:00
|
|
|
from pytorch_lightning.profiler import BaseProfiler
|
2020-03-06 17:00:05 +00:00
|
|
|
from pytorch_lightning.trainer.callback_hook import TrainerCallbackHookMixin
|
2020-08-07 22:33:51 +00:00
|
|
|
from pytorch_lightning.trainer.configuration_validator import ConfigValidator
|
2020-10-09 23:34:09 +00:00
|
|
|
from pytorch_lightning.trainer.connectors.env_vars_connector import overwrite_by_env_vars
|
2020-03-24 18:55:27 +00:00
|
|
|
from pytorch_lightning.trainer.data_loading import TrainerDataLoadingMixin
|
2019-12-04 16:39:14 +00:00
|
|
|
from pytorch_lightning.trainer.logging import TrainerLoggingMixin
|
|
|
|
from pytorch_lightning.trainer.model_hooks import TrainerModelHooksMixin
|
2020-04-02 15:48:53 +00:00
|
|
|
from pytorch_lightning.trainer.optimizers import TrainerOptimizersMixin
|
2020-08-09 10:24:09 +00:00
|
|
|
from pytorch_lightning.trainer.states import TrainerState, trainer_state
|
2019-12-04 16:39:14 +00:00
|
|
|
from pytorch_lightning.trainer.training_tricks import TrainerTrainingTricksMixin
|
2020-09-12 03:33:09 +00:00
|
|
|
from pytorch_lightning.utilities import rank_zero_warn
|
2020-07-20 23:00:20 +00:00
|
|
|
from pytorch_lightning.utilities.debugging import InternalDebugger
|
2020-07-24 15:42:15 +00:00
|
|
|
from pytorch_lightning.utilities.exceptions import MisconfigurationException
|
2020-09-09 12:45:04 +00:00
|
|
|
from pytorch_lightning.trainer.evaluation_loop import EvaluationLoop
|
|
|
|
from pytorch_lightning.trainer.training_loop import TrainLoop
|
2020-09-01 19:48:28 +00:00
|
|
|
from pytorch_lightning.accelerators.accelerator_connector import AcceleratorConnector
|
2020-09-12 03:33:09 +00:00
|
|
|
from pytorch_lightning.trainer.connectors.logger_connector import LoggerConnector
|
|
|
|
from pytorch_lightning.trainer.connectors.optimizer_connector import OptimizerConnector
|
|
|
|
from pytorch_lightning.trainer.connectors.training_trick_connector import TrainingTricksConnector
|
|
|
|
from pytorch_lightning.trainer.connectors.callback_connector import CallbackConnector
|
|
|
|
from pytorch_lightning.trainer.connectors.model_connector import ModelConnector
|
|
|
|
from pytorch_lightning.trainer.connectors.debugging_connector import DebuggingConnector
|
2020-09-12 11:05:21 +00:00
|
|
|
from pytorch_lightning.trainer.connectors.checkpoint_connector import CheckpointConnector
|
2020-09-12 15:07:15 +00:00
|
|
|
from pytorch_lightning.trainer.connectors.slurm_connector import SLURMConnector
|
2020-09-06 21:50:47 +00:00
|
|
|
from pytorch_lightning import _logger as log
|
2020-09-08 22:46:42 +00:00
|
|
|
from pytorch_lightning.tuner.tuning import Tuner
|
2020-09-12 03:33:09 +00:00
|
|
|
from pytorch_lightning.trainer.connectors.precision_connector import PrecisionConnector
|
|
|
|
from pytorch_lightning.trainer.connectors.profiler_connector import ProfilerConnector
|
|
|
|
from pytorch_lightning.trainer.connectors.data_connector import DataConnector
|
2020-09-28 11:53:57 +00:00
|
|
|
from pytorch_lightning.utilities.cloud_io import load as pl_load
|
2020-08-31 16:12:02 +00:00
|
|
|
from pytorch_lightning.utilities.model_utils import is_overridden
|
2020-09-10 00:03:18 +00:00
|
|
|
from pytorch_lightning.trainer.properties import TrainerProperties
|
2020-10-10 02:03:46 +00:00
|
|
|
from pytorch_lightning.plugins.plugin_connector import PluginConnector
|
2020-10-11 05:05:14 +00:00
|
|
|
from pytorch_lightning.accelerators.accelerator import Accelerator
|
|
|
|
from pytorch_lightning.accelerators.cpu_accelerator import CPUAccelerator
|
2020-11-10 21:13:41 +00:00
|
|
|
from pytorch_lightning.utilities.memory import recursive_detach
|
2020-06-30 22:09:16 +00:00
|
|
|
|
2020-07-09 15:36:21 +00:00
|
|
|
# warnings to ignore in trainer
|
2020-07-24 15:42:15 +00:00
|
|
|
warnings.filterwarnings(
|
|
|
|
'ignore', message='torch.distributed.reduce_op is deprecated, ' 'please use torch.distributed.ReduceOp instead'
|
|
|
|
)
|
2019-10-04 19:35:02 +00:00
|
|
|
|
2019-05-14 00:40:07 +00:00
|
|
|
try:
|
|
|
|
from apex import amp
|
2019-08-05 21:28:04 +00:00
|
|
|
except ImportError:
|
2020-08-08 09:07:32 +00:00
|
|
|
amp = None
|
2019-03-31 01:45:16 +00:00
|
|
|
|
2019-07-09 00:12:27 +00:00
|
|
|
|
2020-03-06 17:00:05 +00:00
|
|
|
class Trainer(
|
2020-09-10 00:03:18 +00:00
|
|
|
TrainerProperties,
|
2020-06-19 15:00:46 +00:00
|
|
|
TrainerCallbackHookMixin,
|
|
|
|
TrainerModelHooksMixin,
|
2020-04-02 15:48:53 +00:00
|
|
|
TrainerOptimizersMixin,
|
2020-03-06 17:00:05 +00:00
|
|
|
TrainerLoggingMixin,
|
|
|
|
TrainerTrainingTricksMixin,
|
|
|
|
TrainerDataLoadingMixin,
|
|
|
|
):
|
2020-10-09 23:34:09 +00:00
|
|
|
@overwrite_by_env_vars
|
2019-12-04 11:57:10 +00:00
|
|
|
def __init__(
|
2020-06-12 18:37:52 +00:00
|
|
|
self,
|
|
|
|
logger: Union[LightningLoggerBase, Iterable[LightningLoggerBase], bool] = True,
|
2020-10-30 03:47:37 +00:00
|
|
|
checkpoint_callback: bool = True,
|
2020-06-12 18:37:52 +00:00
|
|
|
callbacks: Optional[List[Callback]] = None,
|
|
|
|
default_root_dir: Optional[str] = None,
|
|
|
|
gradient_clip_val: float = 0,
|
|
|
|
process_position: int = 0,
|
|
|
|
num_nodes: int = 1,
|
|
|
|
num_processes: int = 1,
|
|
|
|
gpus: Optional[Union[List[int], str, int]] = None,
|
|
|
|
auto_select_gpus: bool = False,
|
2020-06-23 16:06:57 +00:00
|
|
|
tpu_cores: Optional[Union[List[int], str, int]] = None,
|
2020-06-12 18:37:52 +00:00
|
|
|
log_gpu_memory: Optional[str] = None,
|
|
|
|
progress_bar_refresh_rate: int = 1,
|
2020-06-17 12:03:28 +00:00
|
|
|
overfit_batches: Union[int, float] = 0.0,
|
2020-06-12 18:37:52 +00:00
|
|
|
track_grad_norm: Union[int, float, str] = -1,
|
|
|
|
check_val_every_n_epoch: int = 1,
|
|
|
|
fast_dev_run: bool = False,
|
|
|
|
accumulate_grad_batches: Union[int, Dict[int, int], List[list]] = 1,
|
|
|
|
max_epochs: int = 1000,
|
|
|
|
min_epochs: int = 1,
|
|
|
|
max_steps: Optional[int] = None,
|
|
|
|
min_steps: Optional[int] = None,
|
2020-06-17 17:42:28 +00:00
|
|
|
limit_train_batches: Union[int, float] = 1.0,
|
2020-06-17 12:03:28 +00:00
|
|
|
limit_val_batches: Union[int, float] = 1.0,
|
|
|
|
limit_test_batches: Union[int, float] = 1.0,
|
2020-06-17 17:42:28 +00:00
|
|
|
val_check_interval: Union[int, float] = 1.0,
|
2020-10-06 14:27:06 +00:00
|
|
|
flush_logs_every_n_steps: int = 100,
|
|
|
|
log_every_n_steps: int = 50,
|
2020-10-10 13:21:08 +00:00
|
|
|
accelerator: Optional[Union[str, Accelerator]] = None,
|
2020-08-05 23:12:11 +00:00
|
|
|
sync_batchnorm: bool = False,
|
2020-06-12 18:37:52 +00:00
|
|
|
precision: int = 32,
|
2020-10-10 02:03:46 +00:00
|
|
|
weights_summary: Optional[str] = 'top',
|
2020-06-12 18:37:52 +00:00
|
|
|
weights_save_path: Optional[str] = None,
|
|
|
|
num_sanity_val_steps: int = 2,
|
|
|
|
truncated_bptt_steps: Optional[int] = None,
|
|
|
|
resume_from_checkpoint: Optional[str] = None,
|
2020-10-27 10:57:16 +00:00
|
|
|
profiler: Optional[Union[BaseProfiler, bool, str]] = None,
|
2020-06-12 18:37:52 +00:00
|
|
|
benchmark: bool = False,
|
|
|
|
deterministic: bool = False,
|
|
|
|
reload_dataloaders_every_epoch: bool = False,
|
|
|
|
auto_lr_find: Union[bool, str] = False,
|
|
|
|
replace_sampler_ddp: bool = True,
|
|
|
|
terminate_on_nan: bool = False,
|
|
|
|
auto_scale_batch_size: Union[str, bool] = False,
|
2020-06-13 16:00:14 +00:00
|
|
|
prepare_data_per_node: bool = True,
|
2020-10-15 21:02:50 +00:00
|
|
|
plugins: Optional[list] = None,
|
2020-08-13 14:03:13 +00:00
|
|
|
amp_backend: str = 'native',
|
2020-10-06 14:27:06 +00:00
|
|
|
amp_level: str = 'O2',
|
2020-10-10 13:21:08 +00:00
|
|
|
distributed_backend: Optional[str] = None,
|
2020-11-14 04:43:42 +00:00
|
|
|
automatic_optimization: Optional[bool] = None,
|
2020-11-10 21:13:41 +00:00
|
|
|
move_metrics_to_cpu: bool = False,
|
2019-12-04 11:57:10 +00:00
|
|
|
):
|
2020-09-30 12:34:19 +00:00
|
|
|
r"""
|
|
|
|
Customize every aspect of training via flags
|
|
|
|
|
|
|
|
Args:
|
|
|
|
|
2020-10-10 13:21:08 +00:00
|
|
|
accelerator: Previously known as distributed_backend (dp, ddp, ddp2, etc...).
|
|
|
|
Can also take in an accelerator object for custom hardware.
|
|
|
|
|
2020-09-30 12:34:19 +00:00
|
|
|
accumulate_grad_batches: Accumulates grads every k batches or as set up in the dict.
|
|
|
|
|
|
|
|
amp_backend: The mixed precision backend to use ("native" or "apex")
|
|
|
|
|
|
|
|
amp_level: The optimization level to use (O1, O2, etc...).
|
|
|
|
|
2020-10-06 02:28:38 +00:00
|
|
|
auto_lr_find: If set to True, will make trainer.tune() run a learning rate finder,
|
|
|
|
trying to optimize initial learning for faster convergence. trainer.tune() method will
|
|
|
|
set the suggested learning rate in self.lr or self.learning_rate in the LightningModule.
|
|
|
|
To use a different key set a string instead of True with the key name.
|
2020-09-30 12:34:19 +00:00
|
|
|
|
|
|
|
auto_scale_batch_size: If set to True, will `initially` run a batch size
|
|
|
|
finder trying to find the largest batch size that fits into memory.
|
|
|
|
The result will be stored in self.batch_size in the LightningModule.
|
|
|
|
Additionally, can be set to either `power` that estimates the batch size through
|
|
|
|
a power search or `binsearch` that estimates the batch size through a binary search.
|
|
|
|
|
|
|
|
auto_select_gpus: If enabled and `gpus` is an integer, pick available
|
|
|
|
gpus automatically. This is especially useful when
|
|
|
|
GPUs are configured to be in "exclusive mode", such
|
|
|
|
that only one process at a time can access them.
|
|
|
|
|
|
|
|
benchmark: If true enables cudnn.benchmark.
|
|
|
|
|
|
|
|
callbacks: Add a list of callbacks.
|
|
|
|
|
2020-10-30 03:47:37 +00:00
|
|
|
checkpoint_callback: If ``True``, enable checkpointing.
|
|
|
|
It will configure a default ModelCheckpoint callback if there is no user-defined ModelCheckpoint in
|
|
|
|
:paramref:`~pytorch_lightning.trainer.trainer.Trainer.callbacks`. Default: ``True``.
|
|
|
|
|
|
|
|
.. warning:: Passing a ModelCheckpoint instance to this argument is deprecated since
|
2020-11-22 06:35:54 +00:00
|
|
|
v1.1 and will be unsupported from v1.3. Use `callbacks` argument instead.
|
2020-09-30 12:34:19 +00:00
|
|
|
|
|
|
|
check_val_every_n_epoch: Check val every n train epochs.
|
|
|
|
|
|
|
|
default_root_dir: Default path for logs and weights when no logger/ckpt_callback passed.
|
|
|
|
Default: ``os.getcwd()``.
|
|
|
|
Can be remote file paths such as `s3://mybucket/path` or 'hdfs://path/'
|
|
|
|
|
|
|
|
deterministic: If true enables cudnn.deterministic.
|
|
|
|
|
2020-10-10 13:21:08 +00:00
|
|
|
distributed_backend: deprecated. Please use 'accelerator'
|
2020-09-30 12:34:19 +00:00
|
|
|
|
|
|
|
fast_dev_run: runs 1 batch of train, test and val to find any bugs (ie: a sort of unit test).
|
|
|
|
|
2020-10-06 14:27:06 +00:00
|
|
|
flush_logs_every_n_steps: How often to flush logs to disk (defaults to every 100 steps).
|
|
|
|
|
2020-09-30 12:34:19 +00:00
|
|
|
gpus: number of gpus to train on (int) or which GPUs to train on (list or str) applied per node
|
|
|
|
|
|
|
|
gradient_clip_val: 0 means don't clip.
|
|
|
|
|
|
|
|
limit_train_batches: How much of training dataset to check (floats = percent, int = num_batches)
|
|
|
|
|
|
|
|
limit_val_batches: How much of validation dataset to check (floats = percent, int = num_batches)
|
|
|
|
|
|
|
|
limit_test_batches: How much of test dataset to check (floats = percent, int = num_batches)
|
|
|
|
|
|
|
|
logger: Logger (or iterable collection of loggers) for experiment tracking.
|
|
|
|
|
|
|
|
log_gpu_memory: None, 'min_max', 'all'. Might slow performance
|
|
|
|
|
2020-10-06 14:27:06 +00:00
|
|
|
log_every_n_steps: How often to log within steps (defaults to every 50 steps).
|
|
|
|
|
2020-10-10 20:44:15 +00:00
|
|
|
automatic_optimization: If False you are responsible for calling .backward, .step, zero_grad.
|
2020-11-14 04:43:42 +00:00
|
|
|
If False you are responsible for calling .backward, .step, zero_grad in LightningModule.
|
|
|
|
This argument has been moved to LightningModule. It is deprecated here in v1.1 and
|
|
|
|
will be removed in v1.3.
|
2020-10-10 20:44:15 +00:00
|
|
|
|
2020-09-30 12:34:19 +00:00
|
|
|
prepare_data_per_node: If True, each LOCAL_RANK=0 will call prepare data.
|
|
|
|
Otherwise only NODE_RANK=0, LOCAL_RANK=0 will prepare data
|
|
|
|
|
|
|
|
process_position: orders the progress bar when running multiple models on same machine.
|
|
|
|
|
|
|
|
progress_bar_refresh_rate: How often to refresh progress bar (in steps). Value ``0`` disables progress bar.
|
|
|
|
Ignored when a custom callback is passed to :paramref:`~Trainer.callbacks`.
|
|
|
|
|
2020-10-27 10:57:16 +00:00
|
|
|
profiler: To profile individual steps during training and assist in identifying bottlenecks. Passing bool
|
|
|
|
value is deprecated in v1.1 and will be removed in v1.3.
|
2020-09-30 12:34:19 +00:00
|
|
|
|
|
|
|
overfit_batches: Overfit a percent of training data (float) or a set number of batches (int). Default: 0.0
|
|
|
|
|
2020-10-10 02:03:46 +00:00
|
|
|
plugins: Plugins allow modification of core behavior like ddp and amp.
|
|
|
|
|
2020-09-30 12:34:19 +00:00
|
|
|
precision: Full precision (32), half precision (16). Can be used on CPU, GPU or TPUs.
|
|
|
|
|
|
|
|
max_epochs: Stop training once this number of epochs is reached.
|
|
|
|
|
|
|
|
min_epochs: Force training for at least these many epochs
|
|
|
|
|
|
|
|
max_steps: Stop training after this number of steps. Disabled by default (None).
|
|
|
|
|
|
|
|
min_steps: Force training for at least these number of steps. Disabled by default (None).
|
|
|
|
|
|
|
|
num_nodes: number of GPU nodes for distributed training.
|
|
|
|
|
2020-11-11 17:05:24 +00:00
|
|
|
num_processes: number of processes for distributed training with distributed_backend="ddp_cpu"
|
|
|
|
|
2020-09-30 12:34:19 +00:00
|
|
|
num_sanity_val_steps: Sanity check runs n validation batches before starting the training routine.
|
|
|
|
Set it to `-1` to run all batches in all validation dataloaders. Default: 2
|
|
|
|
|
|
|
|
reload_dataloaders_every_epoch: Set to True to reload dataloaders every epoch.
|
|
|
|
|
|
|
|
replace_sampler_ddp: Explicitly enables or disables sampler replacement. If not specified this
|
|
|
|
will toggled automatically when DDP is used. By default it will add ``shuffle=True`` for
|
|
|
|
train sampler and ``shuffle=False`` for val/test sampler. If you want to customize it,
|
|
|
|
you can set ``replace_sampler_ddp=False`` and add your own distributed sampler.
|
|
|
|
|
|
|
|
resume_from_checkpoint: To resume training from a specific checkpoint pass in the path here.
|
|
|
|
This can be a URL.
|
|
|
|
|
|
|
|
sync_batchnorm: Synchronize batch norm layers between process groups/whole world.
|
|
|
|
|
|
|
|
terminate_on_nan: If set to True, will terminate training (by raising a `ValueError`) at the
|
|
|
|
end of each training batch, if any of the parameters or the loss are NaN or +/-inf.
|
|
|
|
|
|
|
|
tpu_cores: How many TPU cores to train on (1 or 8) / Single TPU to train on [1]
|
|
|
|
|
|
|
|
track_grad_norm: -1 no tracking. Otherwise tracks that p-norm. May be set to 'inf' infinity-norm.
|
|
|
|
|
|
|
|
truncated_bptt_steps: Truncated back prop breaks performs backprop every k steps of much longer
|
|
|
|
sequence.
|
|
|
|
|
|
|
|
val_check_interval: How often to check the validation set. Use float to check within a training epoch,
|
|
|
|
use int to check every n steps (batches).
|
|
|
|
|
|
|
|
weights_summary: Prints a summary of the weights when training begins.
|
|
|
|
|
|
|
|
weights_save_path: Where to save weights if specified. Will override default_root_dir
|
2020-11-21 01:45:09 +00:00
|
|
|
for checkpoints only. Use this if for whatever reason you need the checkpoints
|
|
|
|
stored in a different place than the logs written in `default_root_dir`.
|
|
|
|
Can be remote file paths such as `s3://mybucket/path` or 'hdfs://path/'
|
|
|
|
Defaults to `default_root_dir`.
|
2020-11-10 21:13:41 +00:00
|
|
|
|
2020-11-21 01:45:09 +00:00
|
|
|
move_metrics_to_cpu: Whether to force internal logged metrics to be moved to CPU.
|
|
|
|
This can save some GPU memory but can make training slower. Use with attention.
|
2020-09-30 12:34:19 +00:00
|
|
|
"""
|
2020-05-17 13:14:54 +00:00
|
|
|
super().__init__()
|
2019-07-18 16:04:19 +00:00
|
|
|
|
2020-09-10 17:21:04 +00:00
|
|
|
# init connectors
|
2020-09-08 22:46:42 +00:00
|
|
|
self.dev_debugger = InternalDebugger(self)
|
|
|
|
self.config_validator = ConfigValidator(self)
|
|
|
|
self.data_connector = DataConnector(self)
|
2020-09-10 17:21:04 +00:00
|
|
|
self.optimizer_connector = OptimizerConnector(self)
|
2020-09-08 22:46:42 +00:00
|
|
|
self.accelerator_connector = AcceleratorConnector(self)
|
|
|
|
self.logger_connector = LoggerConnector(self)
|
2020-09-09 04:24:20 +00:00
|
|
|
self.model_connector = ModelConnector(self)
|
2020-09-10 12:55:30 +00:00
|
|
|
self.precision_connector = PrecisionConnector(self)
|
2020-09-10 12:07:55 +00:00
|
|
|
self.callback_connector = CallbackConnector(self)
|
2020-09-10 12:55:30 +00:00
|
|
|
self.debugging_connector = DebuggingConnector(self)
|
2020-09-10 14:51:35 +00:00
|
|
|
self.training_tricks_connector = TrainingTricksConnector(self)
|
2020-09-10 17:21:04 +00:00
|
|
|
self.profile_connector = ProfilerConnector(self)
|
2020-09-12 11:05:21 +00:00
|
|
|
self.checkpoint_connector = CheckpointConnector(self)
|
2020-09-12 15:07:15 +00:00
|
|
|
self.slurm_connector = SLURMConnector(self)
|
2020-09-08 22:46:42 +00:00
|
|
|
self.tuner = Tuner(self)
|
|
|
|
self.accelerator_backend = None
|
|
|
|
self.evaluation_loop = EvaluationLoop(self)
|
|
|
|
self.train_loop = TrainLoop(self)
|
2020-10-10 02:03:46 +00:00
|
|
|
self.plugin_connector = PluginConnector(self)
|
2020-09-08 22:46:42 +00:00
|
|
|
|
2020-06-29 01:36:46 +00:00
|
|
|
# training state
|
2020-09-10 14:51:35 +00:00
|
|
|
self.weights_summary = weights_summary
|
2020-06-29 01:36:46 +00:00
|
|
|
self.model = None
|
2020-09-10 17:21:04 +00:00
|
|
|
self.shown_warnings = set()
|
2020-06-29 01:36:46 +00:00
|
|
|
|
|
|
|
# init callbacks
|
2020-10-04 02:18:05 +00:00
|
|
|
# Declare attributes to be set in callback_connector on_trainer_init
|
2020-09-10 12:07:55 +00:00
|
|
|
self.callback_connector.on_trainer_init(
|
|
|
|
callbacks,
|
|
|
|
checkpoint_callback,
|
|
|
|
progress_bar_refresh_rate,
|
2020-09-10 14:51:35 +00:00
|
|
|
process_position,
|
|
|
|
default_root_dir,
|
|
|
|
weights_save_path,
|
2020-10-06 14:27:06 +00:00
|
|
|
resume_from_checkpoint,
|
2020-09-10 12:07:55 +00:00
|
|
|
)
|
2020-06-29 01:36:46 +00:00
|
|
|
|
2020-09-10 14:51:35 +00:00
|
|
|
# hook
|
|
|
|
self.on_init_start()
|
2020-06-02 22:51:09 +00:00
|
|
|
|
2020-09-10 17:21:04 +00:00
|
|
|
# init optimizer + lr scheduler related flags
|
|
|
|
self.optimizer_connector.on_trainer_init()
|
|
|
|
|
|
|
|
# init data flags
|
|
|
|
self.data_connector.on_trainer_init(
|
2020-10-06 14:27:06 +00:00
|
|
|
check_val_every_n_epoch, reload_dataloaders_every_epoch, prepare_data_per_node
|
2020-09-10 17:21:04 +00:00
|
|
|
)
|
|
|
|
|
2020-09-10 14:51:35 +00:00
|
|
|
# init training tricks
|
|
|
|
self.training_tricks_connector.on_trainer_init(
|
2020-10-06 14:27:06 +00:00
|
|
|
gradient_clip_val, track_grad_norm, accumulate_grad_batches, truncated_bptt_steps, terminate_on_nan
|
2020-09-10 14:51:35 +00:00
|
|
|
)
|
2020-06-02 22:51:09 +00:00
|
|
|
|
2020-09-10 11:24:42 +00:00
|
|
|
# init accelerator related flags
|
|
|
|
self.accelerator_connector.on_trainer_init(
|
|
|
|
num_processes,
|
|
|
|
tpu_cores,
|
2020-10-10 13:21:08 +00:00
|
|
|
accelerator,
|
2020-09-10 11:24:42 +00:00
|
|
|
distributed_backend,
|
|
|
|
auto_select_gpus,
|
2020-09-10 12:07:55 +00:00
|
|
|
gpus,
|
|
|
|
num_nodes,
|
|
|
|
log_gpu_memory,
|
|
|
|
sync_batchnorm,
|
2020-09-10 14:51:35 +00:00
|
|
|
benchmark,
|
2020-09-10 17:21:04 +00:00
|
|
|
replace_sampler_ddp,
|
2020-10-04 12:48:46 +00:00
|
|
|
deterministic,
|
2020-09-10 11:24:42 +00:00
|
|
|
)
|
|
|
|
|
2020-09-10 12:55:30 +00:00
|
|
|
# init train loop related flags
|
2020-11-14 04:43:42 +00:00
|
|
|
# TODO: deprecate in 1.2.0
|
|
|
|
if automatic_optimization is None:
|
|
|
|
automatic_optimization = True
|
|
|
|
else:
|
|
|
|
rank_zero_warn(
|
|
|
|
"Disable automatic optimization with the trainer flag is deprecated and will be removed in v1.3.0!"
|
|
|
|
"Please use the property on the LightningModule for disabling automatic optimization"
|
|
|
|
)
|
2020-10-10 20:44:15 +00:00
|
|
|
self.train_loop.on_trainer_init(
|
|
|
|
max_epochs,
|
|
|
|
min_epochs,
|
|
|
|
max_steps,
|
|
|
|
min_steps,
|
|
|
|
num_sanity_val_steps,
|
|
|
|
automatic_optimization
|
|
|
|
)
|
2020-09-10 17:21:04 +00:00
|
|
|
self.evaluation_loop.on_trainer_init()
|
2020-07-27 21:56:55 +00:00
|
|
|
|
2020-09-10 17:21:04 +00:00
|
|
|
# configure tuner
|
|
|
|
self.tuner.on_trainer_init(auto_lr_find, auto_scale_batch_size)
|
2019-07-08 13:42:13 +00:00
|
|
|
|
2020-02-07 03:01:21 +00:00
|
|
|
# configure profiler
|
2020-09-10 17:21:04 +00:00
|
|
|
self.profile_connector.on_trainer_init(profiler)
|
2020-02-07 03:01:21 +00:00
|
|
|
|
2020-09-10 14:51:35 +00:00
|
|
|
# init logger flags
|
2020-11-10 21:13:41 +00:00
|
|
|
self.logger_connector.on_trainer_init(
|
|
|
|
logger,
|
|
|
|
flush_logs_every_n_steps,
|
|
|
|
log_every_n_steps,
|
|
|
|
move_metrics_to_cpu
|
|
|
|
)
|
2019-09-06 04:29:38 +00:00
|
|
|
|
2020-09-10 12:55:30 +00:00
|
|
|
# init debugging flags
|
|
|
|
self.debugging_connector.on_init_start(
|
|
|
|
limit_train_batches,
|
|
|
|
limit_val_batches,
|
|
|
|
limit_test_batches,
|
|
|
|
val_check_interval,
|
|
|
|
overfit_batches,
|
2020-10-06 14:27:06 +00:00
|
|
|
fast_dev_run,
|
2020-09-10 12:55:30 +00:00
|
|
|
)
|
2019-09-06 04:29:38 +00:00
|
|
|
|
2020-09-10 12:55:30 +00:00
|
|
|
# set precision
|
|
|
|
self.precision_connector.on_trainer_init(precision, amp_level, amp_backend)
|
2020-05-13 23:17:04 +00:00
|
|
|
|
2020-10-10 02:03:46 +00:00
|
|
|
# last thing are the plugins which override whatever the trainer used by default
|
|
|
|
self.plugin_connector.on_trainer_init(plugins)
|
|
|
|
|
2020-02-26 04:17:27 +00:00
|
|
|
# Callback system
|
2020-03-03 04:51:32 +00:00
|
|
|
self.on_init_end()
|
2020-02-26 04:17:27 +00:00
|
|
|
|
2020-02-23 02:23:30 +00:00
|
|
|
def fit(
|
2020-07-24 15:42:15 +00:00
|
|
|
self,
|
|
|
|
model: LightningModule,
|
|
|
|
train_dataloader: Optional[DataLoader] = None,
|
|
|
|
val_dataloaders: Optional[Union[DataLoader, List[DataLoader]]] = None,
|
|
|
|
datamodule: Optional[LightningDataModule] = None,
|
2020-02-23 02:23:30 +00:00
|
|
|
):
|
2020-09-30 12:34:19 +00:00
|
|
|
r"""
|
|
|
|
Runs the full optimization routine.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
datamodule: A instance of :class:`LightningDataModule`.
|
|
|
|
|
|
|
|
model: Model to fit.
|
|
|
|
|
|
|
|
train_dataloader: A Pytorch DataLoader with training samples. If the model has
|
|
|
|
a predefined train_dataloader method this will be skipped.
|
|
|
|
|
|
|
|
val_dataloaders: Either a single Pytorch Dataloader or a list of them, specifying validation samples.
|
|
|
|
If the model has a predefined val_dataloaders method this will be skipped
|
|
|
|
|
|
|
|
"""
|
2020-10-07 15:04:10 +00:00
|
|
|
# bookkeeping
|
2020-09-21 02:58:43 +00:00
|
|
|
self._state = TrainerState.RUNNING
|
|
|
|
|
2020-10-07 15:04:10 +00:00
|
|
|
# ----------------------------
|
|
|
|
# LINK DATA
|
|
|
|
# ----------------------------
|
2020-08-31 15:08:22 +00:00
|
|
|
# setup data, etc...
|
2020-09-11 01:58:47 +00:00
|
|
|
self.train_loop.setup_fit(model, train_dataloader, val_dataloaders, datamodule)
|
2020-04-02 15:53:37 +00:00
|
|
|
|
2020-08-27 02:20:00 +00:00
|
|
|
# hook
|
2020-09-01 18:59:09 +00:00
|
|
|
self.data_connector.prepare_data(model)
|
2020-02-19 11:00:08 +00:00
|
|
|
|
2020-10-07 15:04:10 +00:00
|
|
|
# bookkeeping
|
|
|
|
# we reuse fit in .test() but change its behavior using this flag
|
2020-08-02 12:13:31 +00:00
|
|
|
self.testing = os.environ.get('PL_TESTING_MODE', self.testing)
|
2020-06-13 16:00:14 +00:00
|
|
|
|
2020-10-07 15:04:10 +00:00
|
|
|
# ----------------------------
|
|
|
|
# SET UP TRAINING
|
|
|
|
# ----------------------------
|
2020-09-01 19:48:28 +00:00
|
|
|
self.accelerator_backend = self.accelerator_connector.select_accelerator()
|
2020-08-27 01:29:10 +00:00
|
|
|
self.accelerator_backend.setup(model)
|
2020-09-23 08:38:33 +00:00
|
|
|
|
2020-10-07 15:04:10 +00:00
|
|
|
# ----------------------------
|
|
|
|
# INSPECT THESE FOR MAIN LOOPS
|
|
|
|
# ----------------------------
|
|
|
|
# assign training and eval functions... inspect these to see the train and eval loops :)
|
|
|
|
self.accelerator_backend.train_loop = self.train
|
|
|
|
self.accelerator_backend.validation_loop = self.run_evaluation
|
|
|
|
self.accelerator_backend.test_loop = self.run_evaluation
|
|
|
|
|
|
|
|
# ----------------------------
|
|
|
|
# TRAIN
|
|
|
|
# ----------------------------
|
2020-09-23 08:38:33 +00:00
|
|
|
# hook
|
|
|
|
self.call_hook('on_fit_start')
|
|
|
|
|
2020-08-27 01:29:10 +00:00
|
|
|
results = self.accelerator_backend.train()
|
|
|
|
self.accelerator_backend.teardown()
|
|
|
|
|
2020-10-07 15:04:10 +00:00
|
|
|
# ----------------------------
|
|
|
|
# POST-Training CLEAN UP
|
|
|
|
# ----------------------------
|
2020-08-27 01:29:10 +00:00
|
|
|
# hook
|
|
|
|
self.call_hook('on_fit_end')
|
|
|
|
|
|
|
|
# hook
|
|
|
|
self.teardown('fit')
|
|
|
|
if self.is_function_implemented('teardown'):
|
|
|
|
model.teardown('fit')
|
|
|
|
|
|
|
|
# return 1 when finished
|
|
|
|
# used for testing or when we need to know that training succeeded
|
2020-09-21 02:58:43 +00:00
|
|
|
|
|
|
|
if self._state != TrainerState.INTERRUPTED:
|
|
|
|
self._state = TrainerState.FINISHED
|
2020-08-27 01:29:10 +00:00
|
|
|
return results or 1
|
|
|
|
|
2020-09-06 21:50:47 +00:00
|
|
|
def train(self):
|
|
|
|
self.run_sanity_check(self.get_model())
|
|
|
|
|
2020-11-02 20:51:43 +00:00
|
|
|
# set stage for logging
|
|
|
|
self.logger_connector.set_stage("train")
|
|
|
|
|
2020-10-23 10:17:50 +00:00
|
|
|
self.checkpoint_connector.has_trained = False
|
|
|
|
|
2020-09-06 21:50:47 +00:00
|
|
|
# enable train mode
|
|
|
|
model = self.get_model()
|
|
|
|
model.train()
|
|
|
|
torch.set_grad_enabled(True)
|
|
|
|
|
|
|
|
# reload data when needed
|
|
|
|
self.train_loop.reset_train_val_dataloaders(model)
|
|
|
|
|
|
|
|
# hook
|
|
|
|
self.train_loop.on_train_start()
|
|
|
|
|
2020-11-03 06:40:35 +00:00
|
|
|
if self.train_loop.should_skip_training():
|
|
|
|
self.train_loop.on_train_end()
|
|
|
|
return
|
|
|
|
|
2020-09-06 21:50:47 +00:00
|
|
|
try:
|
|
|
|
# run all epochs
|
|
|
|
for epoch in range(self.current_epoch, self.max_epochs):
|
|
|
|
|
|
|
|
# hook
|
|
|
|
self.train_loop.on_train_epoch_start(epoch)
|
|
|
|
|
|
|
|
# run train epoch
|
2020-09-06 23:55:18 +00:00
|
|
|
self.train_loop.run_training_epoch()
|
2020-09-06 21:50:47 +00:00
|
|
|
|
|
|
|
if self.max_steps and self.max_steps <= self.global_step:
|
|
|
|
|
|
|
|
# hook
|
|
|
|
self.train_loop.on_train_end()
|
|
|
|
return
|
|
|
|
|
|
|
|
# update LR schedulers
|
2020-09-10 17:21:04 +00:00
|
|
|
self.optimizer_connector.update_learning_rates(interval='epoch')
|
2020-09-06 21:50:47 +00:00
|
|
|
|
|
|
|
# early stopping
|
|
|
|
met_min_epochs = epoch >= self.min_epochs - 1
|
|
|
|
met_min_steps = self.global_step >= self.min_steps if self.min_steps else True
|
|
|
|
|
|
|
|
if self.should_stop:
|
2020-10-06 14:27:06 +00:00
|
|
|
if met_min_epochs and met_min_steps:
|
2020-09-06 21:50:47 +00:00
|
|
|
self.train_loop.on_train_end()
|
|
|
|
return
|
|
|
|
else:
|
2020-10-06 14:27:06 +00:00
|
|
|
log.info(
|
|
|
|
'Trainer was signaled to stop but required minimum epochs'
|
|
|
|
f' ({self.min_epochs}) or minimum steps ({self.min_steps}) has'
|
|
|
|
' not been met. Training will continue...'
|
|
|
|
)
|
2020-09-06 21:50:47 +00:00
|
|
|
|
|
|
|
# hook
|
|
|
|
self.train_loop.on_train_end()
|
|
|
|
|
|
|
|
except KeyboardInterrupt:
|
|
|
|
rank_zero_warn('Detected KeyboardInterrupt, attempting graceful shutdown...')
|
|
|
|
|
|
|
|
# user could press ctrl+c many times... only shutdown once
|
|
|
|
if not self.interrupted:
|
|
|
|
self.interrupted = True
|
|
|
|
self._state = TrainerState.INTERRUPTED
|
|
|
|
self.on_keyboard_interrupt()
|
|
|
|
|
|
|
|
# hook
|
|
|
|
self.train_loop.on_train_end()
|
|
|
|
|
2020-09-09 12:45:04 +00:00
|
|
|
def run_evaluation(self, test_mode: bool = False, max_batches=None):
|
2020-11-02 20:51:43 +00:00
|
|
|
|
|
|
|
# used to know if we are logging for val, test + reset cached results
|
|
|
|
self.logger_connector.set_stage(test_mode, reset=True)
|
|
|
|
|
2020-09-09 12:45:04 +00:00
|
|
|
# bookkeeping
|
|
|
|
self.evaluation_loop.testing = test_mode
|
2020-11-02 20:51:43 +00:00
|
|
|
|
|
|
|
# prepare dataloaders
|
2020-09-09 12:45:04 +00:00
|
|
|
dataloaders, max_batches = self.evaluation_loop.get_evaluation_dataloaders(max_batches)
|
2020-11-02 20:51:43 +00:00
|
|
|
|
|
|
|
# check if we want to skip this evaluation
|
2020-09-09 12:45:04 +00:00
|
|
|
if self.evaluation_loop.should_skip_evaluation(dataloaders, max_batches):
|
|
|
|
return [], []
|
|
|
|
|
2020-11-02 20:51:43 +00:00
|
|
|
# ref model
|
2020-09-09 12:45:04 +00:00
|
|
|
model = self.get_model()
|
2020-10-05 03:02:35 +00:00
|
|
|
|
2020-11-02 20:51:43 +00:00
|
|
|
# enable eval mode + no grads
|
|
|
|
self.evaluation_loop.on_evaluation_model_eval()
|
2020-09-09 12:45:04 +00:00
|
|
|
model.zero_grad()
|
|
|
|
torch.set_grad_enabled(False)
|
|
|
|
|
|
|
|
# hook
|
|
|
|
self.evaluation_loop.on_evaluation_start()
|
|
|
|
|
|
|
|
# set up the eval loop
|
|
|
|
self.evaluation_loop.setup(model, max_batches, dataloaders)
|
|
|
|
|
|
|
|
# hook
|
|
|
|
self.evaluation_loop.on_evaluation_epoch_start()
|
|
|
|
|
|
|
|
# run validation/testing
|
|
|
|
for dataloader_idx, dataloader in enumerate(dataloaders):
|
|
|
|
# bookkeeping
|
|
|
|
dl_outputs = []
|
|
|
|
dataloader = self.accelerator_backend.process_dataloader(dataloader)
|
|
|
|
dl_max_batches = self.evaluation_loop.max_batches[dataloader_idx]
|
|
|
|
|
|
|
|
for batch_idx, batch in enumerate(dataloader):
|
|
|
|
if batch is None:
|
|
|
|
continue
|
|
|
|
|
|
|
|
# stop short when running on limited batches
|
|
|
|
if batch_idx >= dl_max_batches:
|
|
|
|
break
|
|
|
|
|
|
|
|
# hook
|
|
|
|
self.evaluation_loop.on_evaluation_batch_start(batch, batch_idx, dataloader_idx)
|
|
|
|
|
|
|
|
# lightning module methods
|
|
|
|
output = self.evaluation_loop.evaluation_step(test_mode, batch, batch_idx, dataloader_idx)
|
|
|
|
output = self.evaluation_loop.evaluation_step_end(output)
|
|
|
|
|
2020-11-11 17:05:24 +00:00
|
|
|
# hook + store predictions
|
2020-10-08 00:41:56 +00:00
|
|
|
self.evaluation_loop.on_evaluation_batch_end(output, batch, batch_idx, dataloader_idx)
|
2020-09-09 12:45:04 +00:00
|
|
|
|
2020-11-11 17:05:24 +00:00
|
|
|
# log batch metrics
|
|
|
|
self.evaluation_loop.log_evaluation_step_metrics(output, batch_idx)
|
2020-09-29 06:00:28 +00:00
|
|
|
|
|
|
|
# track epoch level outputs
|
2020-11-10 21:13:41 +00:00
|
|
|
dl_outputs = self.track_output_for_epoch_end(dl_outputs, output)
|
2020-09-09 12:45:04 +00:00
|
|
|
|
2020-11-11 17:05:24 +00:00
|
|
|
# store batch level output per dataloader
|
2020-09-09 12:45:04 +00:00
|
|
|
self.evaluation_loop.outputs.append(dl_outputs)
|
|
|
|
|
|
|
|
# lightning module method
|
2020-11-11 17:05:24 +00:00
|
|
|
deprecated_eval_results = self.evaluation_loop.evaluation_epoch_end()
|
2020-09-09 12:45:04 +00:00
|
|
|
|
|
|
|
# hook
|
|
|
|
self.evaluation_loop.on_evaluation_epoch_end()
|
|
|
|
|
2020-11-11 17:05:24 +00:00
|
|
|
# hook
|
|
|
|
self.evaluation_loop.on_evaluation_end()
|
|
|
|
|
|
|
|
# log epoch metrics
|
|
|
|
eval_loop_results = self.evaluation_loop.log_epoch_metrics_on_evaluation_end()
|
|
|
|
|
|
|
|
# save predictions to disk
|
|
|
|
self.evaluation_loop.predictions.to_disk()
|
|
|
|
|
2020-09-09 12:45:04 +00:00
|
|
|
# enable train mode again
|
2020-10-05 03:02:35 +00:00
|
|
|
self.evaluation_loop.on_evaluation_model_train()
|
2020-09-09 12:45:04 +00:00
|
|
|
torch.set_grad_enabled(True)
|
|
|
|
|
2020-10-04 17:36:35 +00:00
|
|
|
return eval_loop_results, deprecated_eval_results
|
2020-09-09 12:45:04 +00:00
|
|
|
|
2020-11-10 21:13:41 +00:00
|
|
|
def track_output_for_epoch_end(self, outputs, output):
|
|
|
|
if output is not None:
|
|
|
|
if isinstance(output, Result):
|
|
|
|
output.detach()
|
|
|
|
if self.move_metrics_to_cpu:
|
|
|
|
output.cpu()
|
|
|
|
elif isinstance(output, dict):
|
|
|
|
output = recursive_detach(output, to_cpu=self.move_metrics_to_cpu)
|
|
|
|
elif isinstance(output, torch.Tensor) and output.is_cuda and self.move_metrics_to_cpu:
|
|
|
|
output = output.cpu()
|
|
|
|
outputs.append(output)
|
|
|
|
return outputs
|
|
|
|
|
2020-09-01 00:36:52 +00:00
|
|
|
def run_test(self):
|
|
|
|
# only load test dataloader for testing
|
|
|
|
# self.reset_test_dataloader(ref_model)
|
|
|
|
eval_loop_results, _ = self.run_evaluation(test_mode=True)
|
2020-07-07 16:24:56 +00:00
|
|
|
|
2020-09-01 00:36:52 +00:00
|
|
|
if len(eval_loop_results) == 0:
|
|
|
|
return 1
|
2019-08-30 22:56:09 +00:00
|
|
|
|
2020-09-01 00:36:52 +00:00
|
|
|
# remove the tensors from the eval results
|
|
|
|
for i, result in enumerate(eval_loop_results):
|
|
|
|
if isinstance(result, dict):
|
|
|
|
for k, v in result.items():
|
|
|
|
if isinstance(v, torch.Tensor):
|
|
|
|
result[k] = v.cpu().item()
|
2020-07-14 18:20:45 +00:00
|
|
|
|
2020-09-01 00:36:52 +00:00
|
|
|
return eval_loop_results
|
2020-07-14 18:20:45 +00:00
|
|
|
|
2020-09-01 00:36:52 +00:00
|
|
|
def run_sanity_check(self, ref_model):
|
|
|
|
using_val_step = ref_model.val_dataloader is not None and is_overridden('validation_step', ref_model)
|
2020-07-25 16:57:40 +00:00
|
|
|
should_sanity_check = using_val_step and self.num_sanity_val_steps > 0 and self.limit_val_batches > 0
|
2020-07-23 11:07:03 +00:00
|
|
|
|
2019-08-30 22:56:09 +00:00
|
|
|
# run tiny validation (if validation defined)
|
|
|
|
# to make sure program won't crash during val
|
2020-07-23 11:07:03 +00:00
|
|
|
if should_sanity_check:
|
2020-02-26 21:55:18 +00:00
|
|
|
self.reset_val_dataloader(ref_model)
|
2020-08-21 18:11:31 +00:00
|
|
|
self.num_sanity_val_batches = [
|
|
|
|
min(self.num_sanity_val_steps, val_batches) for val_batches in self.num_val_batches
|
|
|
|
]
|
2020-04-24 00:46:18 +00:00
|
|
|
|
|
|
|
# hook and callback
|
2020-07-22 17:53:10 +00:00
|
|
|
self.running_sanity_check = True
|
2020-04-24 00:46:18 +00:00
|
|
|
self.on_sanity_check_start()
|
2019-08-24 01:23:27 +00:00
|
|
|
|
2020-08-26 16:28:14 +00:00
|
|
|
# run eval step
|
|
|
|
_, eval_results = self.run_evaluation(test_mode=False, max_batches=self.num_sanity_val_batches)
|
2020-07-01 11:38:00 +00:00
|
|
|
|
|
|
|
# allow no returns from eval
|
|
|
|
if eval_results is not None and len(eval_results) > 0:
|
2020-07-14 18:20:45 +00:00
|
|
|
# when we get a list back, used only the last item
|
|
|
|
if isinstance(eval_results, list):
|
|
|
|
eval_results = eval_results[-1]
|
2020-07-22 17:53:10 +00:00
|
|
|
|
|
|
|
if isinstance(eval_results, EvalResult):
|
|
|
|
callback_metrics = eval_results.callback_metrics
|
|
|
|
else:
|
2020-09-21 02:58:43 +00:00
|
|
|
_, _, _, callback_metrics, _ = self.process_dict_result(eval_results)
|
2020-09-07 13:31:42 +00:00
|
|
|
self.logger_connector.callback_metrics = callback_metrics
|
2019-08-07 11:51:55 +00:00
|
|
|
|
2020-04-24 00:46:18 +00:00
|
|
|
self.on_sanity_check_end()
|
2020-07-22 17:53:10 +00:00
|
|
|
self.running_sanity_check = False
|
2019-11-03 10:42:53 +00:00
|
|
|
|
2020-05-04 12:24:34 +00:00
|
|
|
def test(
|
2020-07-24 15:42:15 +00:00
|
|
|
self,
|
|
|
|
model: Optional[LightningModule] = None,
|
|
|
|
test_dataloaders: Optional[Union[DataLoader, List[DataLoader]]] = None,
|
|
|
|
ckpt_path: Optional[str] = 'best',
|
|
|
|
verbose: bool = True,
|
|
|
|
datamodule: Optional[LightningDataModule] = None,
|
2020-05-04 12:24:34 +00:00
|
|
|
):
|
2020-09-30 12:34:19 +00:00
|
|
|
r"""
|
|
|
|
|
|
|
|
Separates from fit to make sure you never run on your test set until you want to.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
ckpt_path: Either ``best`` or path to the checkpoint you wish to test.
|
|
|
|
If ``None``, use the weights from the last epoch to test. Default to ``best``.
|
|
|
|
|
|
|
|
datamodule: A instance of :class:`LightningDataModule`.
|
|
|
|
|
|
|
|
model: The model to test.
|
|
|
|
|
|
|
|
test_dataloaders: Either a single
|
|
|
|
Pytorch Dataloader or a list of them, specifying validation samples.
|
|
|
|
|
|
|
|
verbose: If True, prints the test results
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
The final test result dictionary. If no test_epoch_end is defined returns a list of dictionaries
|
|
|
|
"""
|
2020-07-07 16:24:56 +00:00
|
|
|
# --------------------
|
|
|
|
# SETUP HOOK
|
|
|
|
# --------------------
|
2020-07-14 18:20:45 +00:00
|
|
|
self.verbose_test = verbose
|
|
|
|
|
2020-11-02 20:51:43 +00:00
|
|
|
self.logger_connector.set_stage("test")
|
|
|
|
|
2020-07-24 15:42:15 +00:00
|
|
|
# If you supply a datamodule you can't supply train_dataloader or val_dataloaders
|
|
|
|
if test_dataloaders and datamodule:
|
|
|
|
raise MisconfigurationException(
|
|
|
|
'You cannot pass test_dataloaders to trainer.test if you supply a datamodule'
|
|
|
|
)
|
|
|
|
|
|
|
|
# Attach datamodule to get setup/prepare_data added to model before the call to it below
|
2020-08-31 15:08:22 +00:00
|
|
|
self.data_connector.attach_datamodule(model or self.get_model(), datamodule, 'test')
|
2020-07-09 22:36:36 +00:00
|
|
|
|
|
|
|
if model is not None:
|
|
|
|
results = self.__test_given_model(model, test_dataloaders)
|
|
|
|
else:
|
|
|
|
results = self.__test_using_best_weights(ckpt_path, test_dataloaders)
|
|
|
|
|
|
|
|
self.teardown('test')
|
|
|
|
|
|
|
|
return results
|
|
|
|
|
|
|
|
def __test_using_best_weights(self, ckpt_path, test_dataloaders):
|
|
|
|
model = self.get_model()
|
2020-06-17 23:49:58 +00:00
|
|
|
|
2020-07-07 16:24:56 +00:00
|
|
|
# if user requests the best checkpoint but we don't have it, error
|
2020-09-28 00:05:02 +00:00
|
|
|
if ckpt_path == 'best' and not self.checkpoint_callback.best_model_path:
|
2020-06-15 12:02:37 +00:00
|
|
|
raise MisconfigurationException(
|
2020-07-24 15:42:15 +00:00
|
|
|
'ckpt_path is "best", but ModelCheckpoint is not configured to save the best model.'
|
|
|
|
)
|
2020-06-15 12:02:37 +00:00
|
|
|
|
2020-07-09 22:36:36 +00:00
|
|
|
# load best weights
|
|
|
|
if ckpt_path is not None:
|
2020-06-15 12:02:37 +00:00
|
|
|
# ckpt_path is 'best' so load the best model
|
|
|
|
if ckpt_path == 'best':
|
|
|
|
ckpt_path = self.checkpoint_callback.best_model_path
|
2020-03-03 04:38:47 +00:00
|
|
|
|
2020-07-10 01:28:11 +00:00
|
|
|
if len(ckpt_path) == 0:
|
2020-07-24 15:42:15 +00:00
|
|
|
rank_zero_warn(
|
|
|
|
f'.test() found no path for the best weights, {ckpt_path}. Please '
|
|
|
|
f'specify a path for a checkpoint .test(ckpt_path=PATH)'
|
|
|
|
)
|
2020-07-10 01:28:11 +00:00
|
|
|
return {}
|
2020-10-01 13:25:33 +00:00
|
|
|
if self.accelerator_backend is not None:
|
|
|
|
self.accelerator_backend.barrier()
|
2020-07-10 01:28:11 +00:00
|
|
|
|
2020-09-28 11:53:57 +00:00
|
|
|
ckpt = pl_load(ckpt_path, map_location=lambda storage, loc: storage)
|
2020-07-09 22:36:36 +00:00
|
|
|
model.load_state_dict(ckpt['state_dict'])
|
2020-04-10 15:44:03 +00:00
|
|
|
|
2020-07-09 22:36:36 +00:00
|
|
|
# attach dataloaders
|
2020-04-10 15:44:03 +00:00
|
|
|
if test_dataloaders is not None:
|
2020-08-31 15:08:22 +00:00
|
|
|
self.data_connector.attach_dataloaders(model, test_dataloaders=test_dataloaders)
|
2020-04-10 15:44:03 +00:00
|
|
|
|
2020-07-09 22:36:36 +00:00
|
|
|
# run tests
|
|
|
|
self.tested_ckpt_path = ckpt_path
|
2020-07-07 16:24:56 +00:00
|
|
|
self.testing = True
|
2020-07-09 22:36:36 +00:00
|
|
|
os.environ['PL_TESTING_MODE'] = '1'
|
2020-07-07 16:24:56 +00:00
|
|
|
self.model = model
|
|
|
|
results = self.fit(model)
|
2020-03-06 11:57:14 +00:00
|
|
|
self.testing = False
|
2020-07-09 22:36:36 +00:00
|
|
|
del os.environ['PL_TESTING_MODE']
|
2020-03-06 11:57:14 +00:00
|
|
|
|
2020-07-09 22:36:36 +00:00
|
|
|
# teardown
|
2020-06-17 23:49:58 +00:00
|
|
|
if self.is_function_implemented('teardown'):
|
2020-06-25 15:10:17 +00:00
|
|
|
model_ref = self.get_model()
|
|
|
|
model_ref.teardown('test')
|
2020-06-17 23:49:58 +00:00
|
|
|
|
2020-07-07 16:24:56 +00:00
|
|
|
return results
|
|
|
|
|
2020-07-09 22:36:36 +00:00
|
|
|
def __test_given_model(self, model, test_dataloaders):
|
|
|
|
|
|
|
|
# attach data
|
|
|
|
if test_dataloaders is not None:
|
2020-08-31 15:08:22 +00:00
|
|
|
self.data_connector.attach_dataloaders(model, test_dataloaders=test_dataloaders)
|
2020-07-09 22:36:36 +00:00
|
|
|
|
|
|
|
# run test
|
|
|
|
# sets up testing so we short circuit to eval
|
|
|
|
self.testing = True
|
|
|
|
self.model = model
|
|
|
|
results = self.fit(model)
|
|
|
|
self.testing = False
|
|
|
|
|
|
|
|
# teardown
|
|
|
|
if self.is_function_implemented('teardown'):
|
|
|
|
model.teardown('test')
|
|
|
|
|
|
|
|
return results
|
|
|
|
|
2020-10-07 15:04:10 +00:00
|
|
|
def tune(
|
|
|
|
self,
|
|
|
|
model: LightningModule,
|
|
|
|
train_dataloader: Optional[DataLoader] = None,
|
|
|
|
val_dataloaders: Optional[Union[DataLoader, List[DataLoader]]] = None,
|
|
|
|
datamodule: Optional[LightningDataModule] = None,
|
|
|
|
):
|
|
|
|
r"""
|
|
|
|
Runs routines to tune hyperparameters before training.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
datamodule: A instance of :class:`LightningDataModule`.
|
|
|
|
|
|
|
|
model: Model to tune.
|
|
|
|
|
|
|
|
train_dataloader: A Pytorch DataLoader with training samples. If the model has
|
|
|
|
a predefined train_dataloader method this will be skipped.
|
|
|
|
|
|
|
|
val_dataloaders: Either a single Pytorch Dataloader or a list of them, specifying validation samples.
|
|
|
|
If the model has a predefined val_dataloaders method this will be skipped
|
|
|
|
|
|
|
|
"""
|
|
|
|
self.tuner.tune(model, train_dataloader, val_dataloaders, datamodule)
|
|
|
|
|
2020-08-02 00:17:57 +00:00
|
|
|
def call_setup_hook(self, model):
|
|
|
|
# call setup after the ddp process has connected
|
|
|
|
stage_name = 'test' if self.testing else 'fit'
|
|
|
|
if self.datamodule is not None:
|
|
|
|
called = self.datamodule.has_setup_test if self.testing else self.datamodule.has_setup_fit
|
|
|
|
if not called:
|
|
|
|
self.datamodule.setup(stage_name)
|
2020-11-14 00:34:46 +00:00
|
|
|
self.setup(model, stage_name)
|
2020-08-02 00:17:57 +00:00
|
|
|
model.setup(stage_name)
|
|
|
|
|
2020-11-05 22:27:04 +00:00
|
|
|
def _reset_result_and_set_hook_fx_name(self, hook_name):
|
|
|
|
model_ref = self.get_model()
|
|
|
|
if model_ref is not None:
|
|
|
|
# used to track current hook name called
|
|
|
|
model_ref._results = Result()
|
|
|
|
model_ref._current_hook_fx_name = hook_name
|
|
|
|
|
|
|
|
def _cache_logged_metrics(self):
|
|
|
|
model_ref = self.get_model()
|
|
|
|
if model_ref is not None:
|
|
|
|
# capture logging for this hook
|
|
|
|
self.logger_connector.cache_logged_metrics()
|
|
|
|
|
2020-08-24 17:46:46 +00:00
|
|
|
def call_hook(self, hook_name, *args, **kwargs):
|
2020-11-11 17:05:24 +00:00
|
|
|
# set hook_name to model + reset Result obj
|
|
|
|
self._reset_result_and_set_hook_fx_name(hook_name)
|
2020-11-05 22:27:04 +00:00
|
|
|
|
2020-08-24 19:48:14 +00:00
|
|
|
# always profile hooks
|
|
|
|
with self.profiler.profile(hook_name):
|
|
|
|
|
|
|
|
# first call trainer hook
|
|
|
|
if hasattr(self, hook_name):
|
|
|
|
trainer_hook = getattr(self, hook_name)
|
|
|
|
trainer_hook(*args, **kwargs)
|
|
|
|
|
|
|
|
# next call hook in lightningModule
|
|
|
|
output = None
|
2020-08-31 16:12:02 +00:00
|
|
|
model_ref = self.get_model()
|
|
|
|
if is_overridden(hook_name, model_ref):
|
2020-08-24 17:46:46 +00:00
|
|
|
hook_fx = getattr(model_ref, hook_name)
|
|
|
|
output = hook_fx(*args, **kwargs)
|
|
|
|
|
2020-08-24 21:50:47 +00:00
|
|
|
# if the PL module doesn't have the hook then call the accelator
|
|
|
|
# used to auto-reduce things for the user with Results obj
|
|
|
|
elif hasattr(self.accelerator_backend, hook_name):
|
|
|
|
accelerator_hook = getattr(self.accelerator_backend, hook_name)
|
|
|
|
output = accelerator_hook(*args, **kwargs)
|
|
|
|
|
2020-11-11 17:05:24 +00:00
|
|
|
# capture logging
|
|
|
|
self._cache_logged_metrics()
|
2020-11-05 22:27:04 +00:00
|
|
|
return output
|