Swap torch.load for fsspec load in ddp spawn backend (#3787)
* Update ddp_spawn_backend.py * Update ddp_cpu_spawn_backend.py * log Co-authored-by: Jirka Borovec <jirka@pytorchlightning.ai>
This commit is contained in:
parent
192fc018f3
commit
3ab730e316
|
@ -49,6 +49,8 @@ The format is based on [Keep a Changelog](http://keepachangelog.com/en/1.0.0/).
|
|||
|
||||
- `row_log_interval` and `log_save_interval` are now based on training loop's `global_step` instead of epoch-internal batch index ([#3667](https://github.com/PyTorchLightning/pytorch-lightning/pull/3667))
|
||||
|
||||
- Swap `torch.load` for `fsspec` load in DDP spawn backend ([#3787](https://github.com/PyTorchLightning/pytorch-lightning/pull/3787))
|
||||
|
||||
### Deprecated
|
||||
|
||||
|
||||
|
|
|
@ -22,7 +22,7 @@ import torch.multiprocessing as mp
|
|||
from pytorch_lightning import _logger as log
|
||||
from pytorch_lightning.accelerators.base_backend import Accelerator
|
||||
from pytorch_lightning.utilities import AMPType
|
||||
from pytorch_lightning.utilities.cloud_io import atomic_save
|
||||
from pytorch_lightning.utilities.cloud_io import atomic_save, load as pl_load
|
||||
from pytorch_lightning.utilities.distributed import rank_zero_only, rank_zero_warn
|
||||
from pytorch_lightning.utilities.distributed import find_free_network_port
|
||||
from pytorch_lightning.distributed.dist import LightningDistributed
|
||||
|
@ -195,7 +195,7 @@ class DDPCPUSpawnBackend(Accelerator):
|
|||
|
||||
# load last weights
|
||||
if last_path is not None and not self.trainer.testing:
|
||||
ckpt = torch.load(last_path, map_location=lambda storage, loc: storage)
|
||||
ckpt = pl_load(last_path, map_location=lambda storage, loc: storage)
|
||||
model.load_state_dict(ckpt)
|
||||
|
||||
self.trainer.model = model
|
||||
|
|
|
@ -22,7 +22,7 @@ import torch.distributed as dist
|
|||
from pytorch_lightning import _logger as log
|
||||
from pytorch_lightning.accelerators.base_backend import Accelerator
|
||||
from pytorch_lightning.utilities import AMPType
|
||||
from pytorch_lightning.utilities.cloud_io import atomic_save
|
||||
from pytorch_lightning.utilities.cloud_io import atomic_save, load as pl_load
|
||||
from pytorch_lightning.utilities.distributed import rank_zero_only, rank_zero_warn
|
||||
from pytorch_lightning.utilities.seed import seed_everything
|
||||
from pytorch_lightning.distributed.dist import LightningDistributed
|
||||
|
@ -210,7 +210,7 @@ class DDPSpawnBackend(Accelerator):
|
|||
|
||||
# load last weights
|
||||
if last_path is not None and not self.trainer.testing:
|
||||
ckpt = torch.load(last_path, map_location=lambda storage, loc: storage)
|
||||
ckpt = pl_load(last_path, map_location=lambda storage, loc: storage)
|
||||
model.load_state_dict(ckpt)
|
||||
|
||||
self.trainer.model = model
|
||||
|
|
Loading…
Reference in New Issue