From 162b9f4f27f81a44258b1be88e1a7aa82bd4a609 Mon Sep 17 00:00:00 2001 From: William Falcon Date: Thu, 18 Jul 2019 11:15:21 -0400 Subject: [PATCH] set dp as default backend --- pytorch_lightning/models/trainer.py | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/pytorch_lightning/models/trainer.py b/pytorch_lightning/models/trainer.py index 57911e09c6..83ea572477 100644 --- a/pytorch_lightning/models/trainer.py +++ b/pytorch_lightning/models/trainer.py @@ -50,7 +50,7 @@ class Trainer(TrainerIO): val_check_interval=0.95, log_save_interval=100, add_log_row_interval=10, lr_scheduler_milestones=None, - use_distributed_dataparallel=True, + distributed_backend='dp', use_amp=False, print_nan_grads=False, print_weights_summary=True, @@ -110,8 +110,8 @@ class Trainer(TrainerIO): # single GPU will also use DP with devices=[0] have_gpus = self.data_parallel_device_ids is not None and len(self.data_parallel_device_ids) > 0 if have_gpus: - self.use_ddp = use_distributed_dataparallel - self.use_dp = not self.use_ddp + self.use_dp = distributed_backend == 'dp' + self.use_ddp = distributed_backend == 'ddp' # process info self.proc_rank = 0