From 55b69f9fc5d3bda1575044fcb374713080cd854c Mon Sep 17 00:00:00 2001 From: William Falcon Date: Wed, 3 Jul 2019 16:22:43 -0400 Subject: [PATCH] added single node distdataparallel --- pytorch_lightning/models/trainer.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/pytorch_lightning/models/trainer.py b/pytorch_lightning/models/trainer.py index e4bde254a0..d3b62c4568 100644 --- a/pytorch_lightning/models/trainer.py +++ b/pytorch_lightning/models/trainer.py @@ -281,7 +281,7 @@ class Trainer(TrainerIO): if self.on_gpu: rank = 0 self.experiment = self.experiment.get_meta_copy() - mp.spawn(self.__dp_train, nprocs=len(self.data_parallel_device_ids), args=(self, )) + mp.spawn(self.dp_train, nprocs=len(self.data_parallel_device_ids), args=(self, )) else: self.__run_pretrain_routine(model) @@ -291,7 +291,7 @@ class Trainer(TrainerIO): # del state['experiment'] return state - def __dp_train(self, gpu_nb, proc_rank, model): + def dp_train(self, gpu_nb, proc_rank, model): """ Entry point into a DP thread :param gpu_nb: