From 0ac7a8590b5a923e02191d0ad8324abf4ed1a1fd Mon Sep 17 00:00:00 2001 From: William Falcon Date: Thu, 18 Jul 2019 17:59:16 -0400 Subject: [PATCH] added slurm managed flag catch for non-slurm peeps --- pytorch_lightning/models/trainer.py | 1 + 1 file changed, 1 insertion(+) diff --git a/pytorch_lightning/models/trainer.py b/pytorch_lightning/models/trainer.py index 2f07083b3c..f7bb46c9fc 100644 --- a/pytorch_lightning/models/trainer.py +++ b/pytorch_lightning/models/trainer.py @@ -393,6 +393,7 @@ class Trainer(TrainerIO): You requested {nb_requested_gpus} GPUs but launched {nb_slurm_tasks} slurm tasks. We will launch {nb_requested_gpus} processes for you. We recommend you let slurm manage the processes by setting: --ntasks-per-node={nb_requested_gpus} + If you're not using SLURM, ignore this message! """ warnings.warn(msg) mp.spawn(self.ddp_train, nprocs=len(self.data_parallel_device_ids), args=(model, ))