added slurm managed flag catch for non-slurm peeps
This commit is contained in:
parent
6e12431e6b
commit
0ac7a8590b
|
@ -393,6 +393,7 @@ class Trainer(TrainerIO):
|
||||||
You requested {nb_requested_gpus} GPUs but launched {nb_slurm_tasks} slurm tasks.
|
You requested {nb_requested_gpus} GPUs but launched {nb_slurm_tasks} slurm tasks.
|
||||||
We will launch {nb_requested_gpus} processes for you.
|
We will launch {nb_requested_gpus} processes for you.
|
||||||
We recommend you let slurm manage the processes by setting: --ntasks-per-node={nb_requested_gpus}
|
We recommend you let slurm manage the processes by setting: --ntasks-per-node={nb_requested_gpus}
|
||||||
|
If you're not using SLURM, ignore this message!
|
||||||
"""
|
"""
|
||||||
warnings.warn(msg)
|
warnings.warn(msg)
|
||||||
mp.spawn(self.ddp_train, nprocs=len(self.data_parallel_device_ids), args=(model, ))
|
mp.spawn(self.ddp_train, nprocs=len(self.data_parallel_device_ids), args=(model, ))
|
||||||
|
|
Loading…
Reference in New Issue