lightning/pytorch_lightning/trainer
Asaf Manor 09668df726
Update optimizers.py (#1383)
2020-04-07 09:09:23 -04:00
..
__init__.py add trainer attribute to denote if interrupted (#1368) 2020-04-05 11:12:41 -04:00
auto_mix_precision.py Set precision=16 when use_amp is passed as True (#1145) 2020-04-06 08:13:24 -04:00
callback_config.py proper checkpoint implementation (#1043) 2020-03-04 23:02:19 -05:00
callback_hook.py cleaning imports (#1032) 2020-03-12 12:41:37 -04:00
data_loading.py Add warning for few workers (#1378) 2020-04-05 11:07:16 -04:00
deprecated_api.py Simplify progress bar args (#1108) 2020-04-03 00:53:00 +02:00
distrib_data_parallel.py load_spawn_weights only in proc rank 0 (#1385) 2020-04-06 10:17:16 -04:00
distrib_parts.py Set precision=16 when use_amp is passed as True (#1145) 2020-04-06 08:13:24 -04:00
evaluation_loop.py Fix for incorrect run on the validation set with overwritten validation_epoch_end and test_end (#1353) 2020-04-03 09:25:32 -04:00
ignored_warnings.py
logging.py cleaning imports (#1032) 2020-03-12 12:41:37 -04:00
model_hooks.py quick patch __code__ (#1352) 2020-04-03 08:40:02 -04:00
optimizers.py Update optimizers.py (#1383) 2020-04-07 09:09:23 -04:00
supporters.py Fix unimplemented type() on TPU (#1396) 2020-04-06 20:29:55 -04:00
trainer.py Set precision=16 when use_amp is passed as True (#1145) 2020-04-06 08:13:24 -04:00
training_io.py Checkpointing interval (#1272) 2020-03-30 18:37:02 -04:00
training_loop.py Set precision=16 when use_amp is passed as True (#1145) 2020-04-06 08:13:24 -04:00
training_tricks.py nan detection and intervention (#1097) 2020-03-19 09:24:45 -04:00