lightning/pytorch_lightning/core
Lezwon Castelino 69833dad5b
Added check to verify xla device is TPU (#3274)
* tpu device check

* replaced with xmp spawn

* Revert "replaced with xmp spawn"

This reverts commit 6835380f

* replaced all instances of XLA_AVAILABLE

* moved inner_f to global scope

* made refactors

* added changelog

* added TPU_AVAILABLE variable

* fix codefactor issues

* removed form trainer and early stopping

* add TORCHXLA_AVAILABLE check

* added tests

* refactoring

* Update pytorch_lightning/utilities/xla_device_utils.py

Co-authored-by: Adrian Wälchli <aedu.waelchli@gmail.com>

* updated function names

* fixed bug

* updated CHANGELOG.md

* added todo

* added type hints

* isort and black

Co-authored-by: Adrian Wälchli <aedu.waelchli@gmail.com>
Co-authored-by: William Falcon <waf2107@columbia.edu>
2020-10-06 19:54:37 +02:00
..
__init__.py document lightiningmodule better (#2920) 2020-08-11 19:39:43 -04:00
datamodule.py Add trainer attribute to datamodule (#3749) 2020-10-01 00:41:19 +05:30
decorators.py added copyright notices (#3062) 2020-08-19 22:03:22 -04:00
grads.py added copyright notices (#3062) 2020-08-19 22:03:22 -04:00
hooks.py Fixes #2551 (#3858) 2020-10-04 23:02:35 -04:00
lightning.py Added check to verify xla device is TPU (#3274) 2020-10-06 19:54:37 +02:00
memory.py Refactor GPUStatsMonitor to improve training speed (#3257) 2020-09-04 06:02:16 -04:00
saving.py Fix for `load_from_checkpoint` (#2776) 2020-10-05 12:44:23 -04:00
step_result.py Fixes #3668, #3887 as a bonus (#3888) 2020-10-05 21:30:41 -04:00