lightning/tests
Jirka Borovec b3e9dff32d
rename callback FineTune arg `round` (#9711)
* rename CB Tune arg round

Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
2021-10-06 09:39:36 +01:00
..
accelerators Enable auto parameters tying for TPUs (#9525) 2021-10-06 10:16:44 +02:00
base Keep global step update in the loop (#8856) 2021-09-14 19:21:39 +05:30
callbacks rename callback FineTune arg `round` (#9711) 2021-10-06 09:39:36 +01:00
checkpointing Add support for `torch.use_deterministic_algorithms` (#9121) 2021-09-30 04:40:09 +00:00
core Fix `ResultCollection._get_cache` with multielement tensors (#9582) 2021-09-22 14:03:20 +01:00
deprecated_api Enable auto parameters tying for TPUs (#9525) 2021-10-06 10:16:44 +02:00
helpers Remove `_NATIVE_AMP_AVAILABLE` checks (#9747) 2021-09-29 15:34:26 +02:00
loggers Add `enable_progress_bar` to Trainer constructor (#9664) 2021-09-24 22:53:31 -07:00
loops [Feat] Add auto_restart for fault tolerant training (#9722) 2021-10-01 16:37:17 +00:00
models Fix missing arguments when saving hyperparams from parent class only (#9800) 2021-10-06 08:32:29 +01:00
overrides Add support for `torch.use_deterministic_algorithms` (#9121) 2021-09-30 04:40:09 +00:00
plugins Fix broken `test_cpu_amp_precision_context_manager` (#9809) 2021-10-04 12:14:13 +00:00
profiler Remove unnecessary `pytest.param` usage (#9760) 2021-09-30 02:42:11 +00:00
trainer Add warnings regarding unsupported keys in optim config and OneCycleLR (#9666) 2021-10-04 08:25:05 +00:00
tuner Fix some flaky tests in tuner/lr_finder (#9766) 2021-10-01 11:15:16 +05:30
utilities Enable auto parameters tying for TPUs (#9525) 2021-10-06 10:16:44 +02:00
README.md CI: add mdformat (#8673) 2021-08-03 18:19:09 +00:00
__init__.py Replace `yapf` with `black` (#7783) 2021-07-26 13:37:35 +02:00
conftest.py Add support for `torch.use_deterministic_algorithms` (#9121) 2021-09-30 04:40:09 +00:00
mnode_tests.txt Mnodes (#5020) 2021-02-04 20:55:40 +01:00
special_tests.sh Skip reconciliate_processes if used within a cluster environment that creates processes externally (#9389) 2021-09-15 11:54:17 +01:00

README.md

PyTorch-Lightning Tests

Most PL tests train a full MNIST model under various trainer conditions (ddp, ddp2+amp, etc...). This provides testing for most combinations of important settings. The tests expect the model to perform to a reasonable degree of testing accuracy to pass.

Running tests

git clone https://github.com/PyTorchLightning/pytorch-lightning
cd pytorch-lightning

# install dev deps
pip install -r requirements/devel.txt

# run tests
py.test -v

To test models that require GPU make sure to run the above command on a GPU machine. The GPU machine must have at least 2 GPUs to run distributed tests.

Note that this setup will not run tests that require specific packages installed such as Horovod, FairScale, NVIDIA/apex, NVIDIA/DALI, etc. You can rely on our CI to make sure all these tests pass.

Running Coverage

Make sure to run coverage on a GPU machine with at least 2 GPUs and NVIDIA apex installed.

cd pytorch-lightning

# generate coverage (coverage is also installed as part of dev dependencies under requirements/devel.txt)
coverage run --source pytorch_lightning -m py.test pytorch_lightning tests examples -v

# print coverage stats
coverage report -m

# exporting results
coverage xml

Building test image

You can build it on your own, note it takes lots of time, be prepared.

git clone <git-repository>
docker image build -t pytorch_lightning:devel-torch1.9 -f dockers/cuda-extras/Dockerfile --build-arg TORCH_VERSION=1.9 .

To build other versions, select different Dockerfile.

docker image list
docker run --rm -it pytorch_lightning:devel-torch1.9 bash
docker image rm pytorch_lightning:devel-torch1.9