lightning/tests
Luca Antiga 87565cbca0
FP8 + FSDP2 + torch.compile examples for PyTorch Lightning and Fabric (#20440)
* Minimal transformer examples

* [pre-commit.ci] auto fixes from pre-commit.com hooks

for more information, see https://pre-commit.ci

* Add tests for compile after fsdp2/tp

* Add README's

* Add docs

* Rename folder, add cross-reference

* Fix link

* Newline after code-block directive

* Update section name

* Fix reference

* Half standalone tests batch size

* Fix integration tests

* [pre-commit.ci] auto fixes from pre-commit.com hooks

for more information, see https://pre-commit.ci

---------

Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
2024-11-26 15:29:40 +01:00
..
legacy Adding test for legacy checkpoint created with 2.3.3 (#20061) 2024-07-08 18:16:19 -04:00
parity_fabric Drop support for PyTorch 1.11 (#18691) 2023-10-04 20:30:44 +02:00
parity_pytorch precommit: ruff-format (#19434) 2024-02-15 13:39:17 -05:00
tests_fabric FP8 + FSDP2 + torch.compile examples for PyTorch Lightning and Fabric (#20440) 2024-11-26 15:29:40 +01:00
tests_pytorch FP8 + FSDP2 + torch.compile examples for PyTorch Lightning and Fabric (#20440) 2024-11-26 15:29:40 +01:00
README.md Fix standalone test file links (#19165) 2023-12-14 17:18:47 -05:00
run_standalone_tests.sh FP8 + FSDP2 + torch.compile examples for PyTorch Lightning and Fabric (#20440) 2024-11-26 15:29:40 +01:00

README.md

PyTorch-Lightning Tests

Most of the tests in PyTorch Lightning train a BoringModel under various trainer conditions (ddp, amp, etc...). Want to add a new test case and not sure how? Talk to us!

Running tests

Local: Testing your work locally will help you speed up the process since it allows you to focus on particular (failing) test-cases. To setup a local development environment, install both local and test dependencies:

# clone the repo
git clone https://github.com/Lightning-AI/lightning.git
cd lightning

# install required dependencies
export PACKAGE_NAME=pytorch
python -m pip install ".[dev, examples]"
# install pre-commit (optional)
python -m pip install pre-commit
pre-commit install

Additionally, for testing backward compatibility with older versions of PyTorch Lightning, you also need to download all saved version-checkpoints from the public AWS storage. Run the following script to get all saved version-checkpoints:

bash .actions/pull_legacy_checkpoints.sh

Note: These checkpoints are generated to set baselines for maintaining backward compatibility with legacy versions of PyTorch Lightning. Details of checkpoints for back-compatibility can be found here.

You can run the full test suite in your terminal via this make script:

make test

Note: if your computer does not have multi-GPU or TPU these tests are skipped.

GitHub Actions: For convenience, you can also use your own GHActions building which will be triggered with each commit. This is useful if you do not test against all required dependency versions.

Docker: Another option is to utilize the pytorch lightning cuda base docker image. You can then run:

python -m pytest src/lightning/pytorch tests/tests_pytorch -v

You can also run a single test as follows:

python -m pytest -v tests/tests_pytorch/trainer/test_trainer_cli.py::test_default_args

Conditional Tests

To test models that require GPU make sure to run the above command on a GPU machine. The GPU machine must have at least 2 GPUs to run distributed tests.

Note that this setup will not run tests that require specific packages installed You can rely on our CI to make sure all these tests pass.

Standalone Tests

There are certain standalone tests, which you can run using:

./tests/run_standalone_tests.sh tests/tests_pytorch/trainer/
# or run a specific test
./tests/run_standalone_tests.sh -k test_multi_gpu_model_ddp

Running Coverage

Make sure to run coverage on a GPU machine with at least 2 GPUs.

# generate coverage (coverage is also installed as part of dev dependencies)
coverage run --source src/lightning/pytorch -m pytest src/lightning/pytorch tests/tests_pytorch -v

# print coverage stats
coverage report -m

# exporting results
coverage xml