* fix RunningMean
* changelog
* fix none
* Update supporters.py
just needed to multiply by zero for init
* Revert "Update supporters.py"
This reverts commit 7e0da6c6
* fix NaN
* formatting
Co-authored-by: William Falcon <waf2107@columbia.edu>
* added custom mnist without torchvision dep
* move files so it does not conflict with mnist gitignore
* mock torchvision for tests
* fix line too long
* fix line too long
* fix "module level import not at top of file" warning
* move mock imports to __init__.py
* simplify MNIST a lot and download directly the .pt files
* further simplify and clean up mnist
* revert import overrides
* make as before
* drop PIL requirement
* move mnist.py to datasets subfolder
* use logging instead of print
* choose same name as in torchvision
* remove torchvision and pillow also from yml file
* refactor if train
Co-Authored-By: Jirka Borovec <Borda@users.noreply.github.com>
* capitalized class attr
* moved mnist to models
* re-added datsets ignore
* better name for file variable
* Update mnist.py
* move dataset classes to datasets.py
* new line
* update
* update
* fix automerge
* move to base folder
* adapt testingmnist to new mnist base class
* remove temporal fix
* fix datatype
* remove old testingmnist
* readable
* fix import
* fix whitespace
* docstring
Co-Authored-By: Jirka Borovec <Borda@users.noreply.github.com>
* Update tests/base/datasets.py
Co-Authored-By: Jirka Borovec <Borda@users.noreply.github.com>
* changelog
* added types
* Update CHANGELOG.md
Co-Authored-By: Jirka Borovec <Borda@users.noreply.github.com>
* exist->isfile
Co-Authored-By: Jirka Borovec <Borda@users.noreply.github.com>
* index -> idx
* temporary fix for trains error
* better changelog message
Co-authored-by: Jirka Borovec <Borda@users.noreply.github.com>
* adding test
* adding test
* added base parity model
* added base parity model
* added parity test
* added parity test
* added parity test
* added parity test
* added parity test
* added parity test
* added parity test
* added parity test
* added parity test
* added parity test
* added parity test
* added parity test
* added parity test
* added parity test
* added parity test
* move parity to benchmark
* formatting
* fixed gradient acc sched
* move parity to benchmark
* formatting
* fixed gradient acc sched
* skip for CPU
* call last
Co-authored-by: J. Borovec <jirka.borovec@seznam.cz>
* Updated Sequencial Data docs
* Sequntial Data section now contains info on using IterableDatasets
* * Undid reformatting of bullet points
* * added information about val_check_interval
Co-authored-by: Donal Byrne <Donal.Byrne@xperi.com>
* Fix requirement-extra use released Trains package
* Update README.md add Trains and links to the external Visualization section
Co-authored-by: Jirka Borovec <Borda@users.noreply.github.com>
* `add_argparse_args` method fixed (argument types added)
* CHANGELOG.md upd
* autopep8 fixes
* --gpus=0 removed from test (for ci tests)
* typo fixed
* reduce on plateau scheduler fixed
* Trainer cli related tests moved to test_trainer_cli.py
* refactored: get_init_arguments_and_types is a public classmethod of the Trainer now
* test_get_init_arguments_and_types added
* autopep8 fixes
* Trainer cli related tests moved to test_trainer_cli.py
* refactored: get_init_arguments_and_types is a public classmethod of the Trainer now
* test_get_init_arguments_and_types added
* autopep8 fixes
* Trainer cli related tests moved to test_trainer_cli.py
* refactored: get_init_arguments_and_types is a public classmethod of the Trainer now
* test_get_init_arguments_and_types added
* autopep8 fixes
* Trainer cli related tests moved to test_trainer_cli.py
* test_get_init_arguments_and_types added
* autopep8 fixes
* Apply suggestions from code review
* cosmetics
* cosmetics
* Update pytorch_lightning/trainer/trainer.py
Co-Authored-By: Jirka Borovec <Borda@users.noreply.github.com>
* `Trainer.get_init_arguments_and_types` now returns arg types wrapped in tuples (not in sets)
* deprecated args are now ignored in argparser
* get_deprecated_arg_names small refactor
* get_deprecated_arg_names bug fixed
* Trainer cli related tests moved to test_trainer_cli.py
* refactored: get_init_arguments_and_types is a public classmethod of the Trainer now
* test_get_init_arguments_and_types added
* autopep8 fixes
* Trainer cli related tests moved to test_trainer_cli.py
* autopep8 fixes
* Trainer cli related tests moved to test_trainer_cli.py
* Trainer cli related tests moved to test_trainer_cli.py
* test_get_init_arguments_and_types added
* autopep8 fixes
* autopep8 fixes
* Apply suggestions from code review
* cosmetics
* cosmetics
* Update pytorch_lightning/trainer/trainer.py
Co-Authored-By: Jirka Borovec <Borda@users.noreply.github.com>
* `Trainer.get_init_arguments_and_types` now returns arg types wrapped in tuples (not in sets)
* deprecated args are now ignored in argparser
* get_deprecated_arg_names small refactor
* get_deprecated_arg_names bug fixed
* Update pytorch_lightning/trainer/trainer.py
Co-Authored-By: Joe Davison <joe@huggingface.co>
* Update pytorch_lightning/trainer/trainer.py
Co-Authored-By: Joe Davison <joe@huggingface.co>
Co-authored-by: Jirka Borovec <Borda@users.noreply.github.com>
Co-authored-by: Joe Davison <joe@huggingface.co>
Co-authored-by: William Falcon <waf2107@columbia.edu>
* pylint
* model API
* update test
* formatting
* disable logger
* fix checking overwrite
* fix test
* typo
* deprecated model
* fix for DDP
* drop Flake8 in GH actions
* Update pytorch_lightning/trainer/evaluation_loop.py
* fix imports
Co-authored-by: Nic Eggert <nic@eggert.io>
* check for nan values
* test nan detection on loss
* sys.exit
* whitespace
* detect nan and inf values in loss and params
* update
* added documentation
* moved detect nan to training loop, remove flag for print
* blank line
* test
* rename
* deprecate print_nan_grads
* deprecated print_nan_grads
* remove unused imports
* update changelog
* fix line too long
* correct deprecated version
Co-Authored-By: Jirka Borovec <Borda@users.noreply.github.com>
* raise exception instead of sysexit
Co-Authored-By: Jirka Borovec <Borda@users.noreply.github.com>
* raise exception instead of sysexit
Co-Authored-By: Jirka Borovec <Borda@users.noreply.github.com>
* Update pytorch_lightning/trainer/training_tricks.py
Co-Authored-By: Jirka Borovec <Borda@users.noreply.github.com>
* Update pytorch_lightning/trainer/training_tricks.py
Co-Authored-By: Jirka Borovec <Borda@users.noreply.github.com>
* fix test
Co-authored-by: Jirka Borovec <Borda@users.noreply.github.com>
* Update docs so users know the desired manner of configuring learning rate schedulers.
* update list
* as note
Co-authored-by: Jirka Borovec <Borda@users.noreply.github.com>
* check if hparams_type exists in checkpoint dictionary for backward compatibility
* concisely maintain backward compatibility for hparams type
* Bug fix in checkpoint loading (#1132)