11 lines
487 B
ReStructuredText
11 lines
487 B
ReStructuredText
############
|
|
Accelerators
|
|
############
|
|
Accelerators connect a Lightning Trainer to arbitrary accelerators (CPUs, GPUs, TPUs, etc). Accelerators
|
|
also manage distributed accelerators (like DP, DDP, HPC cluster).
|
|
|
|
Accelerators can also be configured to run on arbitrary clusters using Plugins or to link up to arbitrary
|
|
computational strategies like 16-bit precision via AMP and Apex.
|
|
|
|
**For help setting up custom plugin/accelerator please reach out to us at support@pytorchlightning.ai**
|