cleaned up demos
This commit is contained in:
parent
9fc01e3fd3
commit
4d3a8c25d2
|
@ -0,0 +1,11 @@
|
||||||
|
# Examples
|
||||||
|
This folder has 3 sections:
|
||||||
|
|
||||||
|
### Domain templates
|
||||||
|
These are templates to show common approaches such as GANs and RL.
|
||||||
|
|
||||||
|
### Basic examples
|
||||||
|
These show the most common use of Lightning for either CPU or GPU training.
|
||||||
|
|
||||||
|
### Multi-node examples
|
||||||
|
These show how to run jobs on a GPU cluster using lightning.
|
|
@ -0,0 +1,40 @@
|
||||||
|
# Basic Examples
|
||||||
|
Use these examples to test how lightning works.
|
||||||
|
|
||||||
|
### Test on CPU
|
||||||
|
```bash
|
||||||
|
python cpu_template.py
|
||||||
|
```
|
||||||
|
|
||||||
|
### Test on GPUs
|
||||||
|
This demo can train on a single GPU, multiple GPUs or multiple nodes by
|
||||||
|
passing in different flags.
|
||||||
|
|
||||||
|
##### Train on a single GPU
|
||||||
|
```bash
|
||||||
|
python gpu_template.py --gpus 1
|
||||||
|
```
|
||||||
|
|
||||||
|
---
|
||||||
|
##### DataParallel (dp)
|
||||||
|
Train on multiple GPUs using DataParallel.
|
||||||
|
|
||||||
|
```bash
|
||||||
|
python gpu_template.py --gpus 2 --distributed_backend dp
|
||||||
|
```
|
||||||
|
|
||||||
|
##### DistributedDataParallel (ddp)
|
||||||
|
|
||||||
|
Train on multiple GPUs using DistributedDataParallel
|
||||||
|
```bash
|
||||||
|
python gpu_template.py --gpus 2 --distributed_backend ddp
|
||||||
|
```
|
||||||
|
|
||||||
|
##### DistributedDataParallel+DP (ddp2)
|
||||||
|
|
||||||
|
Train on multiple GPUs using DistributedDataParallel + dataparallel.
|
||||||
|
On a single node, uses all GPUs for 1 model. Then shares gradient information
|
||||||
|
across nodes.
|
||||||
|
```bash
|
||||||
|
python gpu_template.py --gpus 2 --distributed_backend ddp2
|
||||||
|
```
|
Loading…
Reference in New Issue