18 lines
662 B
Markdown
18 lines
662 B
Markdown
## Transformers
|
|
|
|
This example contains a simple training loop for next-word prediction with a [Transformer model](https://arxiv.org/abs/1706.03762) on a subset of the [WikiText2](https://www.salesforce.com/products/einstein/ai-research/the-wikitext-dependency-language-modeling-dataset/) dataset.
|
|
It is a simplified version of the [official PyTorch example](https://github.com/pytorch/examples/tree/main/word_language_model).
|
|
|
|
### Train with Fabric
|
|
|
|
```bash
|
|
# CPU
|
|
lightning run model --accelerator=cpu train.py
|
|
|
|
# GPU (CUDA or M1 Mac)
|
|
lightning run model --accelerator=gpu train.py
|
|
|
|
# Multiple GPUs
|
|
lightning run model --accelerator=gpu --devices=4 train.py
|
|
```
|