lightning/examples/fabric/language_model/README.md

18 lines
662 B
Markdown

## Transformers
This example contains a simple training loop for next-word prediction with a [Transformer model](https://arxiv.org/abs/1706.03762) on a subset of the [WikiText2](https://www.salesforce.com/products/einstein/ai-research/the-wikitext-dependency-language-modeling-dataset/) dataset.
It is a simplified version of the [official PyTorch example](https://github.com/pytorch/examples/tree/main/word_language_model).
### Train with Fabric
```bash
# CPU
lightning run model --accelerator=cpu train.py
# GPU (CUDA or M1 Mac)
lightning run model --accelerator=gpu train.py
# Multiple GPUs
lightning run model --accelerator=gpu --devices=4 train.py
```