mirror of https://github.com/explosion/spaCy.git
Update transfomers install docs
* Recommend installing PyTorch separately * Add instructions for `sentencepiece`
This commit is contained in:
parent
fcce3600ed
commit
e3e87e7275
|
@ -204,14 +204,25 @@ drop-in replacements that let you achieve **higher accuracy** in exchange for
|
||||||
> downloaded: 3GB CUDA runtime, 800MB PyTorch, 400MB CuPy, 500MB weights, 200MB
|
> downloaded: 3GB CUDA runtime, 800MB PyTorch, 400MB CuPy, 500MB weights, 200MB
|
||||||
> spaCy and dependencies.
|
> spaCy and dependencies.
|
||||||
|
|
||||||
Once you have CUDA installed, you'll need to install two pip packages,
|
Once you have CUDA installed, we recommend installing PyTorch separately
|
||||||
[`cupy`](https://docs.cupy.dev/en/stable/install.html) and
|
following the
|
||||||
[`spacy-transformers`](https://github.com/explosion/spacy-transformers). `cupy`
|
[PyTorch installation guidelines](https://pytorch.org/get-started/locally/) for
|
||||||
is just like `numpy`, but for GPU. The best way to install it is to choose a
|
your package manager and CUDA version. If you skip this step, pip will install
|
||||||
wheel that matches the version of CUDA you're using. You may also need to set
|
PyTorch as a dependency below, but it may not find the best version for your
|
||||||
the `CUDA_PATH` environment variable if your CUDA runtime is installed in a
|
setup.
|
||||||
non-standard location. Putting it all together, if you had installed CUDA 10.2
|
|
||||||
in `/opt/nvidia/cuda`, you would run:
|
```bash
|
||||||
|
### Example: Install PyTorch 1.7.1 for CUDA 10.1 with pip
|
||||||
|
$ pip install torch==1.7.1+cu101 torchvision==0.8.2+cu101 torchaudio==0.7.2 -f https://download.pytorch.org/whl/torch_stable.html
|
||||||
|
```
|
||||||
|
|
||||||
|
Next, install spaCy with the extras for your CUDA version and transformers. The
|
||||||
|
CUDA extra (e.g., `cuda92`, `cuda102`, `cuda111`) installs the correct version
|
||||||
|
of [`cupy`](https://docs.cupy.dev/en/stable/install.html#installing-cupy), which
|
||||||
|
is just like `numpy`, but for GPU. You may also need to set the `CUDA_PATH`
|
||||||
|
environment variable if your CUDA runtime is installed in a non-standard
|
||||||
|
location. Putting it all together, if you had installed CUDA 10.2 in
|
||||||
|
`/opt/nvidia/cuda`, you would run:
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
### Installation with CUDA
|
### Installation with CUDA
|
||||||
|
@ -219,6 +230,16 @@ $ export CUDA_PATH="/opt/nvidia/cuda"
|
||||||
$ pip install -U %%SPACY_PKG_NAME[cuda102,transformers]%%SPACY_PKG_FLAGS
|
$ pip install -U %%SPACY_PKG_NAME[cuda102,transformers]%%SPACY_PKG_FLAGS
|
||||||
```
|
```
|
||||||
|
|
||||||
|
For [`transformers`](https://huggingface.co/transformers/) v4.0.0+ and models
|
||||||
|
that require [`SentencePiece`](https://github.com/google/sentencepiece) (e.g.,
|
||||||
|
ALBERT, CamemBERT, XLNet, Marian, and T5), install the additional dependencies
|
||||||
|
with:
|
||||||
|
|
||||||
|
```bash
|
||||||
|
### Install sentencepiece
|
||||||
|
$ pip install transformers[sentencepiece]
|
||||||
|
```
|
||||||
|
|
||||||
### Runtime usage {#transformers-runtime}
|
### Runtime usage {#transformers-runtime}
|
||||||
|
|
||||||
Transformer models can be used as **drop-in replacements** for other types of
|
Transformer models can be used as **drop-in replacements** for other types of
|
||||||
|
|
Loading…
Reference in New Issue