2018-12-02 03:22:23 +00:00
|
|
|
<a href="https://explosion.ai"><img src="https://explosion.ai/assets/img/logo.svg" width="125" height="125" align="right" /></a>
|
|
|
|
|
2018-11-26 21:04:35 +00:00
|
|
|
# spaCy: Industrial-strength NLP
|
|
|
|
|
|
|
|
spaCy is a library for advanced Natural Language Processing in Python and
|
2019-10-02 08:37:04 +00:00
|
|
|
Cython. It's built on the very latest research, and was designed from day one to
|
|
|
|
be used in real products. spaCy comes with
|
2019-10-02 08:37:39 +00:00
|
|
|
[pretrained statistical models](https://spacy.io/models) and word vectors, and
|
2019-08-07 11:38:59 +00:00
|
|
|
currently supports tokenization for **50+ languages**. It features
|
2019-04-19 13:23:08 +00:00
|
|
|
state-of-the-art speed, convolutional **neural network models** for tagging,
|
|
|
|
parsing and **named entity recognition** and easy **deep learning** integration.
|
|
|
|
It's commercial open-source software, released under the MIT license.
|
2018-11-26 21:04:35 +00:00
|
|
|
|
2019-10-02 14:47:10 +00:00
|
|
|
💫 **Version 2.2 out now!**
|
|
|
|
[Check out the release notes here.](https://github.com/explosion/spaCy/releases)
|
|
|
|
|
2019-10-02 17:15:22 +00:00
|
|
|
[![Azure Pipelines](<https://img.shields.io/azure-devops/build/explosion-ai/public/8/master.svg?logo=azure-pipelines&style=flat-square&label=build+(3.x)>)](https://dev.azure.com/explosion-ai/public/_build?definitionId=8)
|
2019-10-02 14:47:10 +00:00
|
|
|
[![Current Release Version](https://img.shields.io/github/release/explosion/spacy.svg?style=flat-square&logo=github)](https://github.com/explosion/spaCy/releases)
|
|
|
|
[![pypi Version](https://img.shields.io/pypi/v/spacy.svg?style=flat-square&logo=pypi&logoColor=white)](https://pypi.org/project/spacy/)
|
|
|
|
[![conda Version](https://img.shields.io/conda/vn/conda-forge/spacy.svg?style=flat-square&logo=conda-forge&logoColor=white)](https://anaconda.org/conda-forge/spacy)
|
2018-11-26 21:04:35 +00:00
|
|
|
[![Python wheels](https://img.shields.io/badge/wheels-%E2%9C%93-4c1.svg?longCache=true&style=flat-square&logo=python&logoColor=white)](https://github.com/explosion/wheelwright/releases)
|
2019-10-02 14:47:10 +00:00
|
|
|
[![PyPi downloads](https://img.shields.io/pypi/dm/spacy?style=flat-square&logo=pypi&logoColor=white)](https://pypi.org/project/spacy/)
|
|
|
|
[![Conda downloads](https://img.shields.io/conda/dn/conda-forge/spacy?style=flat-square&logo=conda-forge&logoColor=white)](https://anaconda.org/conda-forge/spacy)
|
2019-10-02 16:49:55 +00:00
|
|
|
[![Model downloads](https://img.shields.io/github/downloads/explosion/spacy-models/total?style=flat-square&label=model+downloads)](https://github.com/explosion/spacy-models/releases)
|
2019-02-24 12:12:13 +00:00
|
|
|
[![Code style: black](https://img.shields.io/badge/code%20style-black-000000.svg?style=flat-square)](https://github.com/ambv/black)
|
2018-11-26 21:04:35 +00:00
|
|
|
[![spaCy on Twitter](https://img.shields.io/twitter/follow/spacy_io.svg?style=social&label=Follow)](https://twitter.com/spacy_io)
|
|
|
|
|
|
|
|
## 📖 Documentation
|
|
|
|
|
2019-03-18 15:07:26 +00:00
|
|
|
| Documentation | |
|
|
|
|
| --------------- | -------------------------------------------------------------- |
|
|
|
|
| [spaCy 101] | New to spaCy? Here's everything you need to know! |
|
|
|
|
| [Usage Guides] | How to use spaCy and its features. |
|
2019-10-02 14:47:10 +00:00
|
|
|
| [New in v2.2] | New features, backwards incompatibilities and migration guide. |
|
2019-03-18 15:07:26 +00:00
|
|
|
| [API Reference] | The detailed reference for spaCy's API. |
|
|
|
|
| [Models] | Download statistical language models for spaCy. |
|
|
|
|
| [Universe] | Libraries, extensions, demos, books and courses. |
|
|
|
|
| [Changelog] | Changes and version history. |
|
|
|
|
| [Contribute] | How to contribute to the spaCy project and code base. |
|
2019-02-24 12:12:13 +00:00
|
|
|
|
|
|
|
[spacy 101]: https://spacy.io/usage/spacy-101
|
2019-10-02 14:47:10 +00:00
|
|
|
[new in v2.2]: https://spacy.io/usage/v2-2
|
2019-02-24 12:12:13 +00:00
|
|
|
[usage guides]: https://spacy.io/usage/
|
|
|
|
[api reference]: https://spacy.io/api/
|
|
|
|
[models]: https://spacy.io/models
|
|
|
|
[universe]: https://spacy.io/universe
|
2019-04-02 15:36:22 +00:00
|
|
|
[changelog]: https://spacy.io/usage#changelog
|
2019-02-24 12:12:13 +00:00
|
|
|
[contribute]: https://github.com/explosion/spaCy/blob/master/CONTRIBUTING.md
|
2018-11-26 21:04:35 +00:00
|
|
|
|
|
|
|
## 💬 Where to ask questions
|
|
|
|
|
2019-10-02 08:37:04 +00:00
|
|
|
The spaCy project is maintained by [@honnibal](https://github.com/honnibal) and
|
|
|
|
[@ines](https://github.com/ines), along with core contributors
|
2019-09-14 10:57:45 +00:00
|
|
|
[@svlandeg](https://github.com/svlandeg) and
|
|
|
|
[@adrianeboyd](https://github.com/adrianeboyd). Please understand that we won't
|
|
|
|
be able to provide individual support via email. We also believe that help is
|
2019-10-02 08:37:04 +00:00
|
|
|
much more valuable if it's shared publicly, so that more people can benefit from
|
|
|
|
it.
|
2018-11-26 21:04:35 +00:00
|
|
|
|
2019-02-24 12:12:13 +00:00
|
|
|
| Type | Platforms |
|
|
|
|
| ------------------------ | ------------------------------------------------------ |
|
2019-02-24 16:40:01 +00:00
|
|
|
| 🚨 **Bug Reports** | [GitHub Issue Tracker] |
|
2019-02-24 12:12:13 +00:00
|
|
|
| 🎁 **Feature Requests** | [GitHub Issue Tracker] |
|
2019-02-24 16:40:01 +00:00
|
|
|
| 👩💻 **Usage Questions** | [Stack Overflow] · [Gitter Chat] · [Reddit User Group] |
|
2019-02-24 12:12:13 +00:00
|
|
|
| 🗯 **General Discussion** | [Gitter Chat] · [Reddit User Group] |
|
2018-11-26 21:04:35 +00:00
|
|
|
|
2019-02-24 12:12:13 +00:00
|
|
|
[github issue tracker]: https://github.com/explosion/spaCy/issues
|
2019-04-02 15:36:22 +00:00
|
|
|
[stack overflow]: https://stackoverflow.com/questions/tagged/spacy
|
2019-02-24 12:12:13 +00:00
|
|
|
[gitter chat]: https://gitter.im/explosion/spaCy
|
|
|
|
[reddit user group]: https://www.reddit.com/r/spacynlp
|
2018-11-26 21:04:35 +00:00
|
|
|
|
|
|
|
## Features
|
|
|
|
|
2019-11-14 23:33:44 +00:00
|
|
|
- Non-destructive **tokenization**
|
|
|
|
- **Named entity** recognition
|
|
|
|
- Support for **50+ languages**
|
|
|
|
- pretrained [statistical models](https://spacy.io/models) and word vectors
|
|
|
|
- State-of-the-art speed
|
|
|
|
- Easy **deep learning** integration
|
|
|
|
- Part-of-speech tagging
|
|
|
|
- Labelled dependency parsing
|
|
|
|
- Syntax-driven sentence segmentation
|
|
|
|
- Built in **visualizers** for syntax and NER
|
|
|
|
- Convenient string-to-hash mapping
|
|
|
|
- Export to numpy data arrays
|
|
|
|
- Efficient binary serialization
|
|
|
|
- Easy **model packaging** and deployment
|
|
|
|
- Robust, rigorously evaluated accuracy
|
2019-02-24 12:12:13 +00:00
|
|
|
|
|
|
|
📖 **For more details, see the
|
2018-11-26 21:04:35 +00:00
|
|
|
[facts, figures and benchmarks](https://spacy.io/usage/facts-figures).**
|
|
|
|
|
|
|
|
## Install spaCy
|
|
|
|
|
|
|
|
For detailed installation instructions, see the
|
|
|
|
[documentation](https://spacy.io/usage).
|
|
|
|
|
2019-11-14 23:33:44 +00:00
|
|
|
- **Operating system**: macOS / OS X · Linux · Windows (Cygwin, MinGW, Visual
|
|
|
|
Studio)
|
2019-12-22 00:53:56 +00:00
|
|
|
- **Python version**: Python 3.6+ (only 64 bit)
|
2019-11-14 23:33:44 +00:00
|
|
|
- **Package managers**: [pip] · [conda] (via `conda-forge`)
|
2018-11-26 21:04:35 +00:00
|
|
|
|
2019-04-02 15:36:22 +00:00
|
|
|
[pip]: https://pypi.org/project/spacy/
|
2018-11-26 21:04:35 +00:00
|
|
|
[conda]: https://anaconda.org/conda-forge/spacy
|
|
|
|
|
Generalize handling of tokenizer special cases (#4259)
* Generalize handling of tokenizer special cases
Handle tokenizer special cases more generally by using the Matcher
internally to match special cases after the affix/token_match
tokenization is complete.
Instead of only matching special cases while processing balanced or
nearly balanced prefixes and suffixes, this recognizes special cases in
a wider range of contexts:
* Allows arbitrary numbers of prefixes/affixes around special cases
* Allows special cases separated by infixes
Existing tests/settings that couldn't be preserved as before:
* The emoticon '")' is no longer a supported special case
* The emoticon ':)' in "example:)" is a false positive again
When merged with #4258 (or the relevant cache bugfix), the affix and
token_match properties should be modified to flush and reload all
special cases to use the updated internal tokenization with the Matcher.
* Remove accidentally added test case
* Really remove accidentally added test
* Reload special cases when necessary
Reload special cases when affixes or token_match are modified. Skip
reloading during initialization.
* Update error code number
* Fix offset and whitespace in Matcher special cases
* Fix offset bugs when merging and splitting tokens
* Set final whitespace on final token in inserted special case
* Improve cache flushing in tokenizer
* Separate cache and specials memory (temporarily)
* Flush cache when adding special cases
* Repeated `self._cache = PreshMap()` and `self._specials = PreshMap()`
are necessary due to this bug:
https://github.com/explosion/preshed/issues/21
* Remove reinitialized PreshMaps on cache flush
* Update UD bin scripts
* Update imports for `bin/`
* Add all currently supported languages
* Update subtok merger for new Matcher validation
* Modify blinded check to look at tokens instead of lemmas (for corpora
with tokens but not lemmas like Telugu)
* Use special Matcher only for cases with affixes
* Reinsert specials cache checks during normal tokenization for special
cases as much as possible
* Additionally include specials cache checks while splitting on infixes
* Since the special Matcher needs consistent affix-only tokenization
for the special cases themselves, introduce the argument
`with_special_cases` in order to do tokenization with or without
specials cache checks
* After normal tokenization, postprocess with special cases Matcher for
special cases containing affixes
* Replace PhraseMatcher with Aho-Corasick
Replace PhraseMatcher with the Aho-Corasick algorithm over numpy arrays
of the hash values for the relevant attribute. The implementation is
based on FlashText.
The speed should be similar to the previous PhraseMatcher. It is now
possible to easily remove match IDs and matches don't go missing with
large keyword lists / vocabularies.
Fixes #4308.
* Restore support for pickling
* Fix internal keyword add/remove for numpy arrays
* Add test for #4248, clean up test
* Improve efficiency of special cases handling
* Use PhraseMatcher instead of Matcher
* Improve efficiency of merging/splitting special cases in document
* Process merge/splits in one pass without repeated token shifting
* Merge in place if no splits
* Update error message number
* Remove UD script modifications
Only used for timing/testing, should be a separate PR
* Remove final traces of UD script modifications
* Update UD bin scripts
* Update imports for `bin/`
* Add all currently supported languages
* Update subtok merger for new Matcher validation
* Modify blinded check to look at tokens instead of lemmas (for corpora
with tokens but not lemmas like Telugu)
* Add missing loop for match ID set in search loop
* Remove cruft in matching loop for partial matches
There was a bit of unnecessary code left over from FlashText in the
matching loop to handle partial token matches, which we don't have with
PhraseMatcher.
* Replace dict trie with MapStruct trie
* Fix how match ID hash is stored/added
* Update fix for match ID vocab
* Switch from map_get_unless_missing to map_get
* Switch from numpy array to Token.get_struct_attr
Access token attributes directly in Doc instead of making a copy of the
relevant values in a numpy array.
Add unsatisfactory warning for hash collision with reserved terminal
hash key. (Ideally it would change the reserved terminal hash and redo
the whole trie, but for now, I'm hoping there won't be collisions.)
* Restructure imports to export find_matches
* Implement full remove()
Remove unnecessary trie paths and free unused maps.
Parallel to Matcher, raise KeyError when attempting to remove a match ID
that has not been added.
* Switch to PhraseMatcher.find_matches
* Switch to local cdef functions for span filtering
* Switch special case reload threshold to variable
Refer to variable instead of hard-coded threshold
* Move more of special case retokenize to cdef nogil
Move as much of the special case retokenization to nogil as possible.
* Rewrap sort as stdsort for OS X
* Rewrap stdsort with specific types
* Switch to qsort
* Fix merge
* Improve cmp functions
* Fix realloc
* Fix realloc again
* Initialize span struct while retokenizing
* Temporarily skip retokenizing
* Revert "Move more of special case retokenize to cdef nogil"
This reverts commit 0b7e52c797cd8ff1548f214bd4186ebb3a7ce8b1.
* Revert "Switch to qsort"
This reverts commit a98d71a942fc9bca531cf5eb05cf89fa88153b60.
* Fix specials check while caching
* Modify URL test with emoticons
The multiple suffix tests result in the emoticon `:>`, which is now
retokenized into one token as a special case after the suffixes are
split off.
* Refactor _apply_special_cases()
* Use cdef ints for span info used in multiple spots
* Modify _filter_special_spans() to prefer earlier
Parallel to #4414, modify _filter_special_spans() so that the earlier
span is preferred for overlapping spans of the same length.
* Replace MatchStruct with Entity
Replace MatchStruct with Entity since the existing Entity struct is
nearly identical.
* Replace Entity with more general SpanC
* Replace MatchStruct with SpanC
* Add error in debug-data if no dev docs are available (see #4575)
* Update azure-pipelines.yml
* Revert "Update azure-pipelines.yml"
This reverts commit ed1060cf59e5895b5fe92ad5b894fd1078ec4c49.
* Use latest wasabi
* Reorganise install_requires
* add dframcy to universe.json (#4580)
* Update universe.json [ci skip]
* Fix multiprocessing for as_tuples=True (#4582)
* Fix conllu script (#4579)
* force extensions to avoid clash between example scripts
* fix arg order and default file encoding
* add example config for conllu script
* newline
* move extension definitions to main function
* few more encodings fixes
* Add load_from_docbin example [ci skip]
TODO: upload the file somewhere
* Update README.md
* Add warnings about 3.8 (resolves #4593) [ci skip]
* Fixed typo: Added space between "recognize" and "various" (#4600)
* Fix DocBin.merge() example (#4599)
* Replace function registries with catalogue (#4584)
* Replace functions registries with catalogue
* Update __init__.py
* Fix test
* Revert unrelated flag [ci skip]
* Bugfix/dep matcher issue 4590 (#4601)
* add contributor agreement for prilopes
* add test for issue #4590
* fix on_match params for DependencyMacther (#4590)
* Minor updates to language example sentences (#4608)
* Add punctuation to Spanish example sentences
* Combine multilanguage examples for lang xx
* Add punctuation to nb examples
* Always realloc to a larger size
Avoid potential (unlikely) edge case and cymem error seen in #4604.
* Add error in debug-data if no dev docs are available (see #4575)
* Update debug-data for GoldCorpus / Example
* Ignore None label in misaligned NER data
2019-11-13 20:24:35 +00:00
|
|
|
> ⚠️ **Important note for Python 3.8:** We can't yet ship pre-compiled binary
|
|
|
|
> wheels for spaCy that work on Python 3.8, as we're still waiting for our CI
|
|
|
|
> providers and other tooling to support it. This means that in order to run
|
|
|
|
> spaCy on Python 3.8, you'll need [a compiler installed](#source) and compile
|
|
|
|
> the library and its Cython dependencies locally. If this is causing problems
|
|
|
|
> for you, the easiest solution is to **use Python 3.7** in the meantime.
|
|
|
|
|
2018-11-26 21:04:35 +00:00
|
|
|
### pip
|
|
|
|
|
2019-10-02 08:37:04 +00:00
|
|
|
Using pip, spaCy releases are available as source packages and binary wheels (as
|
|
|
|
of `v2.0.13`).
|
2018-11-26 21:04:35 +00:00
|
|
|
|
|
|
|
```bash
|
|
|
|
pip install spacy
|
|
|
|
```
|
|
|
|
|
2019-10-02 14:47:10 +00:00
|
|
|
To install additional data tables for lemmatization in **spaCy v2.2+** you can
|
|
|
|
run `pip install spacy[lookups]` or install
|
|
|
|
[`spacy-lookups-data`](https://github.com/explosion/spacy-lookups-data)
|
|
|
|
separately. The lookups package is needed to create blank models with
|
|
|
|
lemmatization data, and to lemmatize in languages that don't yet come with
|
|
|
|
pretrained models and aren't powered by third-party libraries.
|
|
|
|
|
2018-11-26 21:04:35 +00:00
|
|
|
When using pip it is generally recommended to install packages in a virtual
|
|
|
|
environment to avoid modifying system state:
|
|
|
|
|
|
|
|
```bash
|
|
|
|
python -m venv .env
|
|
|
|
source .env/bin/activate
|
|
|
|
pip install spacy
|
|
|
|
```
|
|
|
|
|
|
|
|
### conda
|
|
|
|
|
|
|
|
Thanks to our great community, we've finally re-added conda support. You can now
|
|
|
|
install spaCy via `conda-forge`:
|
|
|
|
|
|
|
|
```bash
|
2019-10-31 12:23:25 +00:00
|
|
|
conda install -c conda-forge spacy
|
2018-11-26 21:04:35 +00:00
|
|
|
```
|
|
|
|
|
2019-10-02 08:37:04 +00:00
|
|
|
For the feedstock including the build recipe and configuration, check out
|
|
|
|
[this repository](https://github.com/conda-forge/spacy-feedstock). Improvements
|
|
|
|
and pull requests to the recipe and setup are always appreciated.
|
2018-11-26 21:04:35 +00:00
|
|
|
|
|
|
|
### Updating spaCy
|
|
|
|
|
|
|
|
Some updates to spaCy may require downloading new statistical models. If you're
|
|
|
|
running spaCy v2.0 or higher, you can use the `validate` command to check if
|
|
|
|
your installed models are compatible and if not, print details on how to update
|
|
|
|
them:
|
|
|
|
|
|
|
|
```bash
|
|
|
|
pip install -U spacy
|
|
|
|
python -m spacy validate
|
|
|
|
```
|
|
|
|
|
|
|
|
If you've trained your own models, keep in mind that your training and runtime
|
|
|
|
inputs must match. After updating spaCy, we recommend **retraining your models**
|
|
|
|
with the new version.
|
|
|
|
|
2019-02-24 12:12:13 +00:00
|
|
|
📖 **For details on upgrading from spaCy 1.x to spaCy 2.x, see the
|
2018-11-26 21:04:35 +00:00
|
|
|
[migration guide](https://spacy.io/usage/v2#migrating).**
|
|
|
|
|
|
|
|
## Download models
|
|
|
|
|
2019-10-02 08:37:04 +00:00
|
|
|
As of v1.7.0, models for spaCy can be installed as **Python packages**. This
|
|
|
|
means that they're a component of your application, just like any other module.
|
|
|
|
Models can be installed using spaCy's `download` command, or manually by
|
|
|
|
pointing pip to a path or URL.
|
2018-11-26 21:04:35 +00:00
|
|
|
|
2019-02-24 12:12:13 +00:00
|
|
|
| Documentation | |
|
|
|
|
| ---------------------- | ------------------------------------------------------------- |
|
|
|
|
| [Available Models] | Detailed model descriptions, accuracy figures and benchmarks. |
|
|
|
|
| [Models Documentation] | Detailed usage instructions. |
|
2018-11-26 21:04:35 +00:00
|
|
|
|
2019-02-24 12:12:13 +00:00
|
|
|
[available models]: https://spacy.io/models
|
|
|
|
[models documentation]: https://spacy.io/docs/usage/models
|
2018-11-26 21:04:35 +00:00
|
|
|
|
|
|
|
```bash
|
2019-03-17 10:48:18 +00:00
|
|
|
# download best-matching version of specific model for your spaCy installation
|
|
|
|
python -m spacy download en_core_web_sm
|
|
|
|
|
2018-11-26 21:04:35 +00:00
|
|
|
# pip install .tar.gz archive from path or URL
|
2019-09-18 19:23:26 +00:00
|
|
|
pip install /Users/you/en_core_web_sm-2.2.0.tar.gz
|
|
|
|
pip install https://github.com/explosion/spacy-models/releases/download/en_core_web_sm-2.2.0/en_core_web_sm-2.2.0.tar.gz
|
2018-11-26 21:04:35 +00:00
|
|
|
```
|
|
|
|
|
|
|
|
### Loading and using models
|
|
|
|
|
2019-03-17 10:48:18 +00:00
|
|
|
To load a model, use `spacy.load()` with the model name, a shortcut link or a
|
|
|
|
path to the model data directory.
|
2018-11-26 21:04:35 +00:00
|
|
|
|
|
|
|
```python
|
|
|
|
import spacy
|
2019-03-17 10:48:18 +00:00
|
|
|
nlp = spacy.load("en_core_web_sm")
|
2019-10-18 16:47:59 +00:00
|
|
|
doc = nlp("This is a sentence.")
|
2018-11-26 21:04:35 +00:00
|
|
|
```
|
|
|
|
|
2019-03-17 10:48:18 +00:00
|
|
|
You can also `import` a model directly via its full name and then call its
|
|
|
|
`load()` method with no arguments.
|
2018-11-26 21:04:35 +00:00
|
|
|
|
|
|
|
```python
|
|
|
|
import spacy
|
|
|
|
import en_core_web_sm
|
|
|
|
|
|
|
|
nlp = en_core_web_sm.load()
|
2019-10-18 16:47:59 +00:00
|
|
|
doc = nlp("This is a sentence.")
|
2018-11-26 21:04:35 +00:00
|
|
|
```
|
|
|
|
|
|
|
|
📖 **For more info and examples, check out the
|
|
|
|
[models documentation](https://spacy.io/docs/usage/models).**
|
|
|
|
|
|
|
|
## Compile from source
|
|
|
|
|
|
|
|
The other way to install spaCy is to clone its
|
|
|
|
[GitHub repository](https://github.com/explosion/spaCy) and build it from
|
|
|
|
source. That is the common way if you want to make changes to the code base.
|
|
|
|
You'll need to make sure that you have a development environment consisting of a
|
|
|
|
Python distribution including header files, a compiler,
|
|
|
|
[pip](https://pip.pypa.io/en/latest/installing/),
|
2019-10-02 08:37:04 +00:00
|
|
|
[virtualenv](https://virtualenv.pypa.io/en/latest/) and
|
|
|
|
[git](https://git-scm.com) installed. The compiler part is the trickiest. How to
|
|
|
|
do that depends on your system. See notes on Ubuntu, OS X and Windows for
|
|
|
|
details.
|
2018-11-26 21:04:35 +00:00
|
|
|
|
|
|
|
```bash
|
|
|
|
# make sure you are using the latest pip
|
|
|
|
python -m pip install -U pip
|
|
|
|
git clone https://github.com/explosion/spaCy
|
|
|
|
cd spaCy
|
|
|
|
|
|
|
|
python -m venv .env
|
|
|
|
source .env/bin/activate
|
|
|
|
export PYTHONPATH=`pwd`
|
|
|
|
pip install -r requirements.txt
|
|
|
|
python setup.py build_ext --inplace
|
|
|
|
```
|
|
|
|
|
|
|
|
Compared to regular install via pip, [requirements.txt](requirements.txt)
|
|
|
|
additionally installs developer dependencies such as Cython. For more details
|
|
|
|
and instructions, see the documentation on
|
2019-04-02 15:36:22 +00:00
|
|
|
[compiling spaCy from source](https://spacy.io/usage#source) and the
|
2019-10-02 08:37:04 +00:00
|
|
|
[quickstart widget](https://spacy.io/usage#section-quickstart) to get the right
|
|
|
|
commands for your platform and Python version.
|
2018-11-26 21:04:35 +00:00
|
|
|
|
|
|
|
### Ubuntu
|
|
|
|
|
|
|
|
Install system-level dependencies via `apt-get`:
|
|
|
|
|
|
|
|
```bash
|
|
|
|
sudo apt-get install build-essential python-dev git
|
|
|
|
```
|
|
|
|
|
|
|
|
### macOS / OS X
|
|
|
|
|
|
|
|
Install a recent version of [XCode](https://developer.apple.com/xcode/),
|
|
|
|
including the so-called "Command Line Tools". macOS and OS X ship with Python
|
|
|
|
and git preinstalled.
|
|
|
|
|
|
|
|
### Windows
|
|
|
|
|
2019-10-02 08:37:04 +00:00
|
|
|
Install a version of the
|
|
|
|
[Visual C++ Build Tools](https://visualstudio.microsoft.com/visual-cpp-build-tools/)
|
|
|
|
or [Visual Studio Express](https://visualstudio.microsoft.com/vs/express/) that
|
2019-12-22 00:53:56 +00:00
|
|
|
matches the version that was used to compile your Python interpreter.
|
2018-11-26 21:04:35 +00:00
|
|
|
|
|
|
|
## Run tests
|
|
|
|
|
2019-09-29 15:32:12 +00:00
|
|
|
spaCy comes with an [extensive test suite](spacy/tests). In order to run the
|
2018-11-26 21:04:35 +00:00
|
|
|
tests, you'll usually want to clone the repository and build spaCy from source.
|
|
|
|
This will also install the required development dependencies and test utilities
|
|
|
|
defined in the `requirements.txt`.
|
|
|
|
|
|
|
|
Alternatively, you can find out where spaCy is installed and run `pytest` on
|
|
|
|
that directory. Don't forget to also install the test utilities via spaCy's
|
|
|
|
`requirements.txt`:
|
|
|
|
|
|
|
|
```bash
|
|
|
|
python -c "import os; import spacy; print(os.path.dirname(spacy.__file__))"
|
|
|
|
pip install -r path/to/requirements.txt
|
|
|
|
python -m pytest <spacy-directory>
|
|
|
|
```
|
|
|
|
|
2019-04-02 15:36:22 +00:00
|
|
|
See [the documentation](https://spacy.io/usage#tests) for more details and
|
2018-11-26 21:04:35 +00:00
|
|
|
examples.
|