mirror of https://github.com/explosion/spaCy.git
Add links to tokenizer API docs to refer relevant information. (#4064)
* Add links to tokenizer API docs to refer relevant information. * Add suggested changes Co-Authored-By: Ines Montani <ines@ines.io>
This commit is contained in:
parent
2cdf7d39e7
commit
a83c0add2e
|
@ -5,7 +5,7 @@ tag: class
|
||||||
source: spacy/tokenizer.pyx
|
source: spacy/tokenizer.pyx
|
||||||
---
|
---
|
||||||
|
|
||||||
Segment text, and create `Doc` objects with the discovered segment boundaries.
|
Segment text, and create `Doc` objects with the discovered segment boundaries. For a deeper understanding, see the docs on [how spaCy's tokenizer works](/usage/linguistic-features#how-tokenizer-works).
|
||||||
|
|
||||||
## Tokenizer.\_\_init\_\_ {#init tag="method"}
|
## Tokenizer.\_\_init\_\_ {#init tag="method"}
|
||||||
|
|
||||||
|
@ -109,7 +109,7 @@ if no suffix rules match.
|
||||||
|
|
||||||
Add a special-case tokenization rule. This mechanism is also used to add custom
|
Add a special-case tokenization rule. This mechanism is also used to add custom
|
||||||
tokenizer exceptions to the language data. See the usage guide on
|
tokenizer exceptions to the language data. See the usage guide on
|
||||||
[adding languages](/usage/adding-languages#tokenizer-exceptions) for more
|
[adding languages](/usage/adding-languages#tokenizer-exceptions) and [linguistic features](/usage/linguistic-features#special-cases) for more
|
||||||
details and examples.
|
details and examples.
|
||||||
|
|
||||||
> #### Example
|
> #### Example
|
||||||
|
|
Loading…
Reference in New Issue