From 60269589576222330fcb6f7f30fdff2ce3b71a95 Mon Sep 17 00:00:00 2001 From: svlandeg Date: Mon, 15 Jul 2019 11:19:34 +0200 Subject: [PATCH] tokenizer doc fix --- spacy/tokenizer.pyx | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/spacy/tokenizer.pyx b/spacy/tokenizer.pyx index 70a693ba1..f19f851c7 100644 --- a/spacy/tokenizer.pyx +++ b/spacy/tokenizer.pyx @@ -348,7 +348,7 @@ cdef class Tokenizer: """Add a special-case tokenization rule. string (unicode): The string to specially tokenize. - token_attrs (iterable): A sequence of dicts, where each dict describes + substrings (iterable): A sequence of dicts, where each dict describes a token and its attributes. The `ORTH` fields of the attributes must exactly match the string when they are concatenated.