diff --git a/spacy/tests/tokenizer/test_tokenizer.py b/spacy/tests/tokenizer/test_tokenizer.py index ed11508b4..b27af6bcd 100644 --- a/spacy/tests/tokenizer/test_tokenizer.py +++ b/spacy/tests/tokenizer/test_tokenizer.py @@ -529,7 +529,7 @@ def test_tokenizer_initial_special_case_explain(en_vocab): token_match=re.compile("^id$").match, rules={ "id": [{"ORTH": "i"}, {"ORTH": "d"}], - } + }, ) tokens = [t.text for t in tokenizer("id")] explain_tokens = [t[1] for t in tokenizer.explain("id")]