From d12a0b643146e58b6b76dbb3514dc2aa62eb0f41 Mon Sep 17 00:00:00 2001 From: luvogels Date: Wed, 26 Apr 2017 23:21:41 +0200 Subject: [PATCH] Hooked up tokenizer tests --- spacy/tests/conftest.py | 7 ++++++- 1 file changed, 6 insertions(+), 1 deletion(-) diff --git a/spacy/tests/conftest.py b/spacy/tests/conftest.py index 37d3180d0..b8ada1d9a 100644 --- a/spacy/tests/conftest.py +++ b/spacy/tests/conftest.py @@ -13,6 +13,8 @@ from ..hu import Hungarian from ..fi import Finnish from ..bn import Bengali from ..he import Hebrew +from ..nb import Norwegian + from ..tokens import Doc from ..strings import StringStore @@ -26,7 +28,7 @@ import pytest LANGUAGES = [English, German, Spanish, Italian, French, Portuguese, Dutch, - Swedish, Hungarian, Finnish, Bengali] + Swedish, Hungarian, Finnish, Bengali, Norwegian] @pytest.fixture(params=LANGUAGES) @@ -88,6 +90,9 @@ def bn_tokenizer(): def he_tokenizer(): return Hebrew.Defaults.create_tokenizer() +@pytest.fixture +def nb_tokenizer(): + return Norwegian.Defaults.create_tokenizer() @pytest.fixture def stringstore():