mirror of https://github.com/explosion/spaCy.git
Tidy up conftest
This commit is contained in:
parent
106d95b01a
commit
743ecf728c
|
@ -9,11 +9,59 @@ def pytest_addoption(parser):
|
|||
parser.addoption("--slow", action="store_true", help="include slow tests")
|
||||
|
||||
|
||||
def pytest_runtest_setup(item):
|
||||
def getopt(opt):
|
||||
# When using 'pytest --pyargs spacy' to test an installed copy of
|
||||
# spacy, pytest skips running our pytest_addoption() hook. Later, when
|
||||
# we call getoption(), pytest raises an error, because it doesn't
|
||||
# recognize the option we're asking about. To avoid this, we need to
|
||||
# pass a default value. We default to False, i.e., we act like all the
|
||||
# options weren't given.
|
||||
return item.config.getoption("--%s" % opt, False)
|
||||
|
||||
for opt in ["slow"]:
|
||||
if opt in item.keywords and not getopt(opt):
|
||||
pytest.skip("need --%s option to run" % opt)
|
||||
|
||||
|
||||
# Fixtures for language tokenizers (languages sorted alphabetically)
|
||||
|
||||
|
||||
@pytest.fixture(scope="module")
|
||||
def tokenizer():
|
||||
return get_lang_class("xx").Defaults.create_tokenizer()
|
||||
|
||||
|
||||
@pytest.fixture(scope="session")
|
||||
def ar_tokenizer():
|
||||
return get_lang_class("ar").Defaults.create_tokenizer()
|
||||
|
||||
|
||||
@pytest.fixture(scope="session")
|
||||
def bn_tokenizer():
|
||||
return get_lang_class("bn").Defaults.create_tokenizer()
|
||||
|
||||
|
||||
@pytest.fixture(scope="session")
|
||||
def ca_tokenizer():
|
||||
return get_lang_class("ca").Defaults.create_tokenizer()
|
||||
|
||||
|
||||
@pytest.fixture(scope="session")
|
||||
def da_tokenizer():
|
||||
return get_lang_class("da").Defaults.create_tokenizer()
|
||||
|
||||
|
||||
@pytest.fixture(scope="session")
|
||||
def de_tokenizer():
|
||||
return get_lang_class("de").Defaults.create_tokenizer()
|
||||
|
||||
|
||||
@pytest.fixture(scope="session")
|
||||
def el_tokenizer():
|
||||
return get_lang_class("el").Defaults.create_tokenizer()
|
||||
|
||||
|
||||
@pytest.fixture(scope="session")
|
||||
def en_tokenizer():
|
||||
return get_lang_class("en").Defaults.create_tokenizer()
|
||||
|
@ -36,8 +84,8 @@ def es_tokenizer():
|
|||
|
||||
|
||||
@pytest.fixture(scope="session")
|
||||
def de_tokenizer():
|
||||
return get_lang_class("de").Defaults.create_tokenizer()
|
||||
def fi_tokenizer():
|
||||
return get_lang_class("fi").Defaults.create_tokenizer()
|
||||
|
||||
|
||||
@pytest.fixture(scope="session")
|
||||
|
@ -45,21 +93,21 @@ def fr_tokenizer():
|
|||
return get_lang_class("fr").Defaults.create_tokenizer()
|
||||
|
||||
|
||||
@pytest.fixture(scope="session")
|
||||
def ga_tokenizer():
|
||||
return get_lang_class("ga").Defaults.create_tokenizer()
|
||||
|
||||
|
||||
@pytest.fixture(scope="session")
|
||||
def he_tokenizer():
|
||||
return get_lang_class("he").Defaults.create_tokenizer()
|
||||
|
||||
|
||||
@pytest.fixture
|
||||
def hu_tokenizer():
|
||||
return get_lang_class("hu").Defaults.create_tokenizer()
|
||||
|
||||
|
||||
@pytest.fixture(scope="session")
|
||||
def fi_tokenizer():
|
||||
return get_lang_class("fi").Defaults.create_tokenizer()
|
||||
|
||||
|
||||
@pytest.fixture(scope="session")
|
||||
def ro_tokenizer():
|
||||
return get_lang_class("ro").Defaults.create_tokenizer()
|
||||
|
||||
|
||||
@pytest.fixture(scope="session")
|
||||
def id_tokenizer():
|
||||
return get_lang_class("id").Defaults.create_tokenizer()
|
||||
|
@ -71,23 +119,9 @@ def it_tokenizer():
|
|||
|
||||
|
||||
@pytest.fixture(scope="session")
|
||||
def sv_tokenizer():
|
||||
return get_lang_class("sv").Defaults.create_tokenizer()
|
||||
|
||||
|
||||
@pytest.fixture(scope="session")
|
||||
def bn_tokenizer():
|
||||
return get_lang_class("bn").Defaults.create_tokenizer()
|
||||
|
||||
|
||||
@pytest.fixture(scope="session")
|
||||
def ga_tokenizer():
|
||||
return get_lang_class("ga").Defaults.create_tokenizer()
|
||||
|
||||
|
||||
@pytest.fixture(scope="session")
|
||||
def he_tokenizer():
|
||||
return get_lang_class("he").Defaults.create_tokenizer()
|
||||
def ja_tokenizer():
|
||||
pytest.importorskip("MeCab")
|
||||
return get_lang_class("ja").Defaults.create_tokenizer()
|
||||
|
||||
|
||||
@pytest.fixture(scope="session")
|
||||
|
@ -96,14 +130,34 @@ def nb_tokenizer():
|
|||
|
||||
|
||||
@pytest.fixture(scope="session")
|
||||
def da_tokenizer():
|
||||
return get_lang_class("da").Defaults.create_tokenizer()
|
||||
def nl_tokenizer():
|
||||
return get_lang_class("nl").Defaults.create_tokenizer()
|
||||
|
||||
|
||||
@pytest.fixture(scope="session")
|
||||
def ja_tokenizer():
|
||||
pytest.importorskip("MeCab")
|
||||
return get_lang_class("ja").Defaults.create_tokenizer()
|
||||
def pl_tokenizer():
|
||||
return get_lang_class("pl").Defaults.create_tokenizer()
|
||||
|
||||
|
||||
@pytest.fixture(scope="session")
|
||||
def pt_tokenizer():
|
||||
return get_lang_class("pt").Defaults.create_tokenizer()
|
||||
|
||||
|
||||
@pytest.fixture(scope="session")
|
||||
def ro_tokenizer():
|
||||
return get_lang_class("ro").Defaults.create_tokenizer()
|
||||
|
||||
|
||||
@pytest.fixture(scope="session")
|
||||
def ru_tokenizer():
|
||||
pytest.importorskip("pymorphy2")
|
||||
return get_lang_class("ru").Defaults.create_tokenizer()
|
||||
|
||||
|
||||
@pytest.fixture(scope="session")
|
||||
def sv_tokenizer():
|
||||
return get_lang_class("sv").Defaults.create_tokenizer()
|
||||
|
||||
|
||||
@pytest.fixture(scope="session")
|
||||
|
@ -117,58 +171,17 @@ def tr_tokenizer():
|
|||
return get_lang_class("tr").Defaults.create_tokenizer()
|
||||
|
||||
|
||||
@pytest.fixture(scope="session")
|
||||
def tt_tokenizer():
|
||||
return get_lang_class("tt").Defaults.create_tokenizer()
|
||||
|
||||
|
||||
@pytest.fixture(scope="session")
|
||||
def uk_tokenizer():
|
||||
pytest.importorskip("pymorphy2")
|
||||
return get_lang_class("uk").Defaults.create_tokenizer()
|
||||
|
||||
|
||||
@pytest.fixture(scope="session")
|
||||
def ca_tokenizer():
|
||||
return get_lang_class("ca").Defaults.create_tokenizer()
|
||||
|
||||
|
||||
@pytest.fixture(scope="session")
|
||||
def pl_tokenizer():
|
||||
return get_lang_class("pl").Defaults.create_tokenizer()
|
||||
|
||||
|
||||
@pytest.fixture(scope="session")
|
||||
def tt_tokenizer():
|
||||
return get_lang_class("tt").Defaults.create_tokenizer()
|
||||
|
||||
|
||||
@pytest.fixture(scope="session")
|
||||
def el_tokenizer():
|
||||
return get_lang_class("el").Defaults.create_tokenizer()
|
||||
|
||||
|
||||
@pytest.fixture(scope="session")
|
||||
def ar_tokenizer():
|
||||
return get_lang_class("ar").Defaults.create_tokenizer()
|
||||
|
||||
|
||||
@pytest.fixture(scope="session")
|
||||
def ur_tokenizer():
|
||||
return get_lang_class("ur").Defaults.create_tokenizer()
|
||||
|
||||
|
||||
@pytest.fixture(scope="session")
|
||||
def ru_tokenizer():
|
||||
pytest.importorskip("pymorphy2")
|
||||
return get_lang_class("ru").Defaults.create_tokenizer()
|
||||
|
||||
|
||||
def pytest_runtest_setup(item):
|
||||
def getopt(opt):
|
||||
# When using 'pytest --pyargs spacy' to test an installed copy of
|
||||
# spacy, pytest skips running our pytest_addoption() hook. Later, when
|
||||
# we call getoption(), pytest raises an error, because it doesn't
|
||||
# recognize the option we're asking about. To avoid this, we need to
|
||||
# pass a default value. We default to False, i.e., we act like all the
|
||||
# options weren't given.
|
||||
return item.config.getoption("--%s" % opt, False)
|
||||
|
||||
for opt in ["slow"]:
|
||||
if opt in item.keywords and not getopt(opt):
|
||||
pytest.skip("need --%s option to run" % opt)
|
||||
|
|
Loading…
Reference in New Issue