spaCy/spacy/lang/uk/tokenizer_exceptions.py

39 lines
1.7 KiB
Python
Raw Blame History

This file contains ambiguous Unicode characters

This file contains Unicode characters that might be confused with other characters. If you think that this is intentional, you can safely ignore this warning. Use the Escape button to reveal them.

# coding: utf8
from __future__ import unicode_literals
# import symbols if you need to use more, add them here
from ...symbols import ORTH, LEMMA, POS, NORM, NOUN
# Add tokenizer exceptions
# Documentation: https://spacy.io/docs/usage/adding-languages#tokenizer-exceptions
# Feel free to use custom logic to generate repetitive exceptions more efficiently.
# If an exception is split into more than one token, the ORTH values combined always
# need to match the original string.
# Exceptions should be added in the following format:
_exc = {}
for exc_data in [
{ORTH: "вул.", LEMMA: "вулиця", NORM: "вулиця", POS: NOUN},
{ORTH: "ім.", LEMMA: "ім'я", NORM: "імені", POS: NOUN},
{ORTH: "просп.", LEMMA: "проспект", NORM: "проспект", POS: NOUN},
{ORTH: "бул.", LEMMA: "бульвар", NORM: "бульвар", POS: NOUN},
{ORTH: "пров.", LEMMA: "провулок", NORM: "провулок", POS: NOUN},
{ORTH: "пл.", LEMMA: "площа", NORM: "площа", POS: NOUN},
{ORTH: "г.", LEMMA: "гора", NORM: "гора", POS: NOUN},
{ORTH: "п.", LEMMA: "пан", NORM: "пан", POS: NOUN},
{ORTH: "м.", LEMMA: "місто", NORM: "місто", POS: NOUN},
{ORTH: "проф.", LEMMA: "професор", NORM: "професор", POS: NOUN},
{ORTH: "акад.", LEMMA: "академік", NORM: "академік", POS: NOUN},
{ORTH: "доц.", LEMMA: "доцент", NORM: "доцент", POS: NOUN},
{ORTH: "оз.", LEMMA: "озеро", NORM: "озеро", POS: NOUN}]:
_exc[exc_data[ORTH]] = [exc_data]
# To keep things clean and readable, it's recommended to only declare the
# TOKENIZER_EXCEPTIONS at the bottom:
TOKENIZER_EXCEPTIONS = _exc