mirror of https://github.com/explosion/spaCy.git
50 lines
1.5 KiB
Python
50 lines
1.5 KiB
Python
# encoding: utf8
|
||
from __future__ import unicode_literals, print_function
|
||
|
||
from os import path
|
||
|
||
from ..language import Language
|
||
from . import language_data
|
||
from .. import util
|
||
from ..lemmatizer import Lemmatizer
|
||
from ..vocab import Vocab
|
||
from ..tokenizer import Tokenizer
|
||
from ..attrs import LANG
|
||
|
||
from ..language_data import update_exc
|
||
from ..language_data import strings_to_exc
|
||
from ..language_data import expand_exc
|
||
from ..language_data import EMOTICONS
|
||
|
||
from .language_data import ORTH_ONLY
|
||
from .language_data import get_time_exc
|
||
|
||
|
||
TOKENIZER_EXCEPTIONS = dict(language_data.TOKENIZER_EXCEPTIONS)
|
||
TOKENIZER_PREFIXES = tuple(language_data.TOKENIZER_PREFIXES)
|
||
TOKENIZER_SUFFIXES = tuple(language_data.TOKENIZER_SUFFIXES)
|
||
TOKENIZER_INFIXES = tuple(language_data.TOKENIZER_INFIXES)
|
||
TAG_MAP = dict(language_data.TAG_MAP)
|
||
STOP_WORDS = set(language_data.STOP_WORDS)
|
||
|
||
|
||
update_exc(TOKENIZER_EXCEPTIONS, strings_to_exc(EMOTICONS))
|
||
update_exc(TOKENIZER_EXCEPTIONS, strings_to_exc(ORTH_ONLY))
|
||
update_exc(TOKENIZER_EXCEPTIONS, get_time_exc(range(1, 12 + 1)))
|
||
update_exc(TOKENIZER_EXCEPTIONS, expand_exc(TOKENIZER_EXCEPTIONS, "'", "’"))
|
||
|
||
|
||
class English(Language):
|
||
lang = 'en'
|
||
|
||
class Defaults(Language.Defaults):
|
||
lex_attr_getters = dict(Language.Defaults.lex_attr_getters)
|
||
lex_attr_getters[LANG] = lambda text: 'en'
|
||
|
||
tokenizer_exceptions = TOKENIZER_EXCEPTIONS
|
||
prefixes = TOKENIZER_PREFIXES
|
||
suffixes = TOKENIZER_SUFFIXES
|
||
infixes = TOKENIZER_INFIXES
|
||
tag_map = TAG_MAP
|
||
stop_words = STOP_WORDS
|