2020-09-29 14:05:48 +00:00
|
|
|
from typing import Union, Dict, Optional, Any, List, IO, TYPE_CHECKING
|
2020-09-28 13:09:59 +00:00
|
|
|
from thinc.api import Config, fix_random_seed, set_gpu_allocator
|
|
|
|
from thinc.api import ConfigValidationError
|
|
|
|
from pathlib import Path
|
|
|
|
import srsly
|
2020-09-29 08:58:50 +00:00
|
|
|
import numpy
|
|
|
|
import tarfile
|
|
|
|
import gzip
|
|
|
|
import zipfile
|
|
|
|
import tqdm
|
2020-09-28 13:09:59 +00:00
|
|
|
|
|
|
|
from .loop import create_before_to_disk_callback
|
|
|
|
from ..lookups import Lookups
|
2020-09-29 08:58:50 +00:00
|
|
|
from ..vectors import Vectors
|
2020-09-28 13:09:59 +00:00
|
|
|
from ..errors import Errors
|
2020-09-29 14:08:39 +00:00
|
|
|
from ..schemas import ConfigSchemaTraining
|
|
|
|
from ..util import registry, load_model_from_config, resolve_dot_names, logger
|
2020-09-28 19:17:10 +00:00
|
|
|
from ..util import load_model, ensure_path, OOV_RANK, DEFAULT_OOV_PROB
|
2020-09-28 13:09:59 +00:00
|
|
|
|
2020-09-29 14:05:48 +00:00
|
|
|
if TYPE_CHECKING:
|
|
|
|
from ..language import Language # noqa: F401
|
2020-09-28 13:09:59 +00:00
|
|
|
|
2020-09-29 14:05:48 +00:00
|
|
|
|
2020-09-29 14:08:39 +00:00
|
|
|
def init_nlp(config: Config, *, use_gpu: int = -1) -> "Language":
|
2020-09-28 13:09:59 +00:00
|
|
|
raw_config = config
|
|
|
|
config = raw_config.interpolate()
|
|
|
|
if config["training"]["seed"] is not None:
|
|
|
|
fix_random_seed(config["training"]["seed"])
|
|
|
|
allocator = config["training"]["gpu_allocator"]
|
|
|
|
if use_gpu >= 0 and allocator:
|
|
|
|
set_gpu_allocator(allocator)
|
|
|
|
# Use original config here before it's resolved to functions
|
|
|
|
sourced_components = get_sourced_components(config)
|
|
|
|
nlp = load_model_from_config(raw_config, auto_fill=True)
|
2020-09-29 14:08:39 +00:00
|
|
|
logger.info("Set up nlp object from config")
|
2020-09-28 13:09:59 +00:00
|
|
|
config = nlp.config.interpolate()
|
|
|
|
# Resolve all training-relevant sections using the filled nlp config
|
|
|
|
T = registry.resolve(config["training"], schema=ConfigSchemaTraining)
|
|
|
|
dot_names = [T["train_corpus"], T["dev_corpus"]]
|
|
|
|
train_corpus, dev_corpus = resolve_dot_names(config, dot_names)
|
|
|
|
optimizer = T["optimizer"]
|
|
|
|
before_to_disk = create_before_to_disk_callback(T["before_to_disk"])
|
|
|
|
# Components that shouldn't be updated during training
|
|
|
|
frozen_components = T["frozen_components"]
|
|
|
|
# Sourced components that require resume_training
|
|
|
|
resume_components = [p for p in sourced_components if p not in frozen_components]
|
2020-09-29 14:08:39 +00:00
|
|
|
logger.info(f"Pipeline: {nlp.pipe_names}")
|
2020-09-28 13:09:59 +00:00
|
|
|
if resume_components:
|
|
|
|
with nlp.select_pipes(enable=resume_components):
|
2020-09-29 14:08:39 +00:00
|
|
|
logger.info(f"Resuming training for: {resume_components}")
|
2020-09-28 13:09:59 +00:00
|
|
|
nlp.resume_training(sgd=optimizer)
|
|
|
|
with nlp.select_pipes(disable=[*frozen_components, *resume_components]):
|
2020-09-29 14:05:48 +00:00
|
|
|
nlp.initialize(lambda: train_corpus(nlp), sgd=optimizer)
|
2020-09-29 14:22:41 +00:00
|
|
|
logger.info("Initialized pipeline components")
|
2020-09-28 19:35:09 +00:00
|
|
|
# Verify the config after calling 'initialize' to ensure labels
|
2020-09-28 13:09:59 +00:00
|
|
|
# are properly initialized
|
|
|
|
verify_config(nlp)
|
|
|
|
nlp = before_to_disk(nlp)
|
|
|
|
return nlp
|
|
|
|
|
|
|
|
|
|
|
|
def must_reinitialize(train_config: Config, init_config: Config) -> bool:
|
|
|
|
# TODO: do this better and more fine-grained
|
|
|
|
return train_config.interpolate().to_str() == init_config.interpolate().to_str()
|
|
|
|
|
|
|
|
|
|
|
|
def init_vocab(
|
2020-09-29 14:05:48 +00:00
|
|
|
nlp: "Language",
|
2020-09-28 13:09:59 +00:00
|
|
|
*,
|
|
|
|
data: Optional[Path] = None,
|
|
|
|
lookups: Optional[Lookups] = None,
|
|
|
|
vectors: Optional[str] = None,
|
2020-09-29 14:05:48 +00:00
|
|
|
) -> "Language":
|
2020-09-28 13:09:59 +00:00
|
|
|
if lookups:
|
|
|
|
nlp.vocab.lookups = lookups
|
2020-09-29 14:08:39 +00:00
|
|
|
logger.info(f"Added vocab lookups: {', '.join(lookups.tables)}")
|
2020-09-28 13:09:59 +00:00
|
|
|
data_path = ensure_path(data)
|
|
|
|
if data_path is not None:
|
|
|
|
lex_attrs = srsly.read_jsonl(data_path)
|
|
|
|
for lexeme in nlp.vocab:
|
|
|
|
lexeme.rank = OOV_RANK
|
|
|
|
for attrs in lex_attrs:
|
|
|
|
if "settings" in attrs:
|
|
|
|
continue
|
|
|
|
lexeme = nlp.vocab[attrs["orth"]]
|
|
|
|
lexeme.set_attrs(**attrs)
|
|
|
|
if len(nlp.vocab):
|
|
|
|
oov_prob = min(lex.prob for lex in nlp.vocab) - 1
|
|
|
|
else:
|
|
|
|
oov_prob = DEFAULT_OOV_PROB
|
|
|
|
nlp.vocab.cfg.update({"oov_prob": oov_prob})
|
2020-09-29 14:22:41 +00:00
|
|
|
logger.info(f"Added {len(nlp.vocab)} lexical entries to the vocab")
|
|
|
|
logger.info("Created vocabulary")
|
2020-09-28 13:09:59 +00:00
|
|
|
if vectors is not None:
|
|
|
|
load_vectors_into_model(nlp, vectors)
|
2020-09-29 14:22:41 +00:00
|
|
|
logger.info(f"Added vectors: {vectors}")
|
2020-09-28 13:09:59 +00:00
|
|
|
|
|
|
|
|
|
|
|
def load_vectors_into_model(
|
2020-09-29 14:05:48 +00:00
|
|
|
nlp: "Language", name: Union[str, Path], *, add_strings: bool = True
|
2020-09-28 13:09:59 +00:00
|
|
|
) -> None:
|
|
|
|
"""Load word vectors from an installed model or path into a model instance."""
|
|
|
|
try:
|
|
|
|
vectors_nlp = load_model(name)
|
|
|
|
except ConfigValidationError as e:
|
|
|
|
title = f"Config validation error for vectors {name}"
|
|
|
|
desc = (
|
|
|
|
"This typically means that there's a problem in the config.cfg included "
|
|
|
|
"with the packaged vectors. Make sure that the vectors package you're "
|
|
|
|
"loading is compatible with the current version of spaCy."
|
|
|
|
)
|
|
|
|
err = ConfigValidationError.from_error(config=None, title=title, desc=desc)
|
|
|
|
raise err from None
|
|
|
|
nlp.vocab.vectors = vectors_nlp.vocab.vectors
|
|
|
|
if add_strings:
|
|
|
|
# I guess we should add the strings from the vectors_nlp model?
|
|
|
|
# E.g. if someone does a similarity query, they might expect the strings.
|
|
|
|
for key in nlp.vocab.vectors.key2row:
|
|
|
|
if key in vectors_nlp.vocab.strings:
|
|
|
|
nlp.vocab.strings.add(vectors_nlp.vocab.strings[key])
|
|
|
|
|
|
|
|
|
2020-09-29 14:05:48 +00:00
|
|
|
def init_tok2vec(
|
2020-09-29 14:47:55 +00:00
|
|
|
nlp: "Language", pretrain_config: Dict[str, Any], init_config: Dict[str, Any]
|
2020-09-28 13:09:59 +00:00
|
|
|
) -> bool:
|
|
|
|
# Load pretrained tok2vec weights - cf. CLI command 'pretrain'
|
|
|
|
P = pretrain_config
|
2020-09-29 14:47:55 +00:00
|
|
|
I = init_config
|
2020-09-28 13:09:59 +00:00
|
|
|
weights_data = None
|
2020-09-29 14:47:55 +00:00
|
|
|
init_tok2vec = ensure_path(I["init_tok2vec"])
|
2020-09-28 13:09:59 +00:00
|
|
|
if init_tok2vec is not None:
|
2020-09-29 14:47:55 +00:00
|
|
|
if P["objective"].get("type") == "vectors" and not I["vectors"]:
|
2020-09-29 18:38:35 +00:00
|
|
|
err = 'need initialize.vectors if pretraining.objective.type is "vectors"'
|
|
|
|
errors = [{"loc": ["initialize"], "msg": err}]
|
2020-09-28 13:09:59 +00:00
|
|
|
raise ConfigValidationError(config=nlp.config, errors=errors)
|
|
|
|
if not init_tok2vec.exists():
|
|
|
|
err = f"can't find pretrained tok2vec: {init_tok2vec}"
|
2020-09-29 18:38:35 +00:00
|
|
|
errors = [{"loc": ["initialize", "init_tok2vec"], "msg": err}]
|
2020-09-28 13:09:59 +00:00
|
|
|
raise ConfigValidationError(config=nlp.config, errors=errors)
|
|
|
|
with init_tok2vec.open("rb") as file_:
|
|
|
|
weights_data = file_.read()
|
|
|
|
if weights_data is not None:
|
|
|
|
tok2vec_component = P["component"]
|
|
|
|
if tok2vec_component is None:
|
|
|
|
desc = (
|
|
|
|
f"To use pretrained tok2vec weights, [pretraining.component] "
|
|
|
|
f"needs to specify the component that should load them."
|
|
|
|
)
|
|
|
|
err = "component can't be null"
|
|
|
|
errors = [{"loc": ["pretraining", "component"], "msg": err}]
|
|
|
|
raise ConfigValidationError(
|
|
|
|
config=nlp.config["pretraining"], errors=errors, desc=desc
|
|
|
|
)
|
|
|
|
layer = nlp.get_pipe(tok2vec_component).model
|
|
|
|
if P["layer"]:
|
|
|
|
layer = layer.get_ref(P["layer"])
|
|
|
|
layer.from_bytes(weights_data)
|
|
|
|
return True
|
|
|
|
return False
|
|
|
|
|
|
|
|
|
2020-09-29 14:05:48 +00:00
|
|
|
def verify_config(nlp: "Language") -> None:
|
2020-09-28 13:09:59 +00:00
|
|
|
"""Perform additional checks based on the config, loaded nlp object and training data."""
|
|
|
|
# TODO: maybe we should validate based on the actual components, the list
|
|
|
|
# in config["nlp"]["pipeline"] instead?
|
|
|
|
for pipe_config in nlp.config["components"].values():
|
|
|
|
# We can't assume that the component name == the factory
|
|
|
|
factory = pipe_config["factory"]
|
|
|
|
if factory == "textcat":
|
|
|
|
verify_textcat_config(nlp, pipe_config)
|
|
|
|
|
|
|
|
|
2020-09-29 14:05:48 +00:00
|
|
|
def verify_textcat_config(nlp: "Language", pipe_config: Dict[str, Any]) -> None:
|
2020-09-28 13:09:59 +00:00
|
|
|
# if 'positive_label' is provided: double check whether it's in the data and
|
|
|
|
# the task is binary
|
|
|
|
if pipe_config.get("positive_label"):
|
|
|
|
textcat_labels = nlp.get_pipe("textcat").labels
|
|
|
|
pos_label = pipe_config.get("positive_label")
|
|
|
|
if pos_label not in textcat_labels:
|
|
|
|
raise ValueError(
|
|
|
|
Errors.E920.format(pos_label=pos_label, labels=textcat_labels)
|
|
|
|
)
|
|
|
|
if len(list(textcat_labels)) != 2:
|
|
|
|
raise ValueError(
|
|
|
|
Errors.E919.format(pos_label=pos_label, labels=textcat_labels)
|
|
|
|
)
|
|
|
|
|
|
|
|
|
|
|
|
def get_sourced_components(config: Union[Dict[str, Any], Config]) -> List[str]:
|
|
|
|
"""RETURNS (List[str]): All sourced components in the original config,
|
2020-09-29 19:39:28 +00:00
|
|
|
e.g. {"source": "en_core_web_sm"}. If the config contains a key
|
|
|
|
"factory", we assume it refers to a component factory.
|
2020-09-28 13:09:59 +00:00
|
|
|
"""
|
|
|
|
return [
|
|
|
|
name
|
|
|
|
for name, cfg in config.get("components", {}).items()
|
|
|
|
if "factory" not in cfg and "source" in cfg
|
|
|
|
]
|
2020-09-29 08:58:50 +00:00
|
|
|
|
|
|
|
|
|
|
|
def convert_vectors(
|
2020-09-29 14:05:48 +00:00
|
|
|
nlp: "Language",
|
2020-09-29 08:58:50 +00:00
|
|
|
vectors_loc: Optional[Path],
|
|
|
|
*,
|
|
|
|
truncate: int,
|
|
|
|
prune: int,
|
|
|
|
name: Optional[str] = None,
|
|
|
|
) -> None:
|
|
|
|
vectors_loc = ensure_path(vectors_loc)
|
|
|
|
if vectors_loc and vectors_loc.parts[-1].endswith(".npz"):
|
|
|
|
nlp.vocab.vectors = Vectors(data=numpy.load(vectors_loc.open("rb")))
|
|
|
|
for lex in nlp.vocab:
|
|
|
|
if lex.rank and lex.rank != OOV_RANK:
|
|
|
|
nlp.vocab.vectors.add(lex.orth, row=lex.rank)
|
|
|
|
else:
|
|
|
|
if vectors_loc:
|
2020-09-29 14:08:39 +00:00
|
|
|
logger.info(f"Reading vectors from {vectors_loc}")
|
|
|
|
vectors_data, vector_keys = read_vectors(vectors_loc, truncate)
|
|
|
|
logger.info(f"Loaded vectors from {vectors_loc}")
|
2020-09-29 08:58:50 +00:00
|
|
|
else:
|
|
|
|
vectors_data, vector_keys = (None, None)
|
|
|
|
if vector_keys is not None:
|
|
|
|
for word in vector_keys:
|
|
|
|
if word not in nlp.vocab:
|
|
|
|
nlp.vocab[word]
|
|
|
|
if vectors_data is not None:
|
|
|
|
nlp.vocab.vectors = Vectors(data=vectors_data, keys=vector_keys)
|
|
|
|
if name is None:
|
|
|
|
# TODO: Is this correct? Does this matter?
|
|
|
|
nlp.vocab.vectors.name = f"{nlp.meta['lang']}_{nlp.meta['name']}.vectors"
|
|
|
|
else:
|
|
|
|
nlp.vocab.vectors.name = name
|
|
|
|
nlp.meta["vectors"]["name"] = nlp.vocab.vectors.name
|
|
|
|
if prune >= 1:
|
|
|
|
nlp.vocab.prune_vectors(prune)
|
|
|
|
|
|
|
|
|
|
|
|
def read_vectors(vectors_loc: Path, truncate_vectors: int):
|
|
|
|
f = open_file(vectors_loc)
|
|
|
|
f = ensure_shape(f)
|
|
|
|
shape = tuple(int(size) for size in next(f).split())
|
|
|
|
if truncate_vectors >= 1:
|
|
|
|
shape = (truncate_vectors, shape[1])
|
|
|
|
vectors_data = numpy.zeros(shape=shape, dtype="f")
|
|
|
|
vectors_keys = []
|
|
|
|
for i, line in enumerate(tqdm.tqdm(f)):
|
|
|
|
line = line.rstrip()
|
|
|
|
pieces = line.rsplit(" ", vectors_data.shape[1])
|
|
|
|
word = pieces.pop(0)
|
|
|
|
if len(pieces) != vectors_data.shape[1]:
|
|
|
|
raise ValueError(Errors.E094.format(line_num=i, loc=vectors_loc))
|
|
|
|
vectors_data[i] = numpy.asarray(pieces, dtype="f")
|
|
|
|
vectors_keys.append(word)
|
|
|
|
if i == truncate_vectors - 1:
|
|
|
|
break
|
|
|
|
return vectors_data, vectors_keys
|
|
|
|
|
|
|
|
|
|
|
|
def open_file(loc: Union[str, Path]) -> IO:
|
|
|
|
"""Handle .gz, .tar.gz or unzipped files"""
|
|
|
|
loc = ensure_path(loc)
|
|
|
|
if tarfile.is_tarfile(str(loc)):
|
|
|
|
return tarfile.open(str(loc), "r:gz")
|
|
|
|
elif loc.parts[-1].endswith("gz"):
|
|
|
|
return (line.decode("utf8") for line in gzip.open(str(loc), "r"))
|
|
|
|
elif loc.parts[-1].endswith("zip"):
|
|
|
|
zip_file = zipfile.ZipFile(str(loc))
|
|
|
|
names = zip_file.namelist()
|
|
|
|
file_ = zip_file.open(names[0])
|
|
|
|
return (line.decode("utf8") for line in file_)
|
|
|
|
else:
|
|
|
|
return loc.open("r", encoding="utf8")
|
|
|
|
|
|
|
|
|
|
|
|
def ensure_shape(lines):
|
|
|
|
"""Ensure that the first line of the data is the vectors shape.
|
|
|
|
If it's not, we read in the data and output the shape as the first result,
|
|
|
|
so that the reader doesn't have to deal with the problem.
|
|
|
|
"""
|
|
|
|
first_line = next(lines)
|
|
|
|
try:
|
|
|
|
shape = tuple(int(size) for size in first_line.split())
|
|
|
|
except ValueError:
|
|
|
|
shape = None
|
|
|
|
if shape is not None:
|
|
|
|
# All good, give the data
|
|
|
|
yield first_line
|
|
|
|
yield from lines
|
|
|
|
else:
|
|
|
|
# Figure out the shape, make it the first value, and then give the
|
|
|
|
# rest of the data.
|
|
|
|
width = len(first_line.split()) - 1
|
|
|
|
captured = [first_line] + list(lines)
|
|
|
|
length = len(captured)
|
|
|
|
yield f"{length} {width}"
|
|
|
|
yield from captured
|