2017-03-23 10:08:41 +00:00
|
|
|
# coding: utf8
|
|
|
|
from __future__ import unicode_literals, division, print_function
|
|
|
|
|
2017-05-22 10:28:58 +00:00
|
|
|
import plac
|
2017-05-17 10:04:50 +00:00
|
|
|
from pathlib import Path
|
2017-05-21 14:07:06 +00:00
|
|
|
import tqdm
|
2017-09-21 00:17:10 +00:00
|
|
|
from thinc.neural._classes.model import Model
|
2017-05-23 08:06:53 +00:00
|
|
|
from timeit import default_timer as timer
|
2018-06-25 11:40:17 +00:00
|
|
|
import shutil
|
💫 Replace ujson, msgpack and dill/pickle/cloudpickle with srsly (#3003)
Remove hacks and wrappers, keep code in sync across our libraries and move spaCy a few steps closer to only depending on packages with binary wheels 🎉
See here: https://github.com/explosion/srsly
Serialization is hard, especially across Python versions and multiple platforms. After dealing with many subtle bugs over the years (encodings, locales, large files) our libraries like spaCy and Prodigy have steadily grown a number of utility functions to wrap the multiple serialization formats we need to support (especially json, msgpack and pickle). These wrapping functions ended up duplicated across our codebases, so we wanted to put them in one place.
At the same time, we noticed that having a lot of small dependencies was making maintainence harder, and making installation slower. To solve this, we've made srsly standalone, by including the component packages directly within it. This way we can provide all the serialization utilities we need in a single binary wheel.
srsly currently includes forks of the following packages:
ujson
msgpack
msgpack-numpy
cloudpickle
* WIP: replace json/ujson with srsly
* Replace ujson in examples
Use regular json instead of srsly to make code easier to read and follow
* Update requirements
* Fix imports
* Fix typos
* Replace msgpack with srsly
* Fix warning
2018-12-03 00:28:22 +00:00
|
|
|
import srsly
|
2018-11-30 19:16:14 +00:00
|
|
|
from wasabi import Printer
|
2017-03-23 10:08:41 +00:00
|
|
|
|
2018-04-03 13:50:31 +00:00
|
|
|
from ._messages import Messages
|
2018-11-30 19:16:14 +00:00
|
|
|
from .._ml import create_default_optimizer
|
2018-02-17 17:11:11 +00:00
|
|
|
from ..attrs import PROB, IS_OOV, CLUSTER, LANG
|
2018-03-27 17:23:02 +00:00
|
|
|
from ..gold import GoldCorpus
|
2017-03-23 10:08:41 +00:00
|
|
|
from .. import util
|
2017-09-25 17:00:47 +00:00
|
|
|
from .. import about
|
2018-11-30 19:16:14 +00:00
|
|
|
|
|
|
|
|
|
|
|
# Take dropout and batch size as generators of values -- dropout
|
|
|
|
# starts high and decays sharply, to force the optimizer to explore.
|
|
|
|
# Batch size starts at 1 and grows, so that we make updates quickly
|
|
|
|
# at the beginning of training.
|
|
|
|
dropout_rates = util.decaying(
|
2018-12-01 13:42:35 +00:00
|
|
|
util.env_opt("dropout_from", 0.1),
|
|
|
|
util.env_opt("dropout_to", 0.1),
|
2018-11-30 19:16:14 +00:00
|
|
|
util.env_opt("dropout_decay", 0.0),
|
|
|
|
)
|
|
|
|
batch_sizes = util.compounding(
|
2018-12-01 13:42:35 +00:00
|
|
|
util.env_opt("batch_from", 750),
|
|
|
|
util.env_opt("batch_to", 750),
|
2018-11-30 19:16:14 +00:00
|
|
|
util.env_opt("batch_compound", 1.001),
|
|
|
|
)
|
2017-03-23 10:08:41 +00:00
|
|
|
|
|
|
|
|
2017-05-22 10:28:58 +00:00
|
|
|
@plac.annotations(
|
2018-11-30 19:16:14 +00:00
|
|
|
lang=("Model language", "positional", None, str),
|
|
|
|
output_path=("Output directory to store model in", "positional", None, Path),
|
|
|
|
train_path=("Location of JSON-formatted training data", "positional", None, Path),
|
|
|
|
dev_path=("Location of JSON-formatted development data", "positional", None, Path),
|
|
|
|
base_model=("Name of model to update (optional)", "option", "b", str),
|
|
|
|
pipeline=("Comma-separated names of pipeline components", "option", "p", str),
|
|
|
|
vectors=("Model to load vectors from", "option", "v", str),
|
|
|
|
n_iter=("Number of iterations", "option", "n", int),
|
|
|
|
n_examples=("Number of examples", "option", "ns", int),
|
2017-06-03 21:10:23 +00:00
|
|
|
use_gpu=("Use GPU", "option", "g", int),
|
2017-09-26 15:59:34 +00:00
|
|
|
version=("Model version", "option", "V", str),
|
2018-11-30 19:16:14 +00:00
|
|
|
meta_path=("Optional path to meta.json to use as base.", "option", "m", Path),
|
|
|
|
init_tok2vec=(
|
|
|
|
"Path to pretrained weights for the token-to-vector parts of the models. See 'spacy pretrain'. Experimental.",
|
|
|
|
"option",
|
|
|
|
"t2v",
|
|
|
|
Path,
|
|
|
|
),
|
|
|
|
parser_multitasks=(
|
|
|
|
"Side objectives for parser CNN, e.g. 'dep' or 'dep,tag'",
|
|
|
|
"option",
|
|
|
|
"pt",
|
|
|
|
str,
|
|
|
|
),
|
|
|
|
entity_multitasks=(
|
|
|
|
"Side objectives for NER CNN, e.g. 'dep' or 'dep,tag'",
|
|
|
|
"option",
|
|
|
|
"et",
|
|
|
|
str,
|
|
|
|
),
|
|
|
|
noise_level=("Amount of corruption for data augmentation", "option", "nl", float),
|
|
|
|
gold_preproc=("Use gold preprocessing", "flag", "G", bool),
|
|
|
|
learn_tokens=("Make parser learn gold-standard tokenization", "flag", "T", bool),
|
|
|
|
verbose=("Display more information for debug", "flag", "VV", bool),
|
|
|
|
debug=("Run data diagnostics before training", "flag", "D", bool),
|
|
|
|
)
|
|
|
|
def train(
|
|
|
|
lang,
|
|
|
|
output_path,
|
|
|
|
train_path,
|
|
|
|
dev_path,
|
|
|
|
base_model=None,
|
|
|
|
pipeline="tagger,parser,ner",
|
|
|
|
vectors=None,
|
|
|
|
n_iter=30,
|
|
|
|
n_examples=0,
|
|
|
|
use_gpu=-1,
|
|
|
|
version="0.0.0",
|
|
|
|
meta_path=None,
|
|
|
|
init_tok2vec=None,
|
|
|
|
parser_multitasks="",
|
|
|
|
entity_multitasks="",
|
|
|
|
noise_level=0.0,
|
|
|
|
gold_preproc=False,
|
|
|
|
learn_tokens=False,
|
|
|
|
verbose=False,
|
|
|
|
debug=False,
|
|
|
|
):
|
2017-05-27 18:01:46 +00:00
|
|
|
"""
|
2018-11-30 19:16:14 +00:00
|
|
|
Train or update a spaCy model. Requires data to be formatted in spaCy's
|
|
|
|
JSON format. To convert data from other formats, use the `spacy convert`
|
|
|
|
command.
|
2017-05-27 18:01:46 +00:00
|
|
|
"""
|
2018-11-30 19:16:14 +00:00
|
|
|
msg = Printer()
|
2018-02-13 11:42:23 +00:00
|
|
|
util.fix_random_seed()
|
2018-11-30 19:16:14 +00:00
|
|
|
util.set_env_log(verbose)
|
|
|
|
|
|
|
|
# Make sure all files and paths exists if they are needed
|
|
|
|
train_path = util.ensure_path(train_path)
|
|
|
|
dev_path = util.ensure_path(dev_path)
|
2017-09-25 17:00:47 +00:00
|
|
|
meta_path = util.ensure_path(meta_path)
|
2018-11-30 19:16:14 +00:00
|
|
|
if not train_path or not train_path.exists():
|
|
|
|
msg.fail(Messages.M050, train_path, exits=1)
|
|
|
|
if not dev_path or not dev_path.exists():
|
|
|
|
msg.fail(Messages.M051, dev_path, exits=1)
|
2017-09-25 17:00:47 +00:00
|
|
|
if meta_path is not None and not meta_path.exists():
|
2018-11-30 19:16:14 +00:00
|
|
|
msg.fail(Messages.M020, meta_path, exits=1)
|
💫 Replace ujson, msgpack and dill/pickle/cloudpickle with srsly (#3003)
Remove hacks and wrappers, keep code in sync across our libraries and move spaCy a few steps closer to only depending on packages with binary wheels 🎉
See here: https://github.com/explosion/srsly
Serialization is hard, especially across Python versions and multiple platforms. After dealing with many subtle bugs over the years (encodings, locales, large files) our libraries like spaCy and Prodigy have steadily grown a number of utility functions to wrap the multiple serialization formats we need to support (especially json, msgpack and pickle). These wrapping functions ended up duplicated across our codebases, so we wanted to put them in one place.
At the same time, we noticed that having a lot of small dependencies was making maintainence harder, and making installation slower. To solve this, we've made srsly standalone, by including the component packages directly within it. This way we can provide all the serialization utilities we need in a single binary wheel.
srsly currently includes forks of the following packages:
ujson
msgpack
msgpack-numpy
cloudpickle
* WIP: replace json/ujson with srsly
* Replace ujson in examples
Use regular json instead of srsly to make code easier to read and follow
* Update requirements
* Fix imports
* Fix typos
* Replace msgpack with srsly
* Fix warning
2018-12-03 00:28:22 +00:00
|
|
|
meta = srsly.read_json(meta_path) if meta_path else {}
|
2017-09-25 17:00:47 +00:00
|
|
|
if not isinstance(meta, dict):
|
2018-11-30 19:16:14 +00:00
|
|
|
msg.fail(Messages.M052, Messages.M053.format(meta_type=type(meta)), exits=1)
|
|
|
|
if output_path.exists() and [p for p in output_path.iterdir() if p.is_dir()]:
|
|
|
|
msg.fail(Messages.M062, Messages.M065)
|
2018-03-27 17:23:02 +00:00
|
|
|
if not output_path.exists():
|
|
|
|
output_path.mkdir()
|
2017-03-23 10:08:41 +00:00
|
|
|
|
2018-11-30 19:16:14 +00:00
|
|
|
# Set up the base model and pipeline. If a base model is specified, load
|
|
|
|
# the model and make sure the pipeline matches the pipeline setting. If
|
|
|
|
# training starts from a blank model, intitalize the language class.
|
|
|
|
pipeline = [p.strip() for p in pipeline.split(",")]
|
|
|
|
msg.text(Messages.M055.format(pipeline=pipeline))
|
|
|
|
if base_model:
|
|
|
|
msg.text(Messages.M056.format(model=base_model))
|
|
|
|
nlp = util.load_model(base_model)
|
|
|
|
if nlp.lang != lang:
|
|
|
|
msg.fail(Messages.M072.format(model_lang=nlp.lang, lang=lang), exits=1)
|
|
|
|
other_pipes = [pipe for pipe in nlp.pipe_names if pipe not in pipeline]
|
|
|
|
nlp.disable_pipes(*other_pipes)
|
|
|
|
for pipe in pipeline:
|
|
|
|
if pipe not in nlp.pipe_names:
|
|
|
|
nlp.add_pipe(nlp.create_pipe(pipe))
|
|
|
|
else:
|
|
|
|
msg.text(Messages.M057.format(model=lang))
|
|
|
|
lang_cls = util.get_lang_class(lang)
|
|
|
|
nlp = lang_cls()
|
|
|
|
for pipe in pipeline:
|
|
|
|
nlp.add_pipe(nlp.create_pipe(pipe))
|
|
|
|
|
|
|
|
if learn_tokens:
|
|
|
|
nlp.add_pipe(nlp.create_pipe("merge_subtokens"))
|
2017-03-23 10:08:41 +00:00
|
|
|
|
2017-09-23 01:00:40 +00:00
|
|
|
if vectors:
|
2018-11-30 19:16:14 +00:00
|
|
|
msg.text(Messages.M058.format(model=vectors))
|
|
|
|
_load_vectors(nlp, vectors)
|
|
|
|
|
|
|
|
# Multitask objectives
|
|
|
|
multitask_options = [("parser", parser_multitasks), ("ner", entity_multitasks)]
|
|
|
|
for pipe_name, multitasks in multitask_options:
|
|
|
|
if multitasks:
|
|
|
|
if pipe_name not in pipeline:
|
|
|
|
msg.fail(Messages.M059.format(pipe=pipe_name))
|
|
|
|
pipe = nlp.get_pipe(pipe_name)
|
|
|
|
for objective in multitasks.split(","):
|
|
|
|
pipe.add_multitask_objective(objective)
|
|
|
|
|
|
|
|
# Prepare training corpus
|
|
|
|
msg.text(Messages.M060.format(limit=n_examples))
|
|
|
|
corpus = GoldCorpus(train_path, dev_path, limit=n_examples)
|
|
|
|
n_train_words = corpus.count_train()
|
|
|
|
|
|
|
|
if base_model:
|
|
|
|
# Start with an existing model, use default optimizer
|
|
|
|
optimizer = create_default_optimizer(Model.ops)
|
|
|
|
else:
|
|
|
|
# Start with a blank model, call begin_training
|
|
|
|
optimizer = nlp.begin_training(lambda: corpus.train_tuples, device=use_gpu)
|
2018-12-01 13:42:35 +00:00
|
|
|
|
2017-09-02 17:46:01 +00:00
|
|
|
nlp._optimizer = None
|
2017-05-25 21:16:30 +00:00
|
|
|
|
2018-11-30 19:16:14 +00:00
|
|
|
# Load in pre-trained weights
|
|
|
|
if init_tok2vec is not None:
|
|
|
|
components = _load_pretrained_tok2vec(nlp, init_tok2vec)
|
|
|
|
msg.text(Messages.M071.format(components=components))
|
|
|
|
|
|
|
|
print(
|
|
|
|
"\nItn. Dep Loss NER Loss UAS NER P. NER R. NER F. Tag % Token % CPU WPS GPU WPS"
|
|
|
|
)
|
2017-05-26 10:52:09 +00:00
|
|
|
try:
|
|
|
|
for i in range(n_iter):
|
2018-11-30 19:16:14 +00:00
|
|
|
train_docs = corpus.train_docs(
|
|
|
|
nlp, noise_level=noise_level, gold_preproc=gold_preproc, max_length=0
|
|
|
|
)
|
2018-03-27 17:23:02 +00:00
|
|
|
words_seen = 0
|
2017-06-05 01:18:37 +00:00
|
|
|
with tqdm.tqdm(total=n_train_words, leave=False) as pbar:
|
2017-05-26 10:52:09 +00:00
|
|
|
losses = {}
|
2018-11-30 19:16:14 +00:00
|
|
|
for batch in util.minibatch_by_words(train_docs, size=batch_sizes):
|
2017-11-03 00:54:54 +00:00
|
|
|
if not batch:
|
|
|
|
continue
|
2017-05-26 10:52:09 +00:00
|
|
|
docs, golds = zip(*batch)
|
2018-11-30 19:16:14 +00:00
|
|
|
nlp.update(
|
|
|
|
docs,
|
|
|
|
golds,
|
|
|
|
sgd=optimizer,
|
|
|
|
drop=next(dropout_rates),
|
|
|
|
losses=losses,
|
|
|
|
)
|
2017-06-05 01:18:37 +00:00
|
|
|
pbar.update(sum(len(doc) for doc in docs))
|
2018-03-27 17:23:02 +00:00
|
|
|
words_seen += sum(len(doc) for doc in docs)
|
2017-05-26 10:52:09 +00:00
|
|
|
with nlp.use_params(optimizer.averages):
|
2017-06-03 18:28:20 +00:00
|
|
|
util.set_env_log(False)
|
2018-11-30 19:16:14 +00:00
|
|
|
epoch_model_path = output_path / ("model%d" % i)
|
2017-06-03 18:28:20 +00:00
|
|
|
nlp.to_disk(epoch_model_path)
|
2017-10-10 17:51:20 +00:00
|
|
|
nlp_loaded = util.load_model_from_path(epoch_model_path)
|
2018-11-30 19:16:14 +00:00
|
|
|
dev_docs = list(corpus.dev_docs(nlp_loaded, gold_preproc=gold_preproc))
|
2017-10-09 13:05:37 +00:00
|
|
|
nwords = sum(len(doc_gold[0]) for doc_gold in dev_docs)
|
|
|
|
start_time = timer()
|
2018-11-30 19:16:14 +00:00
|
|
|
scorer = nlp_loaded.evaluate(dev_docs, debug)
|
2017-10-09 13:05:37 +00:00
|
|
|
end_time = timer()
|
|
|
|
if use_gpu < 0:
|
|
|
|
gpu_wps = None
|
2018-11-30 19:16:14 +00:00
|
|
|
cpu_wps = nwords / (end_time - start_time)
|
2017-10-09 13:05:37 +00:00
|
|
|
else:
|
2018-11-30 19:16:14 +00:00
|
|
|
gpu_wps = nwords / (end_time - start_time)
|
|
|
|
with Model.use_device("cpu"):
|
2017-10-10 17:51:20 +00:00
|
|
|
nlp_loaded = util.load_model_from_path(epoch_model_path)
|
2018-11-30 19:16:14 +00:00
|
|
|
dev_docs = list(
|
|
|
|
corpus.dev_docs(nlp_loaded, gold_preproc=gold_preproc)
|
|
|
|
)
|
2017-10-09 13:05:37 +00:00
|
|
|
start_time = timer()
|
|
|
|
scorer = nlp_loaded.evaluate(dev_docs)
|
|
|
|
end_time = timer()
|
2018-11-30 19:16:14 +00:00
|
|
|
cpu_wps = nwords / (end_time - start_time)
|
|
|
|
acc_loc = output_path / ("model%d" % i) / "accuracy.json"
|
💫 Replace ujson, msgpack and dill/pickle/cloudpickle with srsly (#3003)
Remove hacks and wrappers, keep code in sync across our libraries and move spaCy a few steps closer to only depending on packages with binary wheels 🎉
See here: https://github.com/explosion/srsly
Serialization is hard, especially across Python versions and multiple platforms. After dealing with many subtle bugs over the years (encodings, locales, large files) our libraries like spaCy and Prodigy have steadily grown a number of utility functions to wrap the multiple serialization formats we need to support (especially json, msgpack and pickle). These wrapping functions ended up duplicated across our codebases, so we wanted to put them in one place.
At the same time, we noticed that having a lot of small dependencies was making maintainence harder, and making installation slower. To solve this, we've made srsly standalone, by including the component packages directly within it. This way we can provide all the serialization utilities we need in a single binary wheel.
srsly currently includes forks of the following packages:
ujson
msgpack
msgpack-numpy
cloudpickle
* WIP: replace json/ujson with srsly
* Replace ujson in examples
Use regular json instead of srsly to make code easier to read and follow
* Update requirements
* Fix imports
* Fix typos
* Replace msgpack with srsly
* Fix warning
2018-12-03 00:28:22 +00:00
|
|
|
srsly.write_json(acc_loc, scorer.scores)
|
2018-11-30 19:16:14 +00:00
|
|
|
|
|
|
|
# Update model meta.json
|
|
|
|
meta["lang"] = nlp.lang
|
|
|
|
meta["pipeline"] = nlp.pipe_names
|
|
|
|
meta["spacy_version"] = ">=%s" % about.__version__
|
|
|
|
meta["accuracy"] = scorer.scores
|
|
|
|
meta["speed"] = {"nwords": nwords, "cpu": cpu_wps, "gpu": gpu_wps}
|
|
|
|
meta["vectors"] = {
|
|
|
|
"width": nlp.vocab.vectors_length,
|
|
|
|
"vectors": len(nlp.vocab.vectors),
|
|
|
|
"keys": nlp.vocab.vectors.n_keys,
|
|
|
|
}
|
|
|
|
meta.setdefault("name", "model%d" % i)
|
|
|
|
meta.setdefault("version", version)
|
|
|
|
meta_loc = output_path / ("model%d" % i) / "meta.json"
|
💫 Replace ujson, msgpack and dill/pickle/cloudpickle with srsly (#3003)
Remove hacks and wrappers, keep code in sync across our libraries and move spaCy a few steps closer to only depending on packages with binary wheels 🎉
See here: https://github.com/explosion/srsly
Serialization is hard, especially across Python versions and multiple platforms. After dealing with many subtle bugs over the years (encodings, locales, large files) our libraries like spaCy and Prodigy have steadily grown a number of utility functions to wrap the multiple serialization formats we need to support (especially json, msgpack and pickle). These wrapping functions ended up duplicated across our codebases, so we wanted to put them in one place.
At the same time, we noticed that having a lot of small dependencies was making maintainence harder, and making installation slower. To solve this, we've made srsly standalone, by including the component packages directly within it. This way we can provide all the serialization utilities we need in a single binary wheel.
srsly currently includes forks of the following packages:
ujson
msgpack
msgpack-numpy
cloudpickle
* WIP: replace json/ujson with srsly
* Replace ujson in examples
Use regular json instead of srsly to make code easier to read and follow
* Update requirements
* Fix imports
* Fix typos
* Replace msgpack with srsly
* Fix warning
2018-12-03 00:28:22 +00:00
|
|
|
srsly.write_json(meta_loc, meta)
|
2018-11-30 19:16:14 +00:00
|
|
|
|
|
|
|
util.set_env_log(verbose)
|
2017-09-25 17:00:47 +00:00
|
|
|
|
2018-11-30 19:16:14 +00:00
|
|
|
print_progress(i, losses, scorer.scores, cpu_wps=cpu_wps, gpu_wps=gpu_wps)
|
2017-05-26 10:52:09 +00:00
|
|
|
finally:
|
2018-11-30 19:16:14 +00:00
|
|
|
with msg.loading(Messages.M061):
|
|
|
|
with nlp.use_params(optimizer.averages):
|
|
|
|
final_model_path = output_path / "model-final"
|
|
|
|
nlp.to_disk(final_model_path)
|
|
|
|
msg.good(Messages.M066, util.path2str(final_model_path))
|
|
|
|
|
|
|
|
_collate_best_model(meta, output_path, nlp.pipe_names)
|
|
|
|
|
|
|
|
|
|
|
|
def _load_vectors(nlp, vectors):
|
|
|
|
util.load_model(vectors, vocab=nlp.vocab)
|
|
|
|
for lex in nlp.vocab:
|
|
|
|
values = {}
|
|
|
|
for attr, func in nlp.vocab.lex_attr_getters.items():
|
|
|
|
# These attrs are expected to be set by data. Others should
|
|
|
|
# be set by calling the language functions.
|
|
|
|
if attr not in (CLUSTER, PROB, IS_OOV, LANG):
|
|
|
|
values[lex.vocab.strings[attr]] = func(lex.orth_)
|
|
|
|
lex.set_attrs(**values)
|
|
|
|
lex.is_oov = False
|
2018-06-25 14:36:42 +00:00
|
|
|
|
2018-06-24 21:39:52 +00:00
|
|
|
|
2018-11-15 21:17:16 +00:00
|
|
|
def _load_pretrained_tok2vec(nlp, loc):
|
|
|
|
"""Load pre-trained weights for the 'token-to-vector' part of the component
|
|
|
|
models, which is typically a CNN. See 'spacy pretrain'. Experimental.
|
|
|
|
"""
|
2018-11-30 19:16:14 +00:00
|
|
|
with loc.open("rb") as file_:
|
2018-11-15 21:17:16 +00:00
|
|
|
weights_data = file_.read()
|
|
|
|
loaded = []
|
|
|
|
for name, component in nlp.pipeline:
|
2018-11-30 19:16:14 +00:00
|
|
|
if hasattr(component, "model") and hasattr(component.model, "tok2vec"):
|
2018-11-15 23:34:54 +00:00
|
|
|
component.tok2vec.from_bytes(weights_data)
|
2018-11-15 21:17:16 +00:00
|
|
|
loaded.append(name)
|
|
|
|
return loaded
|
|
|
|
|
|
|
|
|
2018-06-24 21:39:52 +00:00
|
|
|
def _collate_best_model(meta, output_path, components):
|
|
|
|
bests = {}
|
|
|
|
for component in components:
|
|
|
|
bests[component] = _find_best(output_path, component)
|
2018-11-30 19:16:14 +00:00
|
|
|
best_dest = output_path / "model-best"
|
|
|
|
shutil.copytree(output_path / "model-final", best_dest)
|
2018-06-24 21:39:52 +00:00
|
|
|
for component, best_component_src in bests.items():
|
2018-06-25 12:35:24 +00:00
|
|
|
shutil.rmtree(best_dest / component)
|
2018-06-25 21:05:56 +00:00
|
|
|
shutil.copytree(best_component_src / component, best_dest / component)
|
💫 Replace ujson, msgpack and dill/pickle/cloudpickle with srsly (#3003)
Remove hacks and wrappers, keep code in sync across our libraries and move spaCy a few steps closer to only depending on packages with binary wheels 🎉
See here: https://github.com/explosion/srsly
Serialization is hard, especially across Python versions and multiple platforms. After dealing with many subtle bugs over the years (encodings, locales, large files) our libraries like spaCy and Prodigy have steadily grown a number of utility functions to wrap the multiple serialization formats we need to support (especially json, msgpack and pickle). These wrapping functions ended up duplicated across our codebases, so we wanted to put them in one place.
At the same time, we noticed that having a lot of small dependencies was making maintainence harder, and making installation slower. To solve this, we've made srsly standalone, by including the component packages directly within it. This way we can provide all the serialization utilities we need in a single binary wheel.
srsly currently includes forks of the following packages:
ujson
msgpack
msgpack-numpy
cloudpickle
* WIP: replace json/ujson with srsly
* Replace ujson in examples
Use regular json instead of srsly to make code easier to read and follow
* Update requirements
* Fix imports
* Fix typos
* Replace msgpack with srsly
* Fix warning
2018-12-03 00:28:22 +00:00
|
|
|
accs = srsly.read_json(best_component_src / "accuracy.json")
|
2018-06-24 21:39:52 +00:00
|
|
|
for metric in _get_metrics(component):
|
2018-11-30 19:16:14 +00:00
|
|
|
meta["accuracy"][metric] = accs[metric]
|
💫 Replace ujson, msgpack and dill/pickle/cloudpickle with srsly (#3003)
Remove hacks and wrappers, keep code in sync across our libraries and move spaCy a few steps closer to only depending on packages with binary wheels 🎉
See here: https://github.com/explosion/srsly
Serialization is hard, especially across Python versions and multiple platforms. After dealing with many subtle bugs over the years (encodings, locales, large files) our libraries like spaCy and Prodigy have steadily grown a number of utility functions to wrap the multiple serialization formats we need to support (especially json, msgpack and pickle). These wrapping functions ended up duplicated across our codebases, so we wanted to put them in one place.
At the same time, we noticed that having a lot of small dependencies was making maintainence harder, and making installation slower. To solve this, we've made srsly standalone, by including the component packages directly within it. This way we can provide all the serialization utilities we need in a single binary wheel.
srsly currently includes forks of the following packages:
ujson
msgpack
msgpack-numpy
cloudpickle
* WIP: replace json/ujson with srsly
* Replace ujson in examples
Use regular json instead of srsly to make code easier to read and follow
* Update requirements
* Fix imports
* Fix typos
* Replace msgpack with srsly
* Fix warning
2018-12-03 00:28:22 +00:00
|
|
|
srsly.write_json(best_dest / "meta.json", meta)
|
2018-06-24 21:39:52 +00:00
|
|
|
|
|
|
|
|
|
|
|
def _find_best(experiment_dir, component):
|
|
|
|
accuracies = []
|
|
|
|
for epoch_model in experiment_dir.iterdir():
|
|
|
|
if epoch_model.is_dir() and epoch_model.parts[-1] != "model-final":
|
💫 Replace ujson, msgpack and dill/pickle/cloudpickle with srsly (#3003)
Remove hacks and wrappers, keep code in sync across our libraries and move spaCy a few steps closer to only depending on packages with binary wheels 🎉
See here: https://github.com/explosion/srsly
Serialization is hard, especially across Python versions and multiple platforms. After dealing with many subtle bugs over the years (encodings, locales, large files) our libraries like spaCy and Prodigy have steadily grown a number of utility functions to wrap the multiple serialization formats we need to support (especially json, msgpack and pickle). These wrapping functions ended up duplicated across our codebases, so we wanted to put them in one place.
At the same time, we noticed that having a lot of small dependencies was making maintainence harder, and making installation slower. To solve this, we've made srsly standalone, by including the component packages directly within it. This way we can provide all the serialization utilities we need in a single binary wheel.
srsly currently includes forks of the following packages:
ujson
msgpack
msgpack-numpy
cloudpickle
* WIP: replace json/ujson with srsly
* Replace ujson in examples
Use regular json instead of srsly to make code easier to read and follow
* Update requirements
* Fix imports
* Fix typos
* Replace msgpack with srsly
* Fix warning
2018-12-03 00:28:22 +00:00
|
|
|
accs = srsly.read_json(epoch_model / "accuracy.json")
|
2018-06-24 21:39:52 +00:00
|
|
|
scores = [accs.get(metric, 0.0) for metric in _get_metrics(component)]
|
|
|
|
accuracies.append((scores, epoch_model))
|
|
|
|
if accuracies:
|
|
|
|
return max(accuracies)[1]
|
|
|
|
else:
|
|
|
|
return None
|
|
|
|
|
2018-11-30 19:16:14 +00:00
|
|
|
|
2018-06-24 21:39:52 +00:00
|
|
|
def _get_metrics(component):
|
|
|
|
if component == "parser":
|
|
|
|
return ("las", "uas", "token_acc")
|
|
|
|
elif component == "tagger":
|
|
|
|
return ("tags_acc",)
|
|
|
|
elif component == "ner":
|
|
|
|
return ("ents_f", "ents_p", "ents_r")
|
|
|
|
return ("token_acc",)
|
2017-05-19 23:15:50 +00:00
|
|
|
|
|
|
|
|
2017-10-09 13:05:37 +00:00
|
|
|
def print_progress(itn, losses, dev_scores, cpu_wps=0.0, gpu_wps=0.0):
|
2017-05-16 14:17:30 +00:00
|
|
|
scores = {}
|
2018-11-30 19:16:14 +00:00
|
|
|
for col in [
|
|
|
|
"dep_loss",
|
|
|
|
"tag_loss",
|
|
|
|
"uas",
|
|
|
|
"tags_acc",
|
|
|
|
"token_acc",
|
|
|
|
"ents_p",
|
|
|
|
"ents_r",
|
|
|
|
"ents_f",
|
|
|
|
"cpu_wps",
|
|
|
|
"gpu_wps",
|
|
|
|
]:
|
2017-05-16 14:17:30 +00:00
|
|
|
scores[col] = 0.0
|
2018-11-30 19:16:14 +00:00
|
|
|
scores["dep_loss"] = losses.get("parser", 0.0)
|
|
|
|
scores["ner_loss"] = losses.get("ner", 0.0)
|
|
|
|
scores["tag_loss"] = losses.get("tagger", 0.0)
|
2017-05-16 14:17:30 +00:00
|
|
|
scores.update(dev_scores)
|
2018-11-30 19:16:14 +00:00
|
|
|
scores["cpu_wps"] = cpu_wps
|
|
|
|
scores["gpu_wps"] = gpu_wps or 0.0
|
|
|
|
tpl = "".join(
|
|
|
|
(
|
|
|
|
"{:<6d}",
|
|
|
|
"{dep_loss:<10.3f}",
|
|
|
|
"{ner_loss:<10.3f}",
|
|
|
|
"{uas:<8.3f}",
|
|
|
|
"{ents_p:<8.3f}",
|
|
|
|
"{ents_r:<8.3f}",
|
|
|
|
"{ents_f:<8.3f}",
|
|
|
|
"{tags_acc:<8.3f}",
|
|
|
|
"{token_acc:<9.3f}",
|
|
|
|
"{cpu_wps:<9.1f}",
|
|
|
|
"{gpu_wps:.1f}",
|
|
|
|
)
|
|
|
|
)
|
2017-05-16 14:17:30 +00:00
|
|
|
print(tpl.format(itn, **scores))
|