2018-11-30 19:16:14 +00:00
|
|
|
# coding: utf8
|
2018-11-15 21:17:16 +00:00
|
|
|
from __future__ import print_function, unicode_literals
|
2018-11-30 19:16:14 +00:00
|
|
|
|
2018-11-15 21:17:16 +00:00
|
|
|
import plac
|
|
|
|
import random
|
|
|
|
import numpy
|
|
|
|
import time
|
2018-11-15 22:45:36 +00:00
|
|
|
from collections import Counter
|
2018-11-30 19:16:14 +00:00
|
|
|
from pathlib import Path
|
2018-11-29 12:36:43 +00:00
|
|
|
from thinc.v2v import Affine, Maxout
|
|
|
|
from thinc.misc import LayerNorm as LN
|
2019-02-08 13:14:49 +00:00
|
|
|
from thinc.neural.util import prefer_gpu
|
2018-11-30 19:16:14 +00:00
|
|
|
from wasabi import Printer
|
💫 Replace ujson, msgpack and dill/pickle/cloudpickle with srsly (#3003)
Remove hacks and wrappers, keep code in sync across our libraries and move spaCy a few steps closer to only depending on packages with binary wheels 🎉
See here: https://github.com/explosion/srsly
Serialization is hard, especially across Python versions and multiple platforms. After dealing with many subtle bugs over the years (encodings, locales, large files) our libraries like spaCy and Prodigy have steadily grown a number of utility functions to wrap the multiple serialization formats we need to support (especially json, msgpack and pickle). These wrapping functions ended up duplicated across our codebases, so we wanted to put them in one place.
At the same time, we noticed that having a lot of small dependencies was making maintainence harder, and making installation slower. To solve this, we've made srsly standalone, by including the component packages directly within it. This way we can provide all the serialization utilities we need in a single binary wheel.
srsly currently includes forks of the following packages:
ujson
msgpack
msgpack-numpy
cloudpickle
* WIP: replace json/ujson with srsly
* Replace ujson in examples
Use regular json instead of srsly to make code easier to read and follow
* Update requirements
* Fix imports
* Fix typos
* Replace msgpack with srsly
* Fix warning
2018-12-03 00:28:22 +00:00
|
|
|
import srsly
|
2018-11-15 21:17:16 +00:00
|
|
|
|
2018-11-30 19:16:14 +00:00
|
|
|
from ..tokens import Doc
|
|
|
|
from ..attrs import ID, HEAD
|
2019-02-08 13:14:49 +00:00
|
|
|
from .._ml import Tok2Vec, flatten, chain, create_default_optimizer
|
💫 Better support for semi-supervised learning (#3035)
The new spacy pretrain command implemented BERT/ULMFit/etc-like transfer learning, using our Language Modelling with Approximate Outputs version of BERT's cloze task. Pretraining is convenient, but in some ways it's a bit of a strange solution. All we're doing is initialising the weights. At the same time, we're putting a lot of work into our optimisation so that it's less sensitive to initial conditions, and more likely to find good optima. I discuss this a bit in the pseudo-rehearsal blog post: https://explosion.ai/blog/pseudo-rehearsal-catastrophic-forgetting
Support semi-supervised learning in spacy train
One obvious way to improve these pretraining methods is to do multi-task learning, instead of just transfer learning. This has been shown to work very well: https://arxiv.org/pdf/1809.08370.pdf . This patch makes it easy to do this sort of thing.
Add a new argument to spacy train, --raw-text. This takes a jsonl file with unlabelled data that can be used in arbitrary ways to do semi-supervised learning.
Add a new method to the Language class and to pipeline components, .rehearse(). This is like .update(), but doesn't expect GoldParse objects. It takes a batch of Doc objects, and performs an update on some semi-supervised objective.
Move the BERT-LMAO objective out from spacy/cli/pretrain.py into spacy/_ml.py, so we can create a new pipeline component, ClozeMultitask. This can be specified as a parser or NER multitask in the spacy train command. Example usage:
python -m spacy train en ./tmp ~/data/en-core-web/train/nw.json ~/data/en-core-web/dev/nw.json --pipeline parser --raw-textt ~/data/unlabelled/reddit-100k.jsonl --vectors en_vectors_web_lg --parser-multitasks cloze
Implement rehearsal methods for pipeline components
The new --raw-text argument and nlp.rehearse() method also gives us a good place to implement the the idea in the pseudo-rehearsal blog post in the parser. This works as follows:
Add a new nlp.resume_training() method. This allocates copies of pre-trained models in the pipeline, setting things up for the rehearsal updates. It also returns an optimizer object. This also greatly reduces confusion around the nlp.begin_training() method, which randomises the weights, making it not suitable for adding new labels or otherwise fine-tuning a pre-trained model.
Implement rehearsal updates on the Parser class, making it available for the dependency parser and NER. During rehearsal, the initial model is used to supervise the model being trained. The current model is asked to match the predictions of the initial model on some data. This minimises catastrophic forgetting, by keeping the model's predictions close to the original. See the blog post for details.
Implement rehearsal updates for tagger
Implement rehearsal updates for text categoriz
2018-12-10 15:25:33 +00:00
|
|
|
from .._ml import masked_language_model
|
2018-11-30 19:16:14 +00:00
|
|
|
from .. import util
|
2018-11-15 21:17:16 +00:00
|
|
|
|
|
|
|
|
2018-11-30 19:16:14 +00:00
|
|
|
@plac.annotations(
|
|
|
|
texts_loc=("Path to jsonl file with texts to learn from", "positional", None, str),
|
|
|
|
vectors_model=("Name or path to vectors model to learn from"),
|
|
|
|
output_dir=("Directory to write models each epoch", "positional", None, str),
|
|
|
|
width=("Width of CNN layers", "option", "cw", int),
|
|
|
|
depth=("Depth of CNN layers", "option", "cd", int),
|
|
|
|
embed_rows=("Embedding rows", "option", "er", int),
|
|
|
|
use_vectors=("Whether to use the static vectors as input features", "flag", "uv"),
|
|
|
|
dropout=("Dropout", "option", "d", float),
|
2019-03-16 20:38:45 +00:00
|
|
|
batch_size=("Number of words per training batch", "option", "bs", int),
|
|
|
|
max_length=("Max words per example.", "option", "xw", int),
|
|
|
|
min_length=("Min words per example.", "option", "nw", int),
|
2018-11-30 19:16:14 +00:00
|
|
|
seed=("Seed for random number generators", "option", "s", float),
|
|
|
|
nr_iter=("Number of iterations to pretrain", "option", "i", int),
|
|
|
|
)
|
|
|
|
def pretrain(
|
|
|
|
texts_loc,
|
|
|
|
vectors_model,
|
|
|
|
output_dir,
|
|
|
|
width=96,
|
|
|
|
depth=4,
|
|
|
|
embed_rows=2000,
|
|
|
|
use_vectors=False,
|
|
|
|
dropout=0.2,
|
|
|
|
nr_iter=1000,
|
2019-03-16 20:38:45 +00:00
|
|
|
batch_size=3000,
|
|
|
|
max_length=500,
|
|
|
|
min_length=5,
|
2018-11-30 19:16:14 +00:00
|
|
|
seed=0,
|
|
|
|
):
|
|
|
|
"""
|
|
|
|
Pre-train the 'token-to-vector' (tok2vec) layer of pipeline components,
|
|
|
|
using an approximate language-modelling objective. Specifically, we load
|
|
|
|
pre-trained vectors, and train a component like a CNN, BiLSTM, etc to predict
|
|
|
|
vectors which match the pre-trained ones. The weights are saved to a directory
|
|
|
|
after each epoch. You can then pass a path to one of these pre-trained weights
|
|
|
|
files to the 'spacy train' command.
|
|
|
|
|
|
|
|
This technique may be especially helpful if you have little labelled data.
|
|
|
|
However, it's still quite experimental, so your mileage may vary.
|
2018-11-15 21:17:16 +00:00
|
|
|
|
2018-11-30 19:16:14 +00:00
|
|
|
To load the weights back in during 'spacy train', you need to ensure
|
|
|
|
all settings are the same between pretraining and training. The API and
|
|
|
|
errors around this need some improvement.
|
|
|
|
"""
|
|
|
|
config = dict(locals())
|
|
|
|
msg = Printer()
|
|
|
|
util.fix_random_seed(seed)
|
|
|
|
|
|
|
|
has_gpu = prefer_gpu()
|
|
|
|
msg.info("Using GPU" if has_gpu else "Not using GPU")
|
|
|
|
|
|
|
|
output_dir = Path(output_dir)
|
|
|
|
if not output_dir.exists():
|
|
|
|
output_dir.mkdir()
|
|
|
|
msg.good("Created output directory")
|
💫 Replace ujson, msgpack and dill/pickle/cloudpickle with srsly (#3003)
Remove hacks and wrappers, keep code in sync across our libraries and move spaCy a few steps closer to only depending on packages with binary wheels 🎉
See here: https://github.com/explosion/srsly
Serialization is hard, especially across Python versions and multiple platforms. After dealing with many subtle bugs over the years (encodings, locales, large files) our libraries like spaCy and Prodigy have steadily grown a number of utility functions to wrap the multiple serialization formats we need to support (especially json, msgpack and pickle). These wrapping functions ended up duplicated across our codebases, so we wanted to put them in one place.
At the same time, we noticed that having a lot of small dependencies was making maintainence harder, and making installation slower. To solve this, we've made srsly standalone, by including the component packages directly within it. This way we can provide all the serialization utilities we need in a single binary wheel.
srsly currently includes forks of the following packages:
ujson
msgpack
msgpack-numpy
cloudpickle
* WIP: replace json/ujson with srsly
* Replace ujson in examples
Use regular json instead of srsly to make code easier to read and follow
* Update requirements
* Fix imports
* Fix typos
* Replace msgpack with srsly
* Fix warning
2018-12-03 00:28:22 +00:00
|
|
|
srsly.write_json(output_dir / "config.json", config)
|
2018-11-30 19:16:14 +00:00
|
|
|
msg.good("Saved settings to config.json")
|
|
|
|
|
|
|
|
# Load texts from file or stdin
|
|
|
|
if texts_loc != "-": # reading from a file
|
|
|
|
texts_loc = Path(texts_loc)
|
|
|
|
if not texts_loc.exists():
|
|
|
|
msg.fail("Input text file doesn't exist", texts_loc, exits=1)
|
|
|
|
with msg.loading("Loading input texts..."):
|
💫 Replace ujson, msgpack and dill/pickle/cloudpickle with srsly (#3003)
Remove hacks and wrappers, keep code in sync across our libraries and move spaCy a few steps closer to only depending on packages with binary wheels 🎉
See here: https://github.com/explosion/srsly
Serialization is hard, especially across Python versions and multiple platforms. After dealing with many subtle bugs over the years (encodings, locales, large files) our libraries like spaCy and Prodigy have steadily grown a number of utility functions to wrap the multiple serialization formats we need to support (especially json, msgpack and pickle). These wrapping functions ended up duplicated across our codebases, so we wanted to put them in one place.
At the same time, we noticed that having a lot of small dependencies was making maintainence harder, and making installation slower. To solve this, we've made srsly standalone, by including the component packages directly within it. This way we can provide all the serialization utilities we need in a single binary wheel.
srsly currently includes forks of the following packages:
ujson
msgpack
msgpack-numpy
cloudpickle
* WIP: replace json/ujson with srsly
* Replace ujson in examples
Use regular json instead of srsly to make code easier to read and follow
* Update requirements
* Fix imports
* Fix typos
* Replace msgpack with srsly
* Fix warning
2018-12-03 00:28:22 +00:00
|
|
|
texts = list(srsly.read_jsonl(texts_loc))
|
2018-11-30 19:16:14 +00:00
|
|
|
msg.good("Loaded input texts")
|
|
|
|
random.shuffle(texts)
|
|
|
|
else: # reading from stdin
|
|
|
|
msg.text("Reading input text from stdin...")
|
💫 Replace ujson, msgpack and dill/pickle/cloudpickle with srsly (#3003)
Remove hacks and wrappers, keep code in sync across our libraries and move spaCy a few steps closer to only depending on packages with binary wheels 🎉
See here: https://github.com/explosion/srsly
Serialization is hard, especially across Python versions and multiple platforms. After dealing with many subtle bugs over the years (encodings, locales, large files) our libraries like spaCy and Prodigy have steadily grown a number of utility functions to wrap the multiple serialization formats we need to support (especially json, msgpack and pickle). These wrapping functions ended up duplicated across our codebases, so we wanted to put them in one place.
At the same time, we noticed that having a lot of small dependencies was making maintainence harder, and making installation slower. To solve this, we've made srsly standalone, by including the component packages directly within it. This way we can provide all the serialization utilities we need in a single binary wheel.
srsly currently includes forks of the following packages:
ujson
msgpack
msgpack-numpy
cloudpickle
* WIP: replace json/ujson with srsly
* Replace ujson in examples
Use regular json instead of srsly to make code easier to read and follow
* Update requirements
* Fix imports
* Fix typos
* Replace msgpack with srsly
* Fix warning
2018-12-03 00:28:22 +00:00
|
|
|
texts = srsly.read_jsonl("-")
|
2018-11-30 19:16:14 +00:00
|
|
|
|
|
|
|
with msg.loading("Loading model '{}'...".format(vectors_model)):
|
|
|
|
nlp = util.load_model(vectors_model)
|
|
|
|
msg.good("Loaded model '{}'".format(vectors_model))
|
|
|
|
pretrained_vectors = None if not use_vectors else nlp.vocab.vectors.name
|
|
|
|
model = create_pretraining_model(
|
|
|
|
nlp,
|
|
|
|
Tok2Vec(
|
|
|
|
width,
|
|
|
|
embed_rows,
|
|
|
|
conv_depth=depth,
|
|
|
|
pretrained_vectors=pretrained_vectors,
|
|
|
|
bilstm_depth=0, # Requires PyTorch. Experimental.
|
2018-12-17 14:48:27 +00:00
|
|
|
cnn_maxout_pieces=3, # You can try setting this higher
|
2018-12-18 18:19:26 +00:00
|
|
|
subword_features=True, # Set to False for Chinese etc
|
2018-11-30 19:16:14 +00:00
|
|
|
),
|
2018-12-18 18:19:26 +00:00
|
|
|
)
|
2018-11-30 19:16:14 +00:00
|
|
|
optimizer = create_default_optimizer(model.ops)
|
2018-12-18 18:19:26 +00:00
|
|
|
tracker = ProgressTracker(frequency=10000)
|
2018-11-30 19:16:14 +00:00
|
|
|
msg.divider("Pre-training tok2vec layer")
|
|
|
|
row_settings = {"widths": (3, 10, 10, 6, 4), "aligns": ("r", "r", "r", "r", "r")}
|
|
|
|
msg.row(("#", "# Words", "Total Loss", "Loss", "w/s"), **row_settings)
|
|
|
|
for epoch in range(nr_iter):
|
|
|
|
for batch in util.minibatch_by_words(
|
2019-03-16 20:38:45 +00:00
|
|
|
((text, None) for text in texts), size=batch_size
|
2018-11-30 19:16:14 +00:00
|
|
|
):
|
2019-03-16 20:38:45 +00:00
|
|
|
docs = make_docs(
|
|
|
|
nlp,
|
|
|
|
[text for (text, _) in batch],
|
|
|
|
max_length=max_length,
|
|
|
|
min_length=min_length,
|
|
|
|
)
|
2018-11-30 19:16:14 +00:00
|
|
|
loss = make_update(model, docs, optimizer, drop=dropout)
|
|
|
|
progress = tracker.update(epoch, loss, docs)
|
|
|
|
if progress:
|
|
|
|
msg.row(progress, **row_settings)
|
|
|
|
if texts_loc == "-" and tracker.words_per_epoch[epoch] >= 10 ** 7:
|
|
|
|
break
|
|
|
|
with model.use_params(optimizer.averages):
|
|
|
|
with (output_dir / ("model%d.bin" % epoch)).open("wb") as file_:
|
|
|
|
file_.write(model.tok2vec.to_bytes())
|
|
|
|
log = {
|
|
|
|
"nr_word": tracker.nr_word,
|
|
|
|
"loss": tracker.loss,
|
|
|
|
"epoch_loss": tracker.epoch_loss,
|
|
|
|
"epoch": epoch,
|
|
|
|
}
|
|
|
|
with (output_dir / "log.jsonl").open("a") as file_:
|
💫 Replace ujson, msgpack and dill/pickle/cloudpickle with srsly (#3003)
Remove hacks and wrappers, keep code in sync across our libraries and move spaCy a few steps closer to only depending on packages with binary wheels 🎉
See here: https://github.com/explosion/srsly
Serialization is hard, especially across Python versions and multiple platforms. After dealing with many subtle bugs over the years (encodings, locales, large files) our libraries like spaCy and Prodigy have steadily grown a number of utility functions to wrap the multiple serialization formats we need to support (especially json, msgpack and pickle). These wrapping functions ended up duplicated across our codebases, so we wanted to put them in one place.
At the same time, we noticed that having a lot of small dependencies was making maintainence harder, and making installation slower. To solve this, we've made srsly standalone, by including the component packages directly within it. This way we can provide all the serialization utilities we need in a single binary wheel.
srsly currently includes forks of the following packages:
ujson
msgpack
msgpack-numpy
cloudpickle
* WIP: replace json/ujson with srsly
* Replace ujson in examples
Use regular json instead of srsly to make code easier to read and follow
* Update requirements
* Fix imports
* Fix typos
* Replace msgpack with srsly
* Fix warning
2018-12-03 00:28:22 +00:00
|
|
|
file_.write(srsly.json_dumps(log) + "\n")
|
2018-11-30 19:16:14 +00:00
|
|
|
tracker.epoch_loss = 0.0
|
|
|
|
if texts_loc != "-":
|
|
|
|
# Reshuffle the texts if texts were loaded from a file
|
|
|
|
random.shuffle(texts)
|
2018-11-15 22:44:07 +00:00
|
|
|
|
2018-11-28 17:04:58 +00:00
|
|
|
|
2019-02-08 13:14:49 +00:00
|
|
|
def make_update(model, docs, optimizer, drop=0.0, objective="L2"):
|
2018-11-15 21:17:16 +00:00
|
|
|
"""Perform an update over a single batch of documents.
|
|
|
|
|
|
|
|
docs (iterable): A batch of `Doc` objects.
|
|
|
|
drop (float): The droput rate.
|
|
|
|
optimizer (callable): An optimizer.
|
|
|
|
RETURNS loss: A float for the loss.
|
|
|
|
"""
|
|
|
|
predictions, backprop = model.begin_update(docs, drop=drop)
|
2018-12-18 18:19:26 +00:00
|
|
|
loss, gradients = get_vectors_loss(model.ops, docs, predictions, objective)
|
2018-11-15 21:17:16 +00:00
|
|
|
backprop(gradients, sgd=optimizer)
|
2018-11-28 17:04:58 +00:00
|
|
|
# Don't want to return a cupy object here
|
|
|
|
# The gradients are modified in-place by the BERT MLM,
|
|
|
|
# so we get an accurate loss
|
2018-12-18 18:19:26 +00:00
|
|
|
return float(loss)
|
2018-11-15 21:17:16 +00:00
|
|
|
|
|
|
|
|
2019-03-16 20:38:45 +00:00
|
|
|
def make_docs(nlp, batch, min_length, max_length):
|
2018-11-28 17:04:58 +00:00
|
|
|
docs = []
|
|
|
|
for record in batch:
|
|
|
|
text = record["text"]
|
|
|
|
if "tokens" in record:
|
|
|
|
doc = Doc(nlp.vocab, words=record["tokens"])
|
|
|
|
else:
|
|
|
|
doc = nlp.make_doc(text)
|
|
|
|
if "heads" in record:
|
|
|
|
heads = record["heads"]
|
|
|
|
heads = numpy.asarray(heads, dtype="uint64")
|
|
|
|
heads = heads.reshape((len(doc), 1))
|
|
|
|
doc = doc.from_array([HEAD], heads)
|
2018-11-30 20:58:18 +00:00
|
|
|
if len(doc) >= min_length and len(doc) < max_length:
|
2018-11-28 17:04:58 +00:00
|
|
|
docs.append(doc)
|
|
|
|
return docs
|
|
|
|
|
|
|
|
|
2019-02-08 13:14:49 +00:00
|
|
|
def get_vectors_loss(ops, docs, prediction, objective="L2"):
|
2018-11-15 21:17:16 +00:00
|
|
|
"""Compute a mean-squared error loss between the documents' vectors and
|
2018-11-30 19:16:14 +00:00
|
|
|
the prediction.
|
2018-11-15 21:17:16 +00:00
|
|
|
|
|
|
|
Note that this is ripe for customization! We could compute the vectors
|
|
|
|
in some other word, e.g. with an LSTM language model, or use some other
|
|
|
|
type of objective.
|
|
|
|
"""
|
|
|
|
# The simplest way to implement this would be to vstack the
|
|
|
|
# token.vector values, but that's a bit inefficient, especially on GPU.
|
|
|
|
# Instead we fetch the index into the vectors table for each of our tokens,
|
|
|
|
# and look them up all at once. This prevents data copying.
|
|
|
|
ids = ops.flatten([doc.to_array(ID).ravel() for doc in docs])
|
|
|
|
target = docs[0].vocab.vectors.data[ids]
|
2019-02-08 13:14:49 +00:00
|
|
|
if objective == "L2":
|
2018-12-17 14:48:27 +00:00
|
|
|
d_scores = prediction - target
|
2019-02-08 13:14:49 +00:00
|
|
|
loss = (d_scores ** 2).sum()
|
2018-12-17 14:48:27 +00:00
|
|
|
else:
|
2018-12-18 18:19:26 +00:00
|
|
|
raise NotImplementedError(objective)
|
|
|
|
return loss, d_scores
|
2018-12-17 14:48:27 +00:00
|
|
|
|
|
|
|
|
2018-12-18 18:19:26 +00:00
|
|
|
def create_pretraining_model(nlp, tok2vec):
|
2018-11-30 19:16:14 +00:00
|
|
|
"""Define a network for the pretraining. We simply add an output layer onto
|
2018-11-15 21:17:16 +00:00
|
|
|
the tok2vec input model. The tok2vec input model needs to be a model that
|
|
|
|
takes a batch of Doc objects (as a list), and returns a list of arrays.
|
|
|
|
Each array in the output needs to have one row per token in the doc.
|
2018-11-30 19:16:14 +00:00
|
|
|
"""
|
2018-11-15 21:17:16 +00:00
|
|
|
output_size = nlp.vocab.vectors.data.shape[1]
|
2018-11-29 12:36:43 +00:00
|
|
|
output_layer = chain(
|
2019-02-08 13:14:49 +00:00
|
|
|
LN(Maxout(300, pieces=3)), Affine(output_size, drop_factor=0.0)
|
2018-11-29 12:36:43 +00:00
|
|
|
)
|
2018-11-15 23:34:35 +00:00
|
|
|
# This is annoying, but the parser etc have the flatten step after
|
|
|
|
# the tok2vec. To load the weights in cleanly, we need to match
|
|
|
|
# the shape of the models' components exactly. So what we cann
|
|
|
|
# "tok2vec" has to be the same set of processes as what the components do.
|
|
|
|
tok2vec = chain(tok2vec, flatten)
|
2018-11-30 19:16:14 +00:00
|
|
|
model = chain(tok2vec, output_layer)
|
2018-11-28 17:04:58 +00:00
|
|
|
model = masked_language_model(nlp.vocab, model)
|
2018-11-15 23:34:35 +00:00
|
|
|
model.tok2vec = tok2vec
|
2018-11-15 21:17:16 +00:00
|
|
|
model.output_layer = output_layer
|
2018-11-30 19:16:14 +00:00
|
|
|
model.begin_training([nlp.make_doc("Give it a doc to infer shapes")])
|
2018-11-15 21:17:16 +00:00
|
|
|
return model
|
|
|
|
|
|
|
|
|
|
|
|
class ProgressTracker(object):
|
2018-11-29 12:36:43 +00:00
|
|
|
def __init__(self, frequency=1000000):
|
2018-11-28 17:04:58 +00:00
|
|
|
self.loss = 0.0
|
|
|
|
self.prev_loss = 0.0
|
2018-11-15 21:17:16 +00:00
|
|
|
self.nr_word = 0
|
2018-11-15 22:44:07 +00:00
|
|
|
self.words_per_epoch = Counter()
|
2018-11-15 21:17:16 +00:00
|
|
|
self.frequency = frequency
|
|
|
|
self.last_time = time.time()
|
|
|
|
self.last_update = 0
|
2018-11-29 12:36:43 +00:00
|
|
|
self.epoch_loss = 0.0
|
2018-11-15 21:17:16 +00:00
|
|
|
|
|
|
|
def update(self, epoch, loss, docs):
|
|
|
|
self.loss += loss
|
2018-11-29 12:36:43 +00:00
|
|
|
self.epoch_loss += loss
|
2018-11-15 22:44:07 +00:00
|
|
|
words_in_batch = sum(len(doc) for doc in docs)
|
|
|
|
self.words_per_epoch[epoch] += words_in_batch
|
|
|
|
self.nr_word += words_in_batch
|
2018-11-15 21:17:16 +00:00
|
|
|
words_since_update = self.nr_word - self.last_update
|
|
|
|
if words_since_update >= self.frequency:
|
|
|
|
wps = words_since_update / (time.time() - self.last_time)
|
|
|
|
self.last_update = self.nr_word
|
|
|
|
self.last_time = time.time()
|
2018-11-28 17:04:58 +00:00
|
|
|
loss_per_word = self.loss - self.prev_loss
|
|
|
|
status = (
|
|
|
|
epoch,
|
|
|
|
self.nr_word,
|
2019-03-16 20:38:45 +00:00
|
|
|
_smart_round(self.loss, width=10),
|
|
|
|
_smart_round(loss_per_word, width=6),
|
2018-11-28 17:04:58 +00:00
|
|
|
int(wps),
|
|
|
|
)
|
|
|
|
self.prev_loss = float(self.loss)
|
2018-11-15 21:17:16 +00:00
|
|
|
return status
|
|
|
|
else:
|
|
|
|
return None
|
2019-03-16 20:38:45 +00:00
|
|
|
|
|
|
|
|
|
|
|
def _smart_round(figure, width=10, max_decimal=4):
|
|
|
|
"""Round large numbers as integers, smaller numbers as decimals."""
|
|
|
|
n_digits = len(str(int(figure)))
|
|
|
|
n_decimal = width - (n_digits + 1)
|
|
|
|
if n_decimal <= 1:
|
|
|
|
return str(int(figure))
|
|
|
|
else:
|
|
|
|
n_decimal = min(n_decimal, max_decimal)
|
|
|
|
format_str = "%." + str(n_decimal) + "f"
|
|
|
|
return format_str % figure
|