2018-09-24 21:14:06 +00:00
|
|
|
from __future__ import unicode_literals
|
|
|
|
from collections import OrderedDict, defaultdict
|
|
|
|
|
|
|
|
import numpy
|
|
|
|
cimport numpy as np
|
|
|
|
|
|
|
|
from thinc.api import chain
|
2018-09-25 20:12:54 +00:00
|
|
|
from thinc.neural.util import to_categorical, copy_array, get_array_module
|
2018-09-24 21:14:06 +00:00
|
|
|
from . import util
|
2018-09-25 20:12:54 +00:00
|
|
|
from .pipeline import Pipe
|
2018-09-25 08:58:13 +00:00
|
|
|
from ._ml import Tok2Vec, build_morphologizer_model
|
2018-09-24 21:14:06 +00:00
|
|
|
from ._ml import link_vectors_to_models, zero_init, flatten
|
|
|
|
from ._ml import create_default_optimizer
|
|
|
|
from .errors import Errors, TempErrors
|
2019-03-06 23:16:51 +00:00
|
|
|
from .compat import basestring_
|
2018-09-24 21:14:06 +00:00
|
|
|
from .tokens.doc cimport Doc
|
|
|
|
from .vocab cimport Vocab
|
|
|
|
from .morphology cimport Morphology
|
2018-09-26 19:02:13 +00:00
|
|
|
from .morphology import parse_feature, IDS, FIELDS, FIELD_SIZES, NAMES
|
2018-09-25 08:58:13 +00:00
|
|
|
from .pipeline import Pipe
|
2018-09-24 21:14:06 +00:00
|
|
|
|
|
|
|
|
|
|
|
class Morphologizer(Pipe):
|
|
|
|
name = 'morphologizer'
|
|
|
|
|
|
|
|
@classmethod
|
2018-09-26 19:02:13 +00:00
|
|
|
def Model(cls, attr_nums=None, **cfg):
|
2018-09-24 21:14:06 +00:00
|
|
|
if cfg.get('pretrained_dims') and not cfg.get('pretrained_vectors'):
|
|
|
|
raise ValueError(TempErrors.T008)
|
2018-09-26 19:02:13 +00:00
|
|
|
if attr_nums is None:
|
|
|
|
attr_nums = list(FIELD_SIZES)
|
2018-09-24 21:14:06 +00:00
|
|
|
return build_morphologizer_model(attr_nums, **cfg)
|
|
|
|
|
|
|
|
def __init__(self, vocab, model=True, **cfg):
|
|
|
|
self.vocab = vocab
|
|
|
|
self.model = model
|
|
|
|
self.cfg = OrderedDict(sorted(cfg.items()))
|
|
|
|
self.cfg.setdefault('cnn_maxout_pieces', 2)
|
|
|
|
|
|
|
|
@property
|
|
|
|
def labels(self):
|
|
|
|
return self.vocab.morphology.tag_names
|
|
|
|
|
|
|
|
@property
|
|
|
|
def tok2vec(self):
|
|
|
|
if self.model in (None, True, False):
|
|
|
|
return None
|
|
|
|
else:
|
|
|
|
return chain(self.model.tok2vec, flatten)
|
|
|
|
|
|
|
|
def __call__(self, doc):
|
|
|
|
features, tokvecs = self.predict([doc])
|
2018-09-25 08:58:13 +00:00
|
|
|
self.set_annotations([doc], features, tensors=tokvecs)
|
2018-09-24 21:14:06 +00:00
|
|
|
return doc
|
|
|
|
|
|
|
|
def pipe(self, stream, batch_size=128, n_threads=-1):
|
2019-03-06 23:16:51 +00:00
|
|
|
for docs in util.minibatch(stream, size=batch_size):
|
2018-09-24 21:14:06 +00:00
|
|
|
docs = list(docs)
|
|
|
|
features, tokvecs = self.predict(docs)
|
|
|
|
self.set_annotations(docs, features, tensors=tokvecs)
|
|
|
|
yield from docs
|
|
|
|
|
|
|
|
def predict(self, docs):
|
|
|
|
if not any(len(doc) for doc in docs):
|
|
|
|
# Handle case where there are no tokens in any docs.
|
|
|
|
n_labels = self.model.nO
|
|
|
|
guesses = [self.model.ops.allocate((0, n_labels)) for doc in docs]
|
|
|
|
tokvecs = self.model.ops.allocate((0, self.model.tok2vec.nO))
|
|
|
|
return guesses, tokvecs
|
|
|
|
tokvecs = self.model.tok2vec(docs)
|
|
|
|
scores = self.model.softmax(tokvecs)
|
2018-09-26 19:02:13 +00:00
|
|
|
return scores, tokvecs
|
2018-09-24 21:14:06 +00:00
|
|
|
|
2018-09-26 19:02:13 +00:00
|
|
|
def set_annotations(self, docs, batch_scores, tensors=None):
|
2018-09-24 21:14:06 +00:00
|
|
|
if isinstance(docs, Doc):
|
|
|
|
docs = [docs]
|
|
|
|
cdef Doc doc
|
|
|
|
cdef Vocab vocab = self.vocab
|
2018-09-26 19:02:13 +00:00
|
|
|
field_names = list(FIELDS)
|
|
|
|
offsets = [IDS['begin_%s' % field] for field in field_names]
|
2018-09-24 21:14:06 +00:00
|
|
|
for i, doc in enumerate(docs):
|
2018-09-26 19:02:13 +00:00
|
|
|
doc_scores = batch_scores[i]
|
|
|
|
doc_guesses = scores_to_guesses(doc_scores, self.model.softmax.out_sizes)
|
2018-09-24 21:14:06 +00:00
|
|
|
# Convert the neuron indices into feature IDs.
|
2018-09-26 19:02:13 +00:00
|
|
|
doc_feat_ids = self.model.ops.allocate((len(doc), len(field_names)), dtype='i')
|
|
|
|
for j in range(len(doc)):
|
|
|
|
for k, offset in enumerate(offsets):
|
|
|
|
if doc_guesses[j, k] == 0:
|
|
|
|
doc_feat_ids[j, k] = 0
|
|
|
|
else:
|
|
|
|
doc_feat_ids[j, k] = offset + doc_guesses[j, k]
|
|
|
|
# Now add the analysis, and set the hash.
|
|
|
|
try:
|
|
|
|
doc.c[j].morph = self.vocab.morphology.add(doc_feat_ids[j])
|
|
|
|
except:
|
|
|
|
print(offsets)
|
|
|
|
print(doc_guesses[j])
|
|
|
|
print(doc_feat_ids[j])
|
|
|
|
raise
|
2018-09-24 21:14:06 +00:00
|
|
|
|
|
|
|
def update(self, docs, golds, drop=0., sgd=None, losses=None):
|
|
|
|
if losses is not None and self.name not in losses:
|
|
|
|
losses[self.name] = 0.
|
|
|
|
|
|
|
|
tag_scores, bp_tag_scores = self.model.begin_update(docs, drop=drop)
|
|
|
|
loss, d_tag_scores = self.get_loss(docs, golds, tag_scores)
|
|
|
|
bp_tag_scores(d_tag_scores, sgd=sgd)
|
|
|
|
|
|
|
|
if losses is not None:
|
|
|
|
losses[self.name] += loss
|
|
|
|
|
|
|
|
def get_loss(self, docs, golds, scores):
|
|
|
|
guesses = []
|
|
|
|
for doc_scores in scores:
|
|
|
|
guesses.append(scores_to_guesses(doc_scores, self.model.softmax.out_sizes))
|
2018-09-26 19:02:13 +00:00
|
|
|
guesses = self.model.ops.xp.vstack(guesses)
|
|
|
|
scores = self.model.ops.xp.vstack(scores)
|
2018-09-24 21:14:06 +00:00
|
|
|
cdef int idx = 0
|
|
|
|
target = numpy.zeros(scores.shape, dtype='f')
|
2018-09-26 19:02:13 +00:00
|
|
|
field_sizes = self.model.softmax.out_sizes
|
2018-09-24 21:14:06 +00:00
|
|
|
for gold in golds:
|
|
|
|
for features in gold.morphology:
|
|
|
|
if features is None:
|
2018-09-26 19:02:13 +00:00
|
|
|
target[idx] = scores[idx]
|
2018-09-24 21:14:06 +00:00
|
|
|
else:
|
2018-09-26 19:02:13 +00:00
|
|
|
by_field = {}
|
2018-09-24 21:14:06 +00:00
|
|
|
for feature in features:
|
2018-09-26 19:02:13 +00:00
|
|
|
field, column = parse_feature(feature)
|
|
|
|
by_field[field] = column
|
|
|
|
col_offset = 0
|
|
|
|
for field, field_size in enumerate(field_sizes):
|
|
|
|
if field in by_field:
|
|
|
|
target[idx, col_offset + by_field[field]] = 1.
|
|
|
|
else:
|
|
|
|
target[idx, col_offset] = 1.
|
|
|
|
col_offset += field_size
|
2018-09-24 21:14:06 +00:00
|
|
|
idx += 1
|
|
|
|
target = self.model.ops.xp.array(target, dtype='f')
|
|
|
|
d_scores = scores - target
|
|
|
|
loss = (d_scores**2).sum()
|
|
|
|
d_scores = self.model.ops.unflatten(d_scores, [len(d) for d in docs])
|
|
|
|
return float(loss), d_scores
|
|
|
|
|
|
|
|
def use_params(self, params):
|
|
|
|
with self.model.use_params(params):
|
|
|
|
yield
|
2018-09-25 08:58:13 +00:00
|
|
|
|
|
|
|
def scores_to_guesses(scores, out_sizes):
|
2018-09-25 20:12:54 +00:00
|
|
|
xp = get_array_module(scores)
|
|
|
|
guesses = xp.zeros((scores.shape[0], len(out_sizes)), dtype='i')
|
|
|
|
offset = 0
|
|
|
|
for i, size in enumerate(out_sizes):
|
2018-09-26 19:02:13 +00:00
|
|
|
slice_ = scores[:, offset : offset + size]
|
|
|
|
col_guesses = slice_.argmax(axis=1)
|
|
|
|
guesses[:, i] = col_guesses
|
2018-09-25 20:12:54 +00:00
|
|
|
offset += size
|
|
|
|
return guesses
|