2019-02-07 08:42:25 +00:00
|
|
|
# cython: infer_types=True
|
|
|
|
# cython: profile=True
|
|
|
|
from __future__ import unicode_literals
|
|
|
|
|
|
|
|
from cymem.cymem cimport Pool
|
|
|
|
from preshed.maps cimport PreshMap
|
|
|
|
|
|
|
|
from .matcher cimport Matcher
|
|
|
|
from ..vocab cimport Vocab
|
|
|
|
from ..tokens.doc cimport Doc
|
|
|
|
|
|
|
|
from .matcher import unpickle_matcher
|
|
|
|
from ..errors import Errors
|
|
|
|
|
2019-06-16 11:25:32 +00:00
|
|
|
from libcpp cimport bool
|
|
|
|
import numpy
|
2019-02-07 08:42:25 +00:00
|
|
|
|
2019-03-08 10:42:26 +00:00
|
|
|
DELIMITER = "||"
|
2019-02-07 08:42:25 +00:00
|
|
|
INDEX_HEAD = 1
|
|
|
|
INDEX_RELOP = 0
|
|
|
|
|
|
|
|
|
2019-06-16 11:25:32 +00:00
|
|
|
cdef class DependencyMatcher:
|
2019-02-07 08:42:25 +00:00
|
|
|
"""Match dependency parse tree based on pattern rules."""
|
|
|
|
cdef Pool mem
|
|
|
|
cdef readonly Vocab vocab
|
|
|
|
cdef readonly Matcher token_matcher
|
|
|
|
cdef public object _patterns
|
|
|
|
cdef public object _keys_to_token
|
|
|
|
cdef public object _root
|
|
|
|
cdef public object _entities
|
|
|
|
cdef public object _callbacks
|
|
|
|
cdef public object _nodes
|
|
|
|
cdef public object _tree
|
|
|
|
|
|
|
|
def __init__(self, vocab):
|
2019-06-16 11:25:32 +00:00
|
|
|
"""Create the DependencyMatcher.
|
2019-02-07 08:42:25 +00:00
|
|
|
|
|
|
|
vocab (Vocab): The vocabulary object, which must be shared with the
|
|
|
|
documents the matcher will operate on.
|
2019-06-16 11:25:32 +00:00
|
|
|
RETURNS (DependencyMatcher): The newly constructed object.
|
2019-02-07 08:42:25 +00:00
|
|
|
"""
|
|
|
|
size = 20
|
|
|
|
self.token_matcher = Matcher(vocab)
|
|
|
|
self._keys_to_token = {}
|
|
|
|
self._patterns = {}
|
|
|
|
self._root = {}
|
|
|
|
self._nodes = {}
|
|
|
|
self._tree = {}
|
|
|
|
self._entities = {}
|
|
|
|
self._callbacks = {}
|
|
|
|
self.vocab = vocab
|
|
|
|
self.mem = Pool()
|
|
|
|
|
|
|
|
def __reduce__(self):
|
|
|
|
data = (self.vocab, self._patterns,self._tree, self._callbacks)
|
|
|
|
return (unpickle_matcher, data, None, None)
|
|
|
|
|
|
|
|
def __len__(self):
|
2019-03-08 10:42:26 +00:00
|
|
|
"""Get the number of rules, which are edges, added to the dependency
|
|
|
|
tree matcher.
|
2019-02-07 08:42:25 +00:00
|
|
|
|
|
|
|
RETURNS (int): The number of rules.
|
|
|
|
"""
|
|
|
|
return len(self._patterns)
|
|
|
|
|
|
|
|
def __contains__(self, key):
|
|
|
|
"""Check whether the matcher contains rules for a match ID.
|
|
|
|
|
|
|
|
key (unicode): The match ID.
|
|
|
|
RETURNS (bool): Whether the matcher contains rules for this match ID.
|
|
|
|
"""
|
|
|
|
return self._normalize_key(key) in self._patterns
|
|
|
|
|
|
|
|
def validateInput(self, pattern, key):
|
|
|
|
idx = 0
|
|
|
|
visitedNodes = {}
|
|
|
|
for relation in pattern:
|
2019-03-08 10:42:26 +00:00
|
|
|
if "PATTERN" not in relation or "SPEC" not in relation:
|
2019-02-07 08:42:25 +00:00
|
|
|
raise ValueError(Errors.E098.format(key=key))
|
|
|
|
if idx == 0:
|
2019-03-08 10:42:26 +00:00
|
|
|
if not(
|
|
|
|
"NODE_NAME" in relation["SPEC"]
|
|
|
|
and "NBOR_RELOP" not in relation["SPEC"]
|
|
|
|
and "NBOR_NAME" not in relation["SPEC"]
|
|
|
|
):
|
2019-02-07 08:42:25 +00:00
|
|
|
raise ValueError(Errors.E099.format(key=key))
|
2019-03-08 10:42:26 +00:00
|
|
|
visitedNodes[relation["SPEC"]["NODE_NAME"]] = True
|
2019-02-07 08:42:25 +00:00
|
|
|
else:
|
2019-03-08 10:42:26 +00:00
|
|
|
if not(
|
|
|
|
"NODE_NAME" in relation["SPEC"]
|
|
|
|
and "NBOR_RELOP" in relation["SPEC"]
|
|
|
|
and "NBOR_NAME" in relation["SPEC"]
|
|
|
|
):
|
2019-02-07 08:42:25 +00:00
|
|
|
raise ValueError(Errors.E100.format(key=key))
|
2019-03-08 10:42:26 +00:00
|
|
|
if (
|
|
|
|
relation["SPEC"]["NODE_NAME"] in visitedNodes
|
|
|
|
or relation["SPEC"]["NBOR_NAME"] not in visitedNodes
|
|
|
|
):
|
2019-02-07 08:42:25 +00:00
|
|
|
raise ValueError(Errors.E101.format(key=key))
|
2019-03-08 10:42:26 +00:00
|
|
|
visitedNodes[relation["SPEC"]["NODE_NAME"]] = True
|
|
|
|
visitedNodes[relation["SPEC"]["NBOR_NAME"]] = True
|
2019-02-07 08:42:25 +00:00
|
|
|
idx = idx + 1
|
|
|
|
|
2019-10-25 20:21:08 +00:00
|
|
|
def add(self, key, patterns, *_patterns, on_match=None):
|
|
|
|
if patterns is None or hasattr(patterns, "__call__"): # old API
|
|
|
|
on_match = patterns
|
|
|
|
patterns = _patterns
|
2019-02-07 08:42:25 +00:00
|
|
|
for pattern in patterns:
|
|
|
|
if len(pattern) == 0:
|
|
|
|
raise ValueError(Errors.E012.format(key=key))
|
|
|
|
self.validateInput(pattern,key)
|
|
|
|
key = self._normalize_key(key)
|
|
|
|
_patterns = []
|
|
|
|
for pattern in patterns:
|
|
|
|
token_patterns = []
|
|
|
|
for i in range(len(pattern)):
|
2019-03-08 10:42:26 +00:00
|
|
|
token_pattern = [pattern[i]["PATTERN"]]
|
2019-02-07 08:42:25 +00:00
|
|
|
token_patterns.append(token_pattern)
|
|
|
|
# self.patterns.append(token_patterns)
|
|
|
|
_patterns.append(token_patterns)
|
|
|
|
self._patterns.setdefault(key, [])
|
|
|
|
self._callbacks[key] = on_match
|
|
|
|
self._patterns[key].extend(_patterns)
|
2019-03-08 10:42:26 +00:00
|
|
|
# Add each node pattern of all the input patterns individually to the
|
|
|
|
# matcher. This enables only a single instance of Matcher to be used.
|
2019-02-07 08:42:25 +00:00
|
|
|
# Multiple adds are required to track each node pattern.
|
|
|
|
_keys_to_token_list = []
|
|
|
|
for i in range(len(_patterns)):
|
|
|
|
_keys_to_token = {}
|
2019-03-08 10:42:26 +00:00
|
|
|
# TODO: Better ways to hash edges in pattern?
|
2019-02-07 08:42:25 +00:00
|
|
|
for j in range(len(_patterns[i])):
|
2019-03-08 10:42:26 +00:00
|
|
|
k = self._normalize_key(unicode(key) + DELIMITER + unicode(i) + DELIMITER + unicode(j))
|
|
|
|
self.token_matcher.add(k, None, _patterns[i][j])
|
2019-02-07 08:42:25 +00:00
|
|
|
_keys_to_token[k] = j
|
|
|
|
_keys_to_token_list.append(_keys_to_token)
|
|
|
|
self._keys_to_token.setdefault(key, [])
|
|
|
|
self._keys_to_token[key].extend(_keys_to_token_list)
|
|
|
|
_nodes_list = []
|
|
|
|
for pattern in patterns:
|
|
|
|
nodes = {}
|
|
|
|
for i in range(len(pattern)):
|
2019-03-08 10:42:26 +00:00
|
|
|
nodes[pattern[i]["SPEC"]["NODE_NAME"]] = i
|
2019-02-07 08:42:25 +00:00
|
|
|
_nodes_list.append(nodes)
|
|
|
|
self._nodes.setdefault(key, [])
|
|
|
|
self._nodes[key].extend(_nodes_list)
|
2019-03-08 10:42:26 +00:00
|
|
|
# Create an object tree to traverse later on. This data structure
|
|
|
|
# enables easy tree pattern match. Doc-Token based tree cannot be
|
|
|
|
# reused since it is memory-heavy and tightly coupled with the Doc.
|
|
|
|
self.retrieve_tree(patterns, _nodes_list,key)
|
2019-02-07 08:42:25 +00:00
|
|
|
|
2019-03-08 10:42:26 +00:00
|
|
|
def retrieve_tree(self, patterns, _nodes_list, key):
|
2019-02-07 08:42:25 +00:00
|
|
|
_heads_list = []
|
|
|
|
_root_list = []
|
|
|
|
for i in range(len(patterns)):
|
|
|
|
heads = {}
|
|
|
|
root = -1
|
|
|
|
for j in range(len(patterns[i])):
|
|
|
|
token_pattern = patterns[i][j]
|
2019-03-08 10:42:26 +00:00
|
|
|
if ("NBOR_RELOP" not in token_pattern["SPEC"]):
|
|
|
|
heads[j] = ('root', j)
|
2019-02-07 08:42:25 +00:00
|
|
|
root = j
|
|
|
|
else:
|
2019-03-08 10:42:26 +00:00
|
|
|
heads[j] = (
|
|
|
|
token_pattern["SPEC"]["NBOR_RELOP"],
|
|
|
|
_nodes_list[i][token_pattern["SPEC"]["NBOR_NAME"]]
|
|
|
|
)
|
2019-02-07 08:42:25 +00:00
|
|
|
_heads_list.append(heads)
|
|
|
|
_root_list.append(root)
|
|
|
|
_tree_list = []
|
|
|
|
for i in range(len(patterns)):
|
|
|
|
tree = {}
|
|
|
|
for j in range(len(patterns[i])):
|
|
|
|
if(_heads_list[i][j][INDEX_HEAD] == j):
|
|
|
|
continue
|
|
|
|
head = _heads_list[i][j][INDEX_HEAD]
|
|
|
|
if(head not in tree):
|
|
|
|
tree[head] = []
|
2019-03-08 10:42:26 +00:00
|
|
|
tree[head].append((_heads_list[i][j][INDEX_RELOP], j))
|
2019-02-07 08:42:25 +00:00
|
|
|
_tree_list.append(tree)
|
|
|
|
self._tree.setdefault(key, [])
|
|
|
|
self._tree[key].extend(_tree_list)
|
|
|
|
self._root.setdefault(key, [])
|
|
|
|
self._root[key].extend(_root_list)
|
|
|
|
|
|
|
|
def has_key(self, key):
|
|
|
|
"""Check whether the matcher has a rule with a given key.
|
|
|
|
|
|
|
|
key (string or int): The key to check.
|
|
|
|
RETURNS (bool): Whether the matcher has the rule.
|
|
|
|
"""
|
|
|
|
key = self._normalize_key(key)
|
|
|
|
return key in self._patterns
|
|
|
|
|
|
|
|
def get(self, key, default=None):
|
|
|
|
"""Retrieve the pattern stored for a key.
|
|
|
|
|
|
|
|
key (unicode or int): The key to retrieve.
|
|
|
|
RETURNS (tuple): The rule, as an (on_match, patterns) tuple.
|
|
|
|
"""
|
|
|
|
key = self._normalize_key(key)
|
|
|
|
if key not in self._patterns:
|
|
|
|
return default
|
|
|
|
return (self._callbacks[key], self._patterns[key])
|
|
|
|
|
|
|
|
def __call__(self, Doc doc):
|
2019-06-16 11:25:32 +00:00
|
|
|
matched_key_trees = []
|
2019-02-07 08:42:25 +00:00
|
|
|
matches = self.token_matcher(doc)
|
|
|
|
for key in list(self._patterns.keys()):
|
|
|
|
_patterns_list = self._patterns[key]
|
|
|
|
_keys_to_token_list = self._keys_to_token[key]
|
|
|
|
_root_list = self._root[key]
|
|
|
|
_tree_list = self._tree[key]
|
|
|
|
_nodes_list = self._nodes[key]
|
|
|
|
length = len(_patterns_list)
|
|
|
|
for i in range(length):
|
|
|
|
_keys_to_token = _keys_to_token_list[i]
|
|
|
|
_root = _root_list[i]
|
|
|
|
_tree = _tree_list[i]
|
|
|
|
_nodes = _nodes_list[i]
|
|
|
|
id_to_position = {}
|
|
|
|
for i in range(len(_nodes)):
|
|
|
|
id_to_position[i]=[]
|
2019-03-08 10:42:26 +00:00
|
|
|
# TODO: This could be taken outside to improve running time..?
|
2019-02-07 08:42:25 +00:00
|
|
|
for match_id, start, end in matches:
|
|
|
|
if match_id in _keys_to_token:
|
|
|
|
id_to_position[_keys_to_token[match_id]].append(start)
|
2019-03-08 10:42:26 +00:00
|
|
|
_node_operator_map = self.get_node_operator_map(
|
|
|
|
doc,
|
|
|
|
_tree,
|
|
|
|
id_to_position,
|
|
|
|
_nodes,_root
|
|
|
|
)
|
2019-02-07 08:42:25 +00:00
|
|
|
length = len(_nodes)
|
2019-06-16 11:25:32 +00:00
|
|
|
|
|
|
|
matched_trees = []
|
|
|
|
self.recurse(_tree,id_to_position,_node_operator_map,0,[],matched_trees)
|
2020-11-05 08:16:26 +00:00
|
|
|
if len(matched_trees) > 0:
|
|
|
|
matched_key_trees.append((key,matched_trees))
|
|
|
|
for i, (ent_id, nodes) in enumerate(matched_key_trees):
|
|
|
|
on_match = self._callbacks.get(ent_id)
|
|
|
|
if on_match is not None:
|
|
|
|
on_match(self, doc, i, matched_key_trees)
|
2019-06-16 11:25:32 +00:00
|
|
|
return matched_key_trees
|
|
|
|
|
|
|
|
def recurse(self,tree,id_to_position,_node_operator_map,int patternLength,visitedNodes,matched_trees):
|
|
|
|
cdef bool isValid;
|
|
|
|
if(patternLength == len(id_to_position.keys())):
|
|
|
|
isValid = True
|
|
|
|
for node in range(patternLength):
|
|
|
|
if(node in tree):
|
|
|
|
for idx, (relop,nbor) in enumerate(tree[node]):
|
|
|
|
computed_nbors = numpy.asarray(_node_operator_map[visitedNodes[node]][relop])
|
|
|
|
isNbor = False
|
|
|
|
for computed_nbor in computed_nbors:
|
|
|
|
if(computed_nbor.i == visitedNodes[nbor]):
|
|
|
|
isNbor = True
|
|
|
|
isValid = isValid & isNbor
|
|
|
|
if(isValid):
|
|
|
|
matched_trees.append(visitedNodes)
|
|
|
|
return
|
|
|
|
allPatternNodes = numpy.asarray(id_to_position[patternLength])
|
|
|
|
for patternNode in allPatternNodes:
|
|
|
|
self.recurse(tree,id_to_position,_node_operator_map,patternLength+1,visitedNodes+[patternNode],matched_trees)
|
2019-02-07 08:42:25 +00:00
|
|
|
|
|
|
|
# Given a node and an edge operator, to return the list of nodes
|
|
|
|
# from the doc that belong to node+operator. This is used to store
|
|
|
|
# all the results beforehand to prevent unnecessary computation while
|
|
|
|
# pattern matching
|
|
|
|
# _node_operator_map[node][operator] = [...]
|
|
|
|
def get_node_operator_map(self,doc,tree,id_to_position,nodes,root):
|
|
|
|
_node_operator_map = {}
|
|
|
|
all_node_indices = nodes.values()
|
|
|
|
all_operators = []
|
|
|
|
for node in all_node_indices:
|
|
|
|
if node in tree:
|
|
|
|
for child in tree[node]:
|
|
|
|
all_operators.append(child[INDEX_RELOP])
|
|
|
|
all_operators = list(set(all_operators))
|
|
|
|
all_nodes = []
|
|
|
|
for node in all_node_indices:
|
|
|
|
all_nodes = all_nodes + id_to_position[node]
|
|
|
|
all_nodes = list(set(all_nodes))
|
|
|
|
for node in all_nodes:
|
|
|
|
_node_operator_map[node] = {}
|
|
|
|
for operator in all_operators:
|
|
|
|
_node_operator_map[node][operator] = []
|
|
|
|
# Used to invoke methods for each operator
|
|
|
|
switcher = {
|
2019-03-08 10:42:26 +00:00
|
|
|
"<": self.dep,
|
|
|
|
">": self.gov,
|
2019-06-16 11:25:32 +00:00
|
|
|
"<<": self.dep_chain,
|
|
|
|
">>": self.gov_chain,
|
2019-03-08 10:42:26 +00:00
|
|
|
".": self.imm_precede,
|
|
|
|
"$+": self.imm_right_sib,
|
|
|
|
"$-": self.imm_left_sib,
|
|
|
|
"$++": self.right_sib,
|
|
|
|
"$--": self.left_sib
|
2019-02-07 08:42:25 +00:00
|
|
|
}
|
|
|
|
for operator in all_operators:
|
|
|
|
for node in all_nodes:
|
|
|
|
_node_operator_map[node][operator] = switcher.get(operator)(doc,node)
|
|
|
|
return _node_operator_map
|
|
|
|
|
2019-03-08 10:42:26 +00:00
|
|
|
def dep(self, doc, node):
|
2019-06-16 11:25:32 +00:00
|
|
|
return [doc[node].head]
|
2019-02-07 08:42:25 +00:00
|
|
|
|
|
|
|
def gov(self,doc,node):
|
|
|
|
return list(doc[node].children)
|
|
|
|
|
2019-03-08 10:42:26 +00:00
|
|
|
def dep_chain(self, doc, node):
|
2019-02-07 08:42:25 +00:00
|
|
|
return list(doc[node].ancestors)
|
|
|
|
|
2019-03-08 10:42:26 +00:00
|
|
|
def gov_chain(self, doc, node):
|
2019-02-07 08:42:25 +00:00
|
|
|
return list(doc[node].subtree)
|
|
|
|
|
2019-03-08 10:42:26 +00:00
|
|
|
def imm_precede(self, doc, node):
|
|
|
|
if node > 0:
|
|
|
|
return [doc[node - 1]]
|
2019-02-07 08:42:25 +00:00
|
|
|
return []
|
|
|
|
|
2019-03-08 10:42:26 +00:00
|
|
|
def imm_right_sib(self, doc, node):
|
2019-06-16 11:25:32 +00:00
|
|
|
for child in list(doc[node].head.children):
|
|
|
|
if child.i == node - 1:
|
|
|
|
return [doc[child.i]]
|
2019-02-07 08:42:25 +00:00
|
|
|
return []
|
|
|
|
|
2019-03-08 10:42:26 +00:00
|
|
|
def imm_left_sib(self, doc, node):
|
2019-06-16 11:25:32 +00:00
|
|
|
for child in list(doc[node].head.children):
|
|
|
|
if child.i == node + 1:
|
|
|
|
return [doc[child.i]]
|
2019-02-07 08:42:25 +00:00
|
|
|
return []
|
|
|
|
|
2019-03-08 10:42:26 +00:00
|
|
|
def right_sib(self, doc, node):
|
2019-02-07 08:42:25 +00:00
|
|
|
candidate_children = []
|
2019-06-16 11:25:32 +00:00
|
|
|
for child in list(doc[node].head.children):
|
|
|
|
if child.i < node:
|
|
|
|
candidate_children.append(doc[child.i])
|
2019-02-07 08:42:25 +00:00
|
|
|
return candidate_children
|
|
|
|
|
2019-03-08 10:42:26 +00:00
|
|
|
def left_sib(self, doc, node):
|
2019-02-07 08:42:25 +00:00
|
|
|
candidate_children = []
|
2019-06-16 11:25:32 +00:00
|
|
|
for child in list(doc[node].head.children):
|
|
|
|
if child.i > node:
|
|
|
|
candidate_children.append(doc[child.i])
|
2019-02-07 08:42:25 +00:00
|
|
|
return candidate_children
|
|
|
|
|
|
|
|
def _normalize_key(self, key):
|
|
|
|
if isinstance(key, basestring):
|
|
|
|
return self.vocab.strings.add(key)
|
|
|
|
else:
|
|
|
|
return key
|