* Remove the feature array stuff from Tokens class, and replace vector with array-based implementation, with padding.

This commit is contained in:
Matthew Honnibal 2014-10-23 01:57:59 +11:00
parent ea1d4a81eb
commit e5e951ae67
6 changed files with 150 additions and 117 deletions

View File

@ -59,5 +59,5 @@ cdef class Language:
cdef int _find_prefix(self, Py_UNICODE* characters, size_t length) except -1 cdef int _find_prefix(self, Py_UNICODE* characters, size_t length) except -1
cdef int _find_suffix(self, Py_UNICODE* characters, size_t length) except -1 cdef int _find_suffix(self, Py_UNICODE* characters, size_t length) except -1
cdef int _find_infix(self, Py_UNICODE* characters, size_t length) except -1 cdef int _find_infix(self, Py_UNICODE* characters, size_t length) except -1
cdef int _save_cached(self, vector[LexemeC*] *tokens, uint64_t key, size_t n) except -1 cdef int _save_cached(self, LexemeC** tokens, uint64_t key, int n) except -1

View File

@ -101,10 +101,10 @@ cdef class Language:
cdef uint64_t orig_key cdef uint64_t orig_key
cdef int orig_size cdef int orig_size
orig_key = span.key orig_key = span.key
orig_size = tokens.lex.size() orig_size = tokens.length
self._split_affixes(span, &prefixes, &suffixes) self._split_affixes(span, &prefixes, &suffixes)
self._attach_tokens(tokens, start, span, &prefixes, &suffixes) self._attach_tokens(tokens, start, span, &prefixes, &suffixes)
self._save_cached(tokens.lex, orig_key, orig_size) self._save_cached(&tokens.lex[orig_size], orig_key, tokens.length - orig_size)
cdef String* _split_affixes(self, String* string, vector[LexemeC*] *prefixes, cdef String* _split_affixes(self, String* string, vector[LexemeC*] *prefixes,
vector[LexemeC*] *suffixes) except NULL: vector[LexemeC*] *suffixes) except NULL:
@ -177,12 +177,11 @@ cdef class Language:
idx = tokens.push_back(idx, deref(it)) idx = tokens.push_back(idx, deref(it))
preinc(it) preinc(it)
cdef int _save_cached(self, vector[LexemeC*] *tokens, uint64_t key, size_t n) except -1: cdef int _save_cached(self, LexemeC** tokens, uint64_t key, int n) except -1:
assert tokens.size() > n lexemes = <LexemeC**>self._mem.alloc(n + 1, sizeof(LexemeC**))
lexemes = <LexemeC**>self._mem.alloc((tokens.size() - n) + 1, sizeof(LexemeC**)) cdef int i
cdef size_t i, j for i in range(n):
for i, j in enumerate(range(n, tokens.size())): lexemes[i] = tokens[i]
lexemes[i] = tokens.at(j)
lexemes[i + 1] = NULL lexemes[i + 1] = NULL
self.cache.set(key, lexemes) self.cache.set(key, lexemes)

View File

@ -73,6 +73,9 @@ cdef struct LexemeC:
flag_t dist_flags flag_t dist_flags
cdef LexemeC EMPTY_LEXEME
cpdef dict get_lexeme_dict(size_t i, unicode string) cpdef dict get_lexeme_dict(size_t i, unicode string)
cdef char* intern_and_encode(unicode string, size_t* length) except NULL cdef char* intern_and_encode(unicode string, size_t* length) except NULL

View File

@ -1,10 +1,13 @@
from cpython.ref cimport Py_INCREF from cpython.ref cimport Py_INCREF
from cymem.cymem cimport Pool from cymem.cymem cimport Pool
from libc.string cimport memset
import orth import orth
OOV_DIST_FLAGS = 0 OOV_DIST_FLAGS = 0
memset(&EMPTY_LEXEME, 0, sizeof(LexemeC))
cpdef dict get_lexeme_dict(size_t i, unicode string): cpdef dict get_lexeme_dict(size_t i, unicode string):
ints = [None for _ in range(LexInt_N)] ints = [None for _ in range(LexInt_N)]

View File

@ -1,21 +1,24 @@
from cymem.cymem cimport Pool
from spacy.lexeme cimport LexemeC from spacy.lexeme cimport LexemeC
from libcpp.vector cimport vector
from thinc.typedefs cimport atom_t from thinc.typedefs cimport atom_t
cdef class Tokens: cdef class Tokens:
cdef vector[LexemeC*] *lex cdef Pool mem
cdef vector[int] *idx
cdef vector[int] *pos cdef LexemeC** _lex_ptr
cdef int* _idx_ptr
cdef int* _pos_ptr
cdef LexemeC** lex
cdef int* idx
cdef int* pos
cdef int length
cdef int max_length
cdef int extend(self, int i, LexemeC** lexemes, int n) except -1 cdef int extend(self, int i, LexemeC** lexemes, int n) except -1
cdef int push_back(self, int i, LexemeC* lexeme) except -1 cdef int push_back(self, int i, LexemeC* lexeme) except -1
cdef int int_array(self, atom_t* atoms, int i, int* indices, int n_idx,
int* features, int n_feat) except -1
cdef int string_array(self, atom_t* atoms, int i, int* indices, int n_idx,
int* features, int n_feat) except -1
cdef int bool_array(self, atom_t* atoms, int i, int* indices, int n_idx,
int* features, int n_feat) except -1
cpdef int id(self, size_t i) except -1 cpdef int id(self, size_t i) except -1
cpdef float prob(self, size_t i) except 1 cpdef float prob(self, size_t i) except 1

View File

@ -6,6 +6,14 @@ cimport numpy
cimport cython cimport cython
import numpy import numpy
DEF PADDING = 5
cdef int bounds_check(int i, int length, int padding) except -1:
if (i + padding) < 0:
raise IndexError
if (i - padding) >= length:
raise IndexError
cdef class Tokens: cdef class Tokens:
"""A sequence of references to Lexeme objects. """A sequence of references to Lexeme objects.
@ -26,71 +34,58 @@ cdef class Tokens:
>>> tokens.can_noun(1) >>> tokens.can_noun(1)
True True
""" """
def __cinit__(self, string_length=0): def __init__(self, string_length=0):
size = int(string_length / 3) if string_length >= 3 else 1 if string_length >= 3:
self.lex = new vector[LexemeC*]() size = int(string_length / 3.0)
self.idx = new vector[int]() else:
self.pos = new vector[int]() size = 5
self.lex.reserve(size) self.mem = Pool()
self.idx.reserve(size) # Guarantee self.lex[i-x], for any i >= 0 and x < padding is in bounds
self.pos.reserve(size) # However, we need to remember the true starting places, so that we can
# realloc.
self._lex_ptr = <LexemeC**>self.mem.alloc(size + (PADDING*2), sizeof(LexemeC*))
self._idx_ptr = <int*>self.mem.alloc(size + (PADDING*2), sizeof(int))
self._pos_ptr = <int*>self.mem.alloc(size + (PADDING*2), sizeof(int))
self.lex = self._lex_ptr
self.idx = self._idx_ptr
self.pos = self._pos_ptr
for i in range(PADDING):
self.lex[i] = &EMPTY_LEXEME
for i in range(size, PADDING):
self.lex[i] = &EMPTY_LEXEME
self.lex += PADDING
self.idx += PADDING
self.pos += PADDING
def __dealloc__(self): self.max_length = size
del self.lex self.length = 0
del self.idx
del self.pos
def __getitem__(self, i): def __getitem__(self, i):
if i >= self.lex.size(): bounds_check(i, self.length, PADDING)
raise IndexError return Lexeme(<size_t>self.lex[i])
return Lexeme(<size_t>self.lex.at(i))
def __len__(self): def __len__(self):
return self.lex.size() return self.length
cdef int push_back(self, int idx, LexemeC* lexeme) except -1: cdef int push_back(self, int idx, LexemeC* lexeme) except -1:
self.lex.push_back(lexeme) if self.length == self.max_length:
self.idx.push_back(idx) self._realloc(self.length * 2)
self.lex[self.length] = lexeme
self.idx[self.length] = idx
self.pos[self.length] = 0
self.length += 1
return idx + lexeme.ints[<int>LexInt_length] return idx + lexeme.ints[<int>LexInt_length]
cdef int int_array(self, atom_t* output, int i, int* indices, int n_idx, def _realloc(self, new_size):
int* features, int n_feat): self.max_length = new_size
cdef int feat_id, idx n = new_size + (PADDING * 2)
cdef int length = self.lex.size() self._lex_ptr = <LexemeC**>self.mem.realloc(self._lex_ptr, n * sizeof(LexemeC*))
for feat_id in features[:n_feat]: self._idx_ptr = <int*>self.mem.realloc(self._idx_ptr, n * sizeof(int))
for idx in indices[:n_idx]: self._pos_ptr = <int*>self.mem.realloc(self._pos_ptr, n * sizeof(int))
if idx < 0 or idx >= length: self.lex = self._lex_ptr + PADDING
output[i] = 0 self.idx = self._idx_ptr + PADDING
else: self.pos = self._pos_ptr + PADDING
output[i] = self.lex[0][idx].ints[<int>feat_id]
i += 1
return i
cdef int string_array(self, atom_t* output, int i, int* indices, int n_idx,
int* features, int n_feat):
cdef int feat_id, idx
cdef int length = self.lex.size()
for feat_id in features[:n_feat]:
for idx in indices[:n_idx]:
if idx < 0 or idx >= length:
output[i] = 0
else:
output[i] = <atom_t>self.lex[0][idx].strings[<int>feat_id]
i += 1
return i
cdef int bool_array(self, atom_t* output, int i, int* indices, int n_idx,
int* features, int n_feat):
cdef int feat_id, idx
cdef int length = self.lex.size()
for feat_id in features[:n_feat]:
for idx in indices[:n_idx]:
if idx < 0 or idx >= length:
output[i] = 0
else:
output[i] = lexeme_check_dist_flag(self.lex[0][idx], feat_id)
i += 1
return i
cdef int extend(self, int idx, LexemeC** lexemes, int n) except -1: cdef int extend(self, int idx, LexemeC** lexemes, int n) except -1:
cdef int i cdef int i
@ -99,131 +94,161 @@ cdef class Tokens:
elif n == 0: elif n == 0:
i = 0 i = 0
while lexemes[i] != NULL: while lexemes[i] != NULL:
self.lex.push_back(lexemes[i]) idx = self.push_back(idx, lexemes[i])
self.idx.push_back(idx)
idx += lexemes[i].ints[<int>LexInt_length]
i += 1 i += 1
else: else:
for i in range(n): for i in range(n):
self.lex.push_back(lexemes[i]) idx = self.push_back(idx, lexemes[i])
self.idx.push_back(idx)
idx += lexemes[i].ints[<int>LexInt_length]
return idx return idx
cpdef int id(self, size_t i) except -1: cpdef int id(self, size_t i) except -1:
return self.lex.at(i).ints[<int>LexInt_id] bounds_check(i, self.length, PADDING)
return self.lex[i].ints[<int>LexInt_id]
cpdef float prob(self, size_t i) except 1: cpdef float prob(self, size_t i) except 1:
return self.lex.at(i).floats[<int>LexFloat_prob] bounds_check(i, self.length, PADDING)
return self.lex[i].floats[<int>LexFloat_prob]
cpdef int cluster(self, size_t i) except *: cpdef int cluster(self, size_t i) except *:
return self.lex.at(i).ints[<int>LexInt_cluster] bounds_check(i, self.length, PADDING)
return self.lex[i].ints[<int>LexInt_cluster]
cpdef bint check_orth_flag(self, size_t i, size_t flag_id) except *: cpdef bint check_orth_flag(self, size_t i, size_t flag_id) except *:
return lexeme_check_orth_flag(self.lex.at(i), flag_id) bounds_check(i, self.length, PADDING)
return lexeme_check_orth_flag(self.lex[i], flag_id)
cpdef bint check_dist_flag(self, size_t i, size_t flag_id) except *: cpdef bint check_dist_flag(self, size_t i, size_t flag_id) except *:
return lexeme_check_dist_flag(self.lex.at(i), flag_id) bounds_check(i, self.length, PADDING)
return lexeme_check_dist_flag(self.lex[i], flag_id)
cpdef unicode string_view(self, size_t i, size_t view_id): cpdef unicode string_view(self, size_t i, size_t view_id):
return lexeme_get_string(self.lex.at(i), view_id) bounds_check(i, self.length, PADDING)
return lexeme_get_string(self.lex[i], view_id)
# Provide accessor methods for the features supported by the language. # Provide accessor methods for the features supported by the language.
# Without these, clients have to use the underlying string_view and check_flag # Without these, clients have to use the underlying string_view and check_flag
# methods, which requires them to know the IDs. # methods, which requires them to know the IDs.
cpdef unicode string(self, size_t i): cpdef unicode string(self, size_t i):
if i >= self.lex.size(): bounds_check(i, self.length, PADDING)
raise IndexError
return self.orig(i) return self.orig(i)
cpdef unicode orig(self, size_t i): cpdef unicode orig(self, size_t i):
cdef bytes utf8_string = self.lex.at(i).strings[<int>LexStr_orig] bounds_check(i, self.length, PADDING)
cdef bytes utf8_string = self.lex[i].strings[<int>LexStr_orig]
cdef unicode string = utf8_string.decode('utf8') cdef unicode string = utf8_string.decode('utf8')
return string return string
cpdef unicode norm(self, size_t i): cpdef unicode norm(self, size_t i):
cdef bytes utf8_string = self.lex.at(i).strings[<int>LexStr_norm] bounds_check(i, self.length, PADDING)
cdef bytes utf8_string = self.lex[i].strings[<int>LexStr_norm]
cdef unicode string = utf8_string.decode('utf8') cdef unicode string = utf8_string.decode('utf8')
return string return string
cpdef unicode shape(self, size_t i): cpdef unicode shape(self, size_t i):
return lexeme_get_string(self.lex.at(i), LexStr_shape) bounds_check(i, self.length, PADDING)
return lexeme_get_string(self.lex[i], LexStr_shape)
cpdef unicode unsparse(self, size_t i): cpdef unicode unsparse(self, size_t i):
return lexeme_get_string(self.lex.at(i), LexStr_unsparse) bounds_check(i, self.length, PADDING)
return lexeme_get_string(self.lex[i], LexStr_unsparse)
cpdef unicode asciied(self, size_t i): cpdef unicode asciied(self, size_t i):
return lexeme_get_string(self.lex.at(i), LexStr_asciied) bounds_check(i, self.length, PADDING)
return lexeme_get_string(self.lex[i], LexStr_asciied)
cpdef bint is_alpha(self, size_t i) except *: cpdef bint is_alpha(self, size_t i) except *:
return lexeme_check_orth_flag(self.lex.at(i), LexOrth_alpha) bounds_check(i, self.length, PADDING)
return lexeme_check_orth_flag(self.lex[i], LexOrth_alpha)
cpdef bint is_ascii(self, size_t i) except *: cpdef bint is_ascii(self, size_t i) except *:
return lexeme_check_orth_flag(self.lex.at(i), LexOrth_ascii) bounds_check(i, self.length, PADDING)
return lexeme_check_orth_flag(self.lex[i], LexOrth_ascii)
cpdef bint is_digit(self, size_t i) except *: cpdef bint is_digit(self, size_t i) except *:
return lexeme_check_orth_flag(self.lex.at(i), LexOrth_digit) bounds_check(i, self.length, PADDING)
return lexeme_check_orth_flag(self.lex[i], LexOrth_digit)
cpdef bint is_lower(self, size_t i) except *: cpdef bint is_lower(self, size_t i) except *:
return lexeme_check_orth_flag(self.lex.at(i), LexOrth_lower) bounds_check(i, self.length, PADDING)
return lexeme_check_orth_flag(self.lex[i], LexOrth_lower)
cpdef bint is_punct(self, size_t i) except *: cpdef bint is_punct(self, size_t i) except *:
return lexeme_check_orth_flag(self.lex.at(i), LexOrth_punct) bounds_check(i, self.length, PADDING)
return lexeme_check_orth_flag(self.lex[i], LexOrth_punct)
cpdef bint is_space(self, size_t i) except *: cpdef bint is_space(self, size_t i) except *:
return lexeme_check_orth_flag(self.lex.at(i), LexOrth_space) bounds_check(i, self.length, PADDING)
return lexeme_check_orth_flag(self.lex[i], LexOrth_space)
cpdef bint is_title(self, size_t i) except *: cpdef bint is_title(self, size_t i) except *:
return lexeme_check_orth_flag(self.lex.at(i), LexOrth_title) bounds_check(i, self.length, PADDING)
return lexeme_check_orth_flag(self.lex[i], LexOrth_title)
cpdef bint is_upper(self, size_t i) except *: cpdef bint is_upper(self, size_t i) except *:
return lexeme_check_orth_flag(self.lex.at(i), LexOrth_upper) bounds_check(i, self.length, PADDING)
return lexeme_check_orth_flag(self.lex[i], LexOrth_upper)
cpdef bint can_adj(self, size_t i) except *: cpdef bint can_adj(self, size_t i) except *:
return lexeme_check_dist_flag(self.lex.at(i), LexDist_adj) bounds_check(i, self.length, PADDING)
return lexeme_check_dist_flag(self.lex[i], LexDist_adj)
cpdef bint can_adp(self, size_t i) except *: cpdef bint can_adp(self, size_t i) except *:
return lexeme_check_dist_flag(self.lex.at(i), LexDist_adp) bounds_check(i, self.length, PADDING)
return lexeme_check_dist_flag(self.lex[i], LexDist_adp)
cpdef bint can_adv(self, size_t i) except *: cpdef bint can_adv(self, size_t i) except *:
return lexeme_check_dist_flag(self.lex.at(i), LexDist_adv) bounds_check(i, self.length, PADDING)
return lexeme_check_dist_flag(self.lex[i], LexDist_adv)
cpdef bint can_conj(self, size_t i) except *: cpdef bint can_conj(self, size_t i) except *:
return lexeme_check_dist_flag(self.lex.at(i), LexDist_conj) bounds_check(i, self.length, PADDING)
return lexeme_check_dist_flag(self.lex[i], LexDist_conj)
cpdef bint can_det(self, size_t i) except *: cpdef bint can_det(self, size_t i) except *:
return lexeme_check_dist_flag(self.lex.at(i), LexDist_det) bounds_check(i, self.length, PADDING)
return lexeme_check_dist_flag(self.lex[i], LexDist_det)
cpdef bint can_noun(self, size_t i) except *: cpdef bint can_noun(self, size_t i) except *:
return lexeme_check_dist_flag(self.lex.at(i), LexDist_noun) bounds_check(i, self.length, PADDING)
return lexeme_check_dist_flag(self.lex[i], LexDist_noun)
cpdef bint can_num(self, size_t i) except *: cpdef bint can_num(self, size_t i) except *:
return lexeme_check_dist_flag(self.lex.at(i), LexDist_num) bounds_check(i, self.length, PADDING)
return lexeme_check_dist_flag(self.lex[i], LexDist_num)
cpdef bint can_pdt(self, size_t i) except *: cpdef bint can_pdt(self, size_t i) except *:
return lexeme_check_dist_flag(self.lex.at(i), LexDist_pdt) bounds_check(i, self.length, PADDING)
return lexeme_check_dist_flag(self.lex[i], LexDist_pdt)
cpdef bint can_pos(self, size_t i) except *: cpdef bint can_pos(self, size_t i) except *:
return lexeme_check_dist_flag(self.lex.at(i), LexDist_pos) bounds_check(i, self.length, PADDING)
return lexeme_check_dist_flag(self.lex[i], LexDist_pos)
cpdef bint can_pron(self, size_t i) except *: cpdef bint can_pron(self, size_t i) except *:
return lexeme_check_dist_flag(self.lex.at(i), LexDist_pron) bounds_check(i, self.length, PADDING)
return lexeme_check_dist_flag(self.lex[i], LexDist_pron)
cpdef bint can_prt(self, size_t i) except *: cpdef bint can_prt(self, size_t i) except *:
return lexeme_check_dist_flag(self.lex.at(i), LexDist_prt) bounds_check(i, self.length, PADDING)
return lexeme_check_dist_flag(self.lex[i], LexDist_prt)
cpdef bint can_punct(self, size_t i) except *: cpdef bint can_punct(self, size_t i) except *:
return lexeme_check_dist_flag(self.lex.at(i), LexDist_punct) bounds_check(i, self.length, PADDING)
return lexeme_check_dist_flag(self.lex[i], LexDist_punct)
cpdef bint can_verb(self, size_t i) except *: cpdef bint can_verb(self, size_t i) except *:
return lexeme_check_dist_flag(self.lex.at(i), LexDist_verb) bounds_check(i, self.length, PADDING)
return lexeme_check_dist_flag(self.lex[i], LexDist_verb)
cpdef bint oft_lower(self, size_t i) except *: cpdef bint oft_lower(self, size_t i) except *:
return lexeme_check_dist_flag(self.lex.at(i), LexDist_lower) bounds_check(i, self.length, PADDING)
return lexeme_check_dist_flag(self.lex[i], LexDist_lower)
cpdef bint oft_title(self, size_t i) except *: cpdef bint oft_title(self, size_t i) except *:
return lexeme_check_dist_flag(self.lex.at(i), LexDist_title) bounds_check(i, self.length, PADDING)
return lexeme_check_dist_flag(self.lex[i], LexDist_title)
cpdef bint oft_upper(self, size_t i) except *: cpdef bint oft_upper(self, size_t i) except *:
return lexeme_check_dist_flag(self.lex.at(i), LexDist_upper) bounds_check(i, self.length, PADDING)
return lexeme_check_dist_flag(self.lex[i], LexDist_upper)