spaCy/spacy/util.py

109 lines
3.2 KiB
Python
Raw Normal View History

2014-09-25 16:26:22 +00:00
from os import path
import codecs
2015-01-05 06:54:13 +00:00
import json
2014-09-25 16:26:22 +00:00
import re
DATA_DIR = path.join(path.dirname(__file__), '..', 'data')
def utf8open(loc, mode='r'):
return codecs.open(loc, mode, 'utf8')
2014-12-20 18:36:29 +00:00
def read_lang_data(data_dir):
with open(path.join(data_dir, 'specials.json')) as file_:
2015-01-05 06:54:13 +00:00
tokenization = json.load(file_)
2014-09-25 16:26:22 +00:00
prefix = read_prefix(data_dir)
suffix = read_suffix(data_dir)
infix = read_infix(data_dir)
2014-10-30 07:14:42 +00:00
return tokenization, prefix, suffix, infix
2014-09-25 16:26:22 +00:00
def read_prefix(data_dir):
with utf8open(path.join(data_dir, 'prefix.txt')) as file_:
2014-09-25 16:26:22 +00:00
entries = file_.read().split('\n')
expression = '|'.join(['^' + re.escape(piece) for piece in entries if piece.strip()])
2014-09-25 16:26:22 +00:00
return expression
2014-09-25 16:26:22 +00:00
def read_suffix(data_dir):
with utf8open(path.join(data_dir, 'suffix.txt')) as file_:
2014-09-25 16:26:22 +00:00
entries = file_.read().split('\n')
expression = '|'.join([piece + '$' for piece in entries if piece.strip()])
2014-09-25 16:26:22 +00:00
return expression
def read_infix(data_dir):
with utf8open(path.join(data_dir, 'infix.txt')) as file_:
entries = file_.read().split('\n')
expression = '|'.join([piece for piece in entries if piece.strip()])
return expression
2014-09-25 16:26:22 +00:00
def read_tokenization(lang):
loc = path.join(DATA_DIR, lang, 'tokenization')
entries = []
seen = set()
with utf8open(loc) as file_:
for line in file_:
line = line.strip()
if line.startswith('#'):
continue
if not line:
continue
pieces = line.split()
chunk = pieces.pop(0)
assert chunk not in seen, chunk
seen.add(chunk)
entries.append((chunk, list(pieces)))
if chunk[0].isalpha() and chunk[0].islower():
chunk = chunk[0].title() + chunk[1:]
pieces[0] = pieces[0][0].title() + pieces[0][1:]
seen.add(chunk)
entries.append((chunk, pieces))
return entries
def read_detoken_rules(lang):
loc = path.join(DATA_DIR, lang, 'detokenize')
entries = []
with utf8open(loc) as file_:
for line in file_:
entries.append(line.strip())
return entries
2015-04-19 08:31:31 +00:00
def align_tokens(ref, indices):
start = 0
queue = list(indices)
for token in ref:
end = start + len(token)
emit = []
while queue and queue[0][1] <= end:
emit.append(queue.pop(0))
yield token, emit
start = end
assert not queue
2014-10-18 07:02:05 +00:00
def detokenize(token_rules, words):
2015-04-19 08:31:31 +00:00
"""To align with treebanks, return a list of "chunks", where a chunk is a
2014-10-18 07:02:05 +00:00
sequence of tokens that are separated by whitespace in actual strings. Each
chunk should be a tuple of token indices, e.g.
>>> detokenize(["ca<SEP>n't", '<SEP>!'], ["I", "ca", "n't", "!"])
[(0,), (1, 2, 3)]
"""
string = ' '.join(words)
for subtoks in token_rules:
# Algorithmically this is dumb, but writing a little list-based match
# machine? Ain't nobody got time for that.
string = string.replace(subtoks.replace('<SEP>', ' '), subtoks)
positions = []
i = 0
for chunk in string.split():
subtoks = chunk.split('<SEP>')
positions.append(tuple(range(i, i+len(subtoks))))
i += len(subtoks)
return positions