2017-03-15 16:39:54 +00:00
|
|
|
# coding: utf-8
|
2017-03-07 16:15:18 +00:00
|
|
|
from __future__ import unicode_literals
|
|
|
|
|
2017-03-07 19:58:55 +00:00
|
|
|
import pytest
|
|
|
|
import dill as pickle
|
2017-12-07 08:53:30 +00:00
|
|
|
import numpy
|
2018-07-24 21:38:44 +00:00
|
|
|
from spacy.strings import StringStore
|
|
|
|
from spacy.vocab import Vocab
|
|
|
|
from spacy.attrs import NORM
|
2017-03-07 16:15:18 +00:00
|
|
|
|
|
|
|
|
2017-03-15 16:39:54 +00:00
|
|
|
@pytest.mark.parametrize('text1,text2', [('hello', 'bye')])
|
2018-07-24 21:38:44 +00:00
|
|
|
def test_pickle_string_store(text1, text2):
|
|
|
|
stringstore = StringStore()
|
2017-03-15 16:39:54 +00:00
|
|
|
store1 = stringstore[text1]
|
|
|
|
store2 = stringstore[text2]
|
|
|
|
data = pickle.dumps(stringstore, protocol=-1)
|
|
|
|
unpickled = pickle.loads(data)
|
|
|
|
assert unpickled[text1] == store1
|
|
|
|
assert unpickled[text2] == store2
|
|
|
|
assert len(stringstore) == len(unpickled)
|
2017-03-07 16:15:18 +00:00
|
|
|
|
2017-03-07 19:58:55 +00:00
|
|
|
|
2017-03-15 16:39:54 +00:00
|
|
|
@pytest.mark.parametrize('text1,text2', [('dog', 'cat')])
|
|
|
|
def test_pickle_vocab(text1, text2):
|
2017-03-07 19:58:55 +00:00
|
|
|
vocab = Vocab(lex_attr_getters={int(NORM): lambda string: string[:-1]})
|
2017-12-07 08:53:30 +00:00
|
|
|
vocab.set_vector('dog', numpy.ones((5,), dtype='f'))
|
2017-03-15 16:39:54 +00:00
|
|
|
lex1 = vocab[text1]
|
|
|
|
lex2 = vocab[text2]
|
|
|
|
assert lex1.norm_ == text1[:-1]
|
|
|
|
assert lex2.norm_ == text2[:-1]
|
|
|
|
data = pickle.dumps(vocab)
|
|
|
|
unpickled = pickle.loads(data)
|
|
|
|
assert unpickled[text1].orth == lex1.orth
|
|
|
|
assert unpickled[text2].orth == lex2.orth
|
|
|
|
assert unpickled[text1].norm == lex1.norm
|
|
|
|
assert unpickled[text2].norm == lex2.norm
|
|
|
|
assert unpickled[text1].norm != unpickled[text2].norm
|
2017-12-07 08:53:30 +00:00
|
|
|
assert unpickled.vectors is not None
|
|
|
|
assert list(vocab['dog'].vector) == [1.,1.,1.,1.,1.]
|