mirror of https://github.com/explosion/spaCy.git
Unset LayerNorm backwards compat hack
This commit is contained in:
parent
73ac0aa0b5
commit
af75b74208
|
@ -31,8 +31,8 @@ import numpy
|
||||||
import io
|
import io
|
||||||
|
|
||||||
# TODO: Unset this once we don't want to support models previous models.
|
# TODO: Unset this once we don't want to support models previous models.
|
||||||
import thinc.neural._classes.layernorm
|
#import thinc.neural._classes.layernorm
|
||||||
thinc.neural._classes.layernorm.set_compat_six_eight(True)
|
#thinc.neural._classes.layernorm.set_compat_six_eight(True)
|
||||||
|
|
||||||
VECTORS_KEY = 'spacy_pretrained_vectors'
|
VECTORS_KEY = 'spacy_pretrained_vectors'
|
||||||
|
|
||||||
|
|
Loading…
Reference in New Issue