2018-11-26 17:54:27 +00:00
|
|
|
# coding: utf-8
|
|
|
|
from __future__ import unicode_literals
|
|
|
|
|
2019-03-15 17:14:46 +00:00
|
|
|
from spacy.cli.converters import conllu2json
|
|
|
|
|
|
|
|
|
|
|
|
def test_cli_converters_conllu2json():
|
|
|
|
# https://raw.githubusercontent.com/ohenrik/nb_news_ud_sm/master/original_data/no-ud-dev-ner.conllu
|
|
|
|
lines = [
|
|
|
|
"1\tDommer\tdommer\tNOUN\t_\tDefinite=Ind|Gender=Masc|Number=Sing\t2\tappos\t_\tO",
|
|
|
|
"2\tFinn\tFinn\tPROPN\t_\tGender=Masc\t4\tnsubj\t_\tB-PER",
|
|
|
|
"3\tEilertsen\tEilertsen\tPROPN\t_\t_\t2\tname\t_\tI-PER",
|
|
|
|
"4\tavstår\tavstå\tVERB\t_\tMood=Ind|Tense=Pres|VerbForm=Fin\t0\troot\t_\tO",
|
|
|
|
]
|
|
|
|
input_data = "\n".join(lines)
|
|
|
|
converted = conllu2json(input_data, n_sents=1)
|
|
|
|
assert len(converted) == 1
|
|
|
|
assert converted[0]["id"] == 0
|
|
|
|
assert len(converted[0]["paragraphs"]) == 1
|
|
|
|
assert len(converted[0]["paragraphs"][0]["sentences"]) == 1
|
|
|
|
sent = converted[0]["paragraphs"][0]["sentences"][0]
|
|
|
|
assert len(sent["tokens"]) == 4
|
|
|
|
tokens = sent["tokens"]
|
|
|
|
assert [t["orth"] for t in tokens] == ["Dommer", "Finn", "Eilertsen", "avstår"]
|
|
|
|
assert [t["tag"] for t in tokens] == ["NOUN", "PROPN", "PROPN", "VERB"]
|
|
|
|
assert [t["head"] for t in tokens] == [1, 2, -1, 0]
|
|
|
|
assert [t["dep"] for t in tokens] == ["appos", "nsubj", "name", "ROOT"]
|
|
|
|
assert [t["ner"] for t in tokens] == ["O", "B-PER", "L-PER", "O"]
|