2017-01-24 09:55:02 +00:00
|
|
|
|
# coding: utf-8
|
|
|
|
|
from __future__ import unicode_literals
|
|
|
|
|
|
|
|
|
|
import pytest
|
2018-07-24 18:09:06 +00:00
|
|
|
|
from .... import util
|
|
|
|
|
|
|
|
|
|
@pytest.fixture(scope='module')
|
|
|
|
|
def fr_tokenizer():
|
|
|
|
|
return util.get_lang_class('fr').Defaults.create_tokenizer()
|
2017-01-24 09:55:02 +00:00
|
|
|
|
|
|
|
|
|
|
2018-07-24 21:38:44 +00:00
|
|
|
|
@pytest.mark.parametrize('text', [
|
|
|
|
|
"aujourd'hui", "Aujourd'hui", "prud'hommes", "prud’hommal"])
|
|
|
|
|
def test_fr_tokenizer_infix_exceptions(fr_tokenizer, text):
|
2017-01-24 09:55:02 +00:00
|
|
|
|
tokens = fr_tokenizer(text)
|
|
|
|
|
assert len(tokens) == 1
|
|
|
|
|
|
|
|
|
|
|
2018-07-24 21:38:44 +00:00
|
|
|
|
@pytest.mark.parametrize('text,lemma', [
|
|
|
|
|
("janv.", "janvier"),
|
|
|
|
|
("juill.", "juillet"),
|
|
|
|
|
("Dr.", "docteur"),
|
|
|
|
|
("av.", "avant"),
|
|
|
|
|
("sept.", "septembre")])
|
|
|
|
|
def test_fr_tokenizer_handles_abbr(fr_tokenizer, text, lemma):
|
2017-01-24 09:55:02 +00:00
|
|
|
|
tokens = fr_tokenizer(text)
|
|
|
|
|
assert len(tokens) == 1
|
|
|
|
|
assert tokens[0].lemma_ == lemma
|
|
|
|
|
|
|
|
|
|
|
2018-07-24 21:38:44 +00:00
|
|
|
|
def test_fr_tokenizer_handles_exc_in_text(fr_tokenizer):
|
2017-01-24 09:55:02 +00:00
|
|
|
|
text = "Je suis allé au mois de janv. aux prud’hommes."
|
|
|
|
|
tokens = fr_tokenizer(text)
|
|
|
|
|
assert len(tokens) == 10
|
|
|
|
|
assert tokens[6].text == "janv."
|
|
|
|
|
assert tokens[6].lemma_ == "janvier"
|
|
|
|
|
assert tokens[8].text == "prud’hommes"
|
2017-02-10 12:17:05 +00:00
|
|
|
|
|
|
|
|
|
|
2018-07-24 21:38:44 +00:00
|
|
|
|
def test_fr_tokenizer_handles_exc_in_text_2(fr_tokenizer):
|
2017-02-10 12:17:05 +00:00
|
|
|
|
text = "Cette après-midi, je suis allé dans un restaurant italo-mexicain."
|
|
|
|
|
tokens = fr_tokenizer(text)
|
|
|
|
|
assert len(tokens) == 11
|
|
|
|
|
assert tokens[1].text == "après-midi"
|
2017-04-27 08:59:38 +00:00
|
|
|
|
assert tokens[9].text == "italo-mexicain"
|
|
|
|
|
|
2018-07-24 21:38:44 +00:00
|
|
|
|
|
|
|
|
|
def test_fr_tokenizer_handles_title(fr_tokenizer):
|
2017-04-27 08:59:38 +00:00
|
|
|
|
text = "N'est-ce pas génial?"
|
|
|
|
|
tokens = fr_tokenizer(text)
|
|
|
|
|
assert len(tokens) == 6
|
|
|
|
|
assert tokens[0].text == "N'"
|
|
|
|
|
assert tokens[0].lemma_ == "ne"
|
2017-04-27 09:52:14 +00:00
|
|
|
|
assert tokens[1].text == "est"
|
|
|
|
|
assert tokens[1].lemma_ == "être"
|
2017-04-27 08:59:38 +00:00
|
|
|
|
assert tokens[2].text == "-ce"
|
|
|
|
|
assert tokens[2].lemma_ == "ce"
|
|
|
|
|
|
2018-07-24 21:38:44 +00:00
|
|
|
|
|
|
|
|
|
def test_fr_tokenizer_handles_title_2(fr_tokenizer):
|
2017-04-27 08:59:38 +00:00
|
|
|
|
text = "Est-ce pas génial?"
|
|
|
|
|
tokens = fr_tokenizer(text)
|
|
|
|
|
assert len(tokens) == 6
|
|
|
|
|
assert tokens[0].text == "Est"
|
|
|
|
|
assert tokens[0].lemma_ == "être"
|
|
|
|
|
|
2018-07-24 21:38:44 +00:00
|
|
|
|
|
|
|
|
|
def test_fr_tokenizer_handles_title_2(fr_tokenizer):
|
2017-04-27 08:59:38 +00:00
|
|
|
|
text = "Qu'est-ce que tu fais?"
|
|
|
|
|
tokens = fr_tokenizer(text)
|
|
|
|
|
assert len(tokens) == 7
|
|
|
|
|
assert tokens[0].text == "Qu'"
|
2018-07-24 21:38:44 +00:00
|
|
|
|
assert tokens[0].lemma_ == "que"
|