spaCy/spacy/tests/tokenizer/test_emoticons.py

36 lines
1.1 KiB
Python
Raw Normal View History

2014-11-02 02:22:14 +00:00
from __future__ import unicode_literals
2014-12-21 09:43:27 +00:00
import pytest
2014-11-02 02:22:14 +00:00
def test_tweebo_challenge(en_tokenizer):
2014-11-02 02:22:14 +00:00
text = u""":o :/ :'( >:o (: :) >.< XD -__- o.O ;D :-) @_@ :P 8D :1 >:( :D =| ") :> ...."""
tokens = en_tokenizer(text)
assert tokens[0].orth_ == ":o"
assert tokens[1].orth_ == ":/"
assert tokens[2].orth_ == ":'("
assert tokens[3].orth_ == ">:o"
assert tokens[4].orth_ == "(:"
assert tokens[5].orth_ == ":)"
assert tokens[6].orth_ == ">.<"
assert tokens[7].orth_ == "XD"
assert tokens[8].orth_ == "-__-"
assert tokens[9].orth_ == "o.O"
assert tokens[10].orth_ == ";D"
assert tokens[11].orth_ == ":-)"
assert tokens[12].orth_ == "@_@"
assert tokens[13].orth_ == ":P"
assert tokens[14].orth_ == "8D"
assert tokens[15].orth_ == ":1"
assert tokens[16].orth_ == ">:("
assert tokens[17].orth_ == ":D"
assert tokens[18].orth_ == "=|"
assert tokens[19].orth_ == '")'
assert tokens[20].orth_ == ':>'
assert tokens[21].orth_ == '....'
def test_false_positive(en_tokenizer):
text = "example:)"
tokens = en_tokenizer(text)
assert len(tokens) == 3