from __future__ import unicode_literals from ...en import English import pytest @pytest.fixture def en_tokenizer(): return English.Defaults.create_tokenizer() def test_big_ellipsis(en_tokenizer): tokens = en_tokenizer(u'$45...............Asking') assert len(tokens) > 2