From 9d8966a2c0cf8c57e595c9b65f1eb9e2dc3ea1ac Mon Sep 17 00:00:00 2001 From: Henning Peters Date: Wed, 10 Feb 2016 19:24:37 +0100 Subject: [PATCH] Update test_tokenizer.py --- spacy/tests/tokenizer/test_tokenizer.py | 14 +++++++------- 1 file changed, 7 insertions(+), 7 deletions(-) diff --git a/spacy/tests/tokenizer/test_tokenizer.py b/spacy/tests/tokenizer/test_tokenizer.py index ca88c66b7..9b17a01c6 100644 --- a/spacy/tests/tokenizer/test_tokenizer.py +++ b/spacy/tests/tokenizer/test_tokenizer.py @@ -8,13 +8,13 @@ import cloudpickle import tempfile -@pytest.mark.xfail -def test_pickle(en_tokenizer): - file_ = io.BytesIO() - cloudpickle.dump(en_tokenizer, file_) - file_.seek(0) - loaded = pickle.load(file_) - assert loaded is not None +# @pytest.mark.xfail +# def test_pickle(en_tokenizer): +# file_ = io.BytesIO() +# cloudpickle.dump(en_tokenizer, file_) +# file_.seek(0) +# loaded = pickle.load(file_) +# assert loaded is not None def test_no_word(en_tokenizer):