* Fix copying of tokenizer data in init_model

This commit is contained in:
Matthew Honnibal 2015-04-12 04:45:31 +02:00
parent 248a2b4b0f
commit 1629b33082

View File

@ -19,6 +19,7 @@ import plac
from pathlib import Path
from shutil import copyfile
from shutil import copytree
import codecs
from spacy.en import get_lex_props
@ -35,7 +36,7 @@ def setup_tokenizer(lang_data_dir, tok_dir):
src = lang_data_dir / filename
dst = tok_dir / filename
if not dst.exists():
copyfile(src, dst)
copyfile(str(src), str(dst))
def _read_clusters(loc):