spaCy/tests
2014-08-18 19:14:00 +02:00
..
my_test.py * Initial commit. Tests passing for punctuation handling. Need contractions, file transport, tokenize function, etc. 2014-07-05 20:51:42 +02:00
sun.tokens * Working tokenization. en doesn't match PTB perfectly. Need to reorganize before adding more schemes. 2014-07-07 01:15:59 +02:00
sun.txt * Working tokenization. en doesn't match PTB perfectly. Need to reorganize before adding more schemes. 2014-07-07 01:15:59 +02:00
test_contractions.py * Refactor spacy so that chunks return arrays of lexemes, so that there is properly one lexeme per word. 2014-08-18 19:14:00 +02:00
test_group_by.py * Refactor spacy so that chunks return arrays of lexemes, so that there is properly one lexeme per word. 2014-08-18 19:14:00 +02:00
test_hash_table.py * Add hash tests 2014-08-02 21:58:31 +01:00
test_orth.py * Refactor spacy so that chunks return arrays of lexemes, so that there is properly one lexeme per word. 2014-08-18 19:14:00 +02:00
test_post_punct.py * Refactor spacy so that chunks return arrays of lexemes, so that there is properly one lexeme per word. 2014-08-18 19:14:00 +02:00
test_pre_punct.py * Refactor spacy so that chunks return arrays of lexemes, so that there is properly one lexeme per word. 2014-08-18 19:14:00 +02:00
test_ptb_match_wiki_sun.py * Add test to check how well we match ptb tokenizer. Needs more text. 2014-07-07 05:11:31 +02:00
test_rules.py * Refactor spacy so that chunks return arrays of lexemes, so that there is properly one lexeme per word. 2014-08-18 19:14:00 +02:00
test_surround_punct.py * Refactor spacy so that chunks return arrays of lexemes, so that there is properly one lexeme per word. 2014-08-18 19:14:00 +02:00
test_tokenizer.py * Refactor spacy so that chunks return arrays of lexemes, so that there is properly one lexeme per word. 2014-08-18 19:14:00 +02:00
test_vocab.py * Reading in tokenization rules correctly. Passing tests. 2014-07-07 00:02:55 +02:00
test_wiki_sun.py * Upd tokenization test 2014-07-07 05:11:04 +02:00
tokenizer.sed * Working tokenization. en doesn't match PTB perfectly. Need to reorganize before adding more schemes. 2014-07-07 01:15:59 +02:00