Improve defaults

This commit is contained in:
Matthew Honnibal 2017-08-12 19:24:17 -05:00
parent 28e930aae0
commit 3e30712b62
2 changed files with 3 additions and 3 deletions

View File

@ -138,7 +138,7 @@ class TokenVectorEncoder(BaseThincComponent):
name = 'tensorizer' name = 'tensorizer'
@classmethod @classmethod
def Model(cls, width=128, embed_size=7500, **cfg): def Model(cls, width=128, embed_size=4000, **cfg):
"""Create a new statistical model for the class. """Create a new statistical model for the class.
width (int): Output size of the model. width (int): Output size of the model.

View File

@ -238,12 +238,12 @@ cdef class Parser:
Base class of the DependencyParser and EntityRecognizer. Base class of the DependencyParser and EntityRecognizer.
""" """
@classmethod @classmethod
def Model(cls, nr_class, token_vector_width=128, hidden_width=128, depth=1, **cfg): def Model(cls, nr_class, token_vector_width=128, hidden_width=300, depth=1, **cfg):
depth = util.env_opt('parser_hidden_depth', depth) depth = util.env_opt('parser_hidden_depth', depth)
token_vector_width = util.env_opt('token_vector_width', token_vector_width) token_vector_width = util.env_opt('token_vector_width', token_vector_width)
hidden_width = util.env_opt('hidden_width', hidden_width) hidden_width = util.env_opt('hidden_width', hidden_width)
parser_maxout_pieces = util.env_opt('parser_maxout_pieces', 2) parser_maxout_pieces = util.env_opt('parser_maxout_pieces', 2)
embed_size = util.env_opt('embed_size', 7500) embed_size = util.env_opt('embed_size', 4000)
tensors = fine_tune(Tok2Vec(token_vector_width, embed_size, tensors = fine_tune(Tok2Vec(token_vector_width, embed_size,
preprocess=doc2feats())) preprocess=doc2feats()))
if parser_maxout_pieces == 1: if parser_maxout_pieces == 1: