Hack for transformer listener size

This commit is contained in:
Paul O'Leary McCann 2022-03-16 15:15:53 +09:00
parent 7811a1194b
commit 6974f55daa

View File

@ -32,7 +32,12 @@ def build_wl_coref_model(
# span predictor embeddings # span predictor embeddings
sp_embedding_size: int = 64, sp_embedding_size: int = 64,
): ):
# TODO fix this
try:
dim = tok2vec.get_dim("nO") dim = tok2vec.get_dim("nO")
except ValueError:
# happens with transformer listener
dim = 768
with Model.define_operators({">>": chain}): with Model.define_operators({">>": chain}):
# TODO chain tok2vec with these models # TODO chain tok2vec with these models