From 6974f55daae4b4f5dfeae76ac40707992b961b5c Mon Sep 17 00:00:00 2001 From: Paul O'Leary McCann Date: Wed, 16 Mar 2022 15:15:53 +0900 Subject: [PATCH] Hack for transformer listener size --- spacy/ml/models/coref.py | 7 ++++++- 1 file changed, 6 insertions(+), 1 deletion(-) diff --git a/spacy/ml/models/coref.py b/spacy/ml/models/coref.py index c584ac659..f40a4c110 100644 --- a/spacy/ml/models/coref.py +++ b/spacy/ml/models/coref.py @@ -32,7 +32,12 @@ def build_wl_coref_model( # span predictor embeddings sp_embedding_size: int = 64, ): - dim = tok2vec.get_dim("nO") + # TODO fix this + try: + dim = tok2vec.get_dim("nO") + except ValueError: + # happens with transformer listener + dim = 768 with Model.define_operators({">>": chain}): # TODO chain tok2vec with these models