2017-10-03 15:26:20 +03:00
|
|
|
//- 💫 DOCS > USAGE > FACTS & FIGURES > BENCHMARKS > MODEL COMPARISON
|
|
|
|
|
|
|
|
p
|
|
|
|
| In this section, we provide benchmark accuracies for the pre-trained
|
|
|
|
| model pipelines we distribute with spaCy. Evaluations are conducted
|
|
|
|
| end-to-end from raw text, with no "gold standard" pre-processing, over
|
2017-11-06 20:19:00 +03:00
|
|
|
| text from a mix of genres where possible. For are more detailed
|
|
|
|
| comparison of the available models, see the new
|
|
|
|
| #[+a("/models/comparison") model comparison tool].
|
2017-10-03 15:26:20 +03:00
|
|
|
|
|
|
|
+aside("Methodology")
|
|
|
|
| The evaluation was conducted on raw text with no gold standard
|
|
|
|
| information. The parser, tagger and entity recognizer were trained on the
|
|
|
|
| #[+a("https://www.gabormelli.com/RKB/OntoNotes_Corpus") OntoNotes 5]
|
|
|
|
| corpus, the word vectors on #[+a("http://commoncrawl.org") Common Crawl].
|
|
|
|
|
2017-10-06 22:39:06 +03:00
|
|
|
+h(4, "benchmarks-models-english") English
|
|
|
|
|
2017-10-03 15:26:20 +03:00
|
|
|
+table(["Model", "spaCy", "Type", "UAS", "NER F", "POS", "WPS", "Size"])
|
|
|
|
+row
|
2017-11-06 20:19:00 +03:00
|
|
|
+cell #[+a("/models/en#en_core_web_sm") #[code en_core_web_sm]] 2.0.0a8
|
2017-10-03 15:26:20 +03:00
|
|
|
each data in ["2.x", "neural"]
|
2017-11-06 20:19:00 +03:00
|
|
|
+cell("right")=data
|
|
|
|
+cell("right") 91.7
|
|
|
|
+cell("right") 85.3
|
|
|
|
+cell("right") 97.0
|
|
|
|
+cell("right") 10.1k
|
|
|
|
+cell("right") #[strong 35 MB]
|
2017-10-03 15:26:20 +03:00
|
|
|
|
|
|
|
+row
|
2017-11-06 20:19:00 +03:00
|
|
|
+cell #[+a("/models/en#en_core_web_lg") #[code en_core_web_lg]] 2.0.0a3
|
2017-10-03 15:26:20 +03:00
|
|
|
each data in ["2.x", "neural"]
|
2017-11-06 20:19:00 +03:00
|
|
|
+cell("right")=data
|
|
|
|
+cell("right") #[strong 91.9]
|
|
|
|
+cell("right") #[strong 85.9]
|
|
|
|
+cell("right") #[strong 97.2]
|
|
|
|
+cell("right") 5.0k
|
|
|
|
+cell("right") 812 MB
|
2017-10-03 15:26:20 +03:00
|
|
|
|
|
|
|
+row("divider")
|
|
|
|
+cell #[code en_core_web_sm] 1.2.0
|
|
|
|
each data in ["1.x", "linear", 86.6, 78.5, 96.6]
|
2017-11-06 20:19:00 +03:00
|
|
|
+cell("right")=data
|
|
|
|
+cell("right") #[strong 25.7k]
|
|
|
|
+cell("right") 50 MB
|
2017-10-03 15:26:20 +03:00
|
|
|
|
|
|
|
+row
|
|
|
|
+cell #[code en_core_web_md] 1.2.1
|
|
|
|
each data in ["1.x", "linear", 90.6, 81.4, 96.7, "18.8k", "1 GB"]
|
2017-11-06 20:19:00 +03:00
|
|
|
+cell("right")=data
|
2017-10-06 22:39:06 +03:00
|
|
|
|
|
|
|
+h(4, "benchmarks-models-spanish") Spanish
|
|
|
|
|
2017-11-06 20:19:00 +03:00
|
|
|
+aside("Evaluation note")
|
|
|
|
| The NER accuracy refers to the "silver standard" annotations in the
|
|
|
|
| WikiNER corpus. Accuracy on these annotations tends to be higher than
|
|
|
|
| correct human annotations.
|
|
|
|
|
2017-10-06 22:39:06 +03:00
|
|
|
+table(["Model", "spaCy", "Type", "UAS", "NER F", "POS", "WPS", "Size"])
|
|
|
|
+row
|
2017-11-06 20:19:00 +03:00
|
|
|
+cell #[+a("/models/es#es_core_news_sm") #[code es_core_news_sm]] 2.0.0a0
|
|
|
|
+cell("right") 2.x
|
|
|
|
+cell("right") neural
|
|
|
|
+cell("right") 89.8
|
|
|
|
+cell("right") 88.7
|
|
|
|
+cell("right") #[strong 96.9]
|
|
|
|
+cell("right") #[em n/a]
|
|
|
|
+cell("right") #[strong 35 MB]
|
|
|
|
|
|
|
|
+row
|
|
|
|
+cell #[+a("/models/es#es_core_news_md") #[code es_core_news_md]] 2.0.0a0
|
|
|
|
+cell("right") 2.x
|
|
|
|
+cell("right") neural
|
|
|
|
+cell("right") #[strong 90.2]
|
|
|
|
+cell("right") 89.0
|
|
|
|
+cell("right") 97.8
|
|
|
|
+cell("right") #[em n/a]
|
|
|
|
+cell("right") 93 MB
|
2017-10-06 22:39:06 +03:00
|
|
|
|
|
|
|
+row("divider")
|
|
|
|
+cell #[code es_core_web_md] 1.1.0
|
|
|
|
each data in ["1.x", "linear", 87.5]
|
2017-11-06 20:19:00 +03:00
|
|
|
+cell("right")=data
|
|
|
|
+cell("right") #[strong 94.2]
|
|
|
|
+cell("right") 96.7
|
|
|
|
+cell("right") #[em n/a]
|
|
|
|
+cell("right") 377 MB
|