//- 💫 DOCS > USAGE > FACTS & FIGURES > BENCHMARKS > MODEL COMPARISON p | In this section, we provide benchmark accuracies for the pre-trained | model pipelines we distribute with spaCy. Evaluations are conducted | end-to-end from raw text, with no "gold standard" pre-processing, over | text from a mix of genres where possible. For are more detailed | comparison of the available models, see the new | #[+a("/models/comparison") model comparison tool]. +aside("Methodology") | The evaluation was conducted on raw text with no gold standard | information. The parser, tagger and entity recognizer were trained on the | #[+a("https://www.gabormelli.com/RKB/OntoNotes_Corpus") OntoNotes 5] | corpus, the word vectors on #[+a("http://commoncrawl.org") Common Crawl]. +h(4, "benchmarks-models-english") English +table(["Model", "spaCy", "Type", "UAS", "NER F", "POS", "WPS", "Size"]) +row +cell #[+a("/models/en#en_core_web_sm") #[code en_core_web_sm]] 2.0.0 +cell("num") 2.x +cell neural +cell("num") 91.7 +cell("num") 85.3 +cell("num") 97.0 +cell("num") 10.1k +cell("num") #[strong 35MB] +row +cell #[+a("/models/en#en_core_web_md") #[code en_core_web_md]] 2.0.0 +cell("num") 2.x +cell neural +cell("num") 91.7 +cell("num") #[strong 85.9] +cell("num") 97.1 +cell("num") 10.0k +cell("num") 115MB +row +cell #[+a("/models/en#en_core_web_lg") #[code en_core_web_lg]] 2.0.0 +cell("num") 2.x +cell neural +cell("num") #[strong 91.9] +cell("num") #[strong 85.9] +cell("num") #[strong 97.2] +cell("num") 10.0k +cell("num") 812MB +row("divider") +cell #[code en_core_web_sm] 1.2.0 +cell("num") 1.x +cell linear +cell("num") 86.6 +cell("num") 78.5 +cell("num") 96.6 +cell("num") #[strong 25.7k] +cell("num") 50MB +row +cell #[code en_core_web_md] 1.2.1 +cell("num") 1.x +cell linear +cell("num") 90.6 +cell("num") 81.4 +cell("num") 96.7 +cell("num") 18.8k +cell("num") 1GB +h(4, "benchmarks-models-spanish") Spanish +aside("Evaluation note") | The NER accuracy refers to the "silver standard" annotations in the | WikiNER corpus. Accuracy on these annotations tends to be higher than | correct human annotations. +table(["Model", "spaCy", "Type", "UAS", "NER F", "POS", "WPS", "Size"]) +row +cell #[+a("/models/es#es_core_news_sm") #[code es_core_news_sm]] 2.0.0 +cell("num") 2.x +cell("num") neural +cell("num") 89.8 +cell("num") 88.7 +cell("num") #[strong 96.9] +cell("num") #[em n/a] +cell("num") #[strong 35MB] +row +cell #[+a("/models/es#es_core_news_md") #[code es_core_news_md]] 2.0.0 +cell("num") 2.x +cell("num") neural +cell("num") #[strong 90.2] +cell("num") 89.0 +cell("num") 97.8 +cell("num") #[em n/a] +cell("num") 93MB +row("divider") +cell #[code es_core_web_md] 1.1.0 each data in ["1.x", "linear", 87.5] +cell("num")=data +cell("num") #[strong 94.2] +cell("num") 96.7 +cell("num") #[em n/a] +cell("num") 377MB