mirror of
https://github.com/explosion/spaCy.git
synced 2025-03-03 19:08:06 +03:00
Update multi-processing example and add to docs
This commit is contained in:
parent
4eabaafd66
commit
1d69a46cd4
|
@ -1,7 +1,9 @@
|
||||||
"""
|
"""
|
||||||
Example of multi-processing with joblib. Here, we're exporting
|
Example of multi-processing with Joblib. Here, we're exporting
|
||||||
part-of-speech-tagged, true-cased, (very roughly) sentence-separated text, with
|
part-of-speech-tagged, true-cased, (very roughly) sentence-separated text, with
|
||||||
each "sentence" on a newline, and spaces between tokens.
|
each "sentence" on a newline, and spaces between tokens. Data is loaded from
|
||||||
|
the IMDB movie reviews dataset and will be loaded automatically via Thinc's
|
||||||
|
built-in dataset loader.
|
||||||
|
|
||||||
Last updated for: spaCy 2.0.0a18
|
Last updated for: spaCy 2.0.0a18
|
||||||
"""
|
"""
|
||||||
|
|
|
@ -106,7 +106,7 @@
|
||||||
"How Pipelines Work": "pipelines",
|
"How Pipelines Work": "pipelines",
|
||||||
"Custom Components": "custom-components",
|
"Custom Components": "custom-components",
|
||||||
"Developing Extensions": "extensions",
|
"Developing Extensions": "extensions",
|
||||||
"Multi-threading": "multithreading",
|
"Multi-Threading": "multithreading",
|
||||||
"Serialization": "serialization"
|
"Serialization": "serialization"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
|
|
|
@ -38,3 +38,16 @@ p
|
||||||
| the generator in two, and then #[code izip] the extra stream to the
|
| the generator in two, and then #[code izip] the extra stream to the
|
||||||
| document stream. Here's
|
| document stream. Here's
|
||||||
| #[+a(gh("spacy") + "/issues/172#issuecomment-183963403") an example].
|
| #[+a(gh("spacy") + "/issues/172#issuecomment-183963403") an example].
|
||||||
|
|
||||||
|
+h(3, "multi-processing-example") Example: Multi-processing with Joblib
|
||||||
|
|
||||||
|
p
|
||||||
|
| This example shows how to use multiple cores to process text using
|
||||||
|
| spaCy and #[+a("https://pythonhosted.org/joblib/") Joblib]. We're
|
||||||
|
| exporting part-of-speech-tagged, true-cased, (very roughly)
|
||||||
|
| sentence-separated text, with each "sentence" on a newline, and
|
||||||
|
| spaces between tokens. Data is loaded from the IMDB movie reviews
|
||||||
|
| dataset and will be loaded automatically via Thinc's built-in dataset
|
||||||
|
| loader.
|
||||||
|
|
||||||
|
+github("spacy", "examples/parallel_tag.py")
|
||||||
|
|
|
@ -71,6 +71,19 @@ include ../_includes/_mixins
|
||||||
|
|
||||||
+github("spacy", "examples/pipeline/custom_attr_methods.py")
|
+github("spacy", "examples/pipeline/custom_attr_methods.py")
|
||||||
|
|
||||||
|
+h(3, "parallel-tag") Multi-processing with Joblib
|
||||||
|
|
||||||
|
p
|
||||||
|
| This example shows how to use multiple cores to process text using
|
||||||
|
| spaCy and #[+a("https://pythonhosted.org/joblib/") Joblib]. We're
|
||||||
|
| exporting part-of-speech-tagged, true-cased, (very roughly)
|
||||||
|
| sentence-separated text, with each "sentence" on a newline, and
|
||||||
|
| spaces between tokens. Data is loaded from the IMDB movie reviews
|
||||||
|
| dataset and will be loaded automatically via Thinc's built-in dataset
|
||||||
|
| loader.
|
||||||
|
|
||||||
|
+github("spacy", "examples/parallel_tag.py")
|
||||||
|
|
||||||
+section("training")
|
+section("training")
|
||||||
+h(3, "training-ner") Training spaCy's Named Entity Recognizer
|
+h(3, "training-ner") Training spaCy's Named Entity Recognizer
|
||||||
|
|
||||||
|
|
Loading…
Reference in New Issue
Block a user