mirror of
				https://github.com/explosion/spaCy.git
				synced 2025-11-04 01:48:04 +03:00 
			
		
		
		
	Update docs [ci skip]
This commit is contained in:
		
							parent
							
								
									6f3649923c
								
							
						
					
					
						commit
						b7e34c1451
					
				| 
						 | 
					@ -27,12 +27,12 @@ lemmatizers, see the
 | 
				
			||||||
> nlp.add_pipe("lemmatizer", config=config)
 | 
					> nlp.add_pipe("lemmatizer", config=config)
 | 
				
			||||||
> ```
 | 
					> ```
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| Setting     | Type                                       | Description                                                                                                                                                                    | Default    |
 | 
					| Setting     | Type                                       | Description                                                                                                                                                                            | Default    |
 | 
				
			||||||
| ----------- | ------------------------------------------ | ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ | ---------- |
 | 
					| ----------- | ------------------------------------------ | -------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- | ---------- |
 | 
				
			||||||
| `mode`      | str                                        | The lemmatizer mode, e.g. "lookup" or "rule".                                                                                                                                  | `"lookup"` |
 | 
					| `mode`      | str                                        | The lemmatizer mode, e.g. "lookup" or "rule".                                                                                                                                          | `"lookup"` |
 | 
				
			||||||
| `lookups`   | [`Lookups`](/api/lookups)                  | The lookups object containing the tables such as "lemma_rules", "lemma_index", "lemma_exc" and "lemma_lookup". If `None`, default tables are loaded from `spacy-lookups-data`. | `None`     |
 | 
					| `lookups`   | [`Lookups`](/api/lookups)                  | The lookups object containing the tables such as `"lemma_rules"`, `"lemma_index"`, `"lemma_exc"` and `"lemma_lookup"`. If `None`, default tables are loaded from `spacy-lookups-data`. | `None`     |
 | 
				
			||||||
| `overwrite` | bool                                       | Whether to overwrite existing lemmas.                                                                                                                                          | `False`    |
 | 
					| `overwrite` | bool                                       | Whether to overwrite existing lemmas.                                                                                                                                                  | `False`    |
 | 
				
			||||||
| `model`     | [`Model`](https://thinc.ai/docs/api-model) | **Not yet implemented:** the model to use.                                                                                                                                     | `None`     |
 | 
					| `model`     | [`Model`](https://thinc.ai/docs/api-model) | **Not yet implemented:** the model to use.                                                                                                                                             | `None`     |
 | 
				
			||||||
 | 
					
 | 
				
			||||||
```python
 | 
					```python
 | 
				
			||||||
https://github.com/explosion/spaCy/blob/develop/spacy/pipeline/lemmatizer.py
 | 
					https://github.com/explosion/spaCy/blob/develop/spacy/pipeline/lemmatizer.py
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -12,14 +12,15 @@ passed on to the next component.
 | 
				
			||||||
> - **Creates:** Objects, attributes and properties modified and set by the
 | 
					> - **Creates:** Objects, attributes and properties modified and set by the
 | 
				
			||||||
>   component.
 | 
					>   component.
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| Name          | Component                                                          | Creates                                                     | Description                                      |
 | 
					| Name           | Component                                                          | Creates                                                   | Description                                      |
 | 
				
			||||||
| ------------- | ------------------------------------------------------------------ | ----------------------------------------------------------- | ------------------------------------------------ |
 | 
					| -------------- | ------------------------------------------------------------------ | --------------------------------------------------------- | ------------------------------------------------ |
 | 
				
			||||||
| **tokenizer** | [`Tokenizer`](/api/tokenizer)                                      | `Doc`                                                       | Segment text into tokens.                        |
 | 
					| **tokenizer**  | [`Tokenizer`](/api/tokenizer)                                      | `Doc`                                                     | Segment text into tokens.                        |
 | 
				
			||||||
| **tagger**    | [`Tagger`](/api/tagger)                                            | `Doc[i].tag`                                                | Assign part-of-speech tags.                      |
 | 
					| **tagger**     | [`Tagger`](/api/tagger)                                            | `Token.tag`                                               | Assign part-of-speech tags.                      |
 | 
				
			||||||
| **parser**    | [`DependencyParser`](/api/dependencyparser)                        | `Doc[i].head`, `Doc[i].dep`, `Doc.sents`, `Doc.noun_chunks` | Assign dependency labels.                        |
 | 
					| **parser**     | [`DependencyParser`](/api/dependencyparser)                        | `Token.head`, `Token.dep`, `Doc.sents`, `Doc.noun_chunks` | Assign dependency labels.                        |
 | 
				
			||||||
| **ner**       | [`EntityRecognizer`](/api/entityrecognizer)                        | `Doc.ents`, `Doc[i].ent_iob`, `Doc[i].ent_type`             | Detect and label named entities.                 |
 | 
					| **ner**        | [`EntityRecognizer`](/api/entityrecognizer)                        | `Doc.ents`, `Token.ent_iob`, `Token.ent_type`             | Detect and label named entities.                 |
 | 
				
			||||||
| **textcat**   | [`TextCategorizer`](/api/textcategorizer)                          | `Doc.cats`                                                  | Assign document labels.                          |
 | 
					| **lemmatizer** | [`Lemmatizer`](/api/lemmatizer)                                    | `Token.lemma`                                             | Assign base forms.                               |
 | 
				
			||||||
| ...           | [custom components](/usage/processing-pipelines#custom-components) | `Doc._.xxx`, `Token._.xxx`, `Span._.xxx`                    | Assign custom attributes, methods or properties. |
 | 
					| **textcat**    | [`TextCategorizer`](/api/textcategorizer)                          | `Doc.cats`                                                | Assign document labels.                          |
 | 
				
			||||||
 | 
					| ...            | [custom components](/usage/processing-pipelines#custom-components) | `Doc._.xxx`, `Token._.xxx`, `Span._.xxx`                  | Assign custom attributes, methods or properties. |
 | 
				
			||||||
 | 
					
 | 
				
			||||||
The processing pipeline always **depends on the statistical model** and its
 | 
					The processing pipeline always **depends on the statistical model** and its
 | 
				
			||||||
capabilities. For example, a pipeline can only include an entity recognizer
 | 
					capabilities. For example, a pipeline can only include an entity recognizer
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -228,16 +228,13 @@ available pipeline components and component functions.
 | 
				
			||||||
| `entity_linker` | [`EntityLinker`](/api/entitylinker)             | Assign knowledge base IDs to named entities. Should be added after the entity recognizer. |
 | 
					| `entity_linker` | [`EntityLinker`](/api/entitylinker)             | Assign knowledge base IDs to named entities. Should be added after the entity recognizer. |
 | 
				
			||||||
| `entity_ruler`  | [`EntityRuler`](/api/entityruler)               | Assign named entities based on pattern rules and dictionaries.                            |
 | 
					| `entity_ruler`  | [`EntityRuler`](/api/entityruler)               | Assign named entities based on pattern rules and dictionaries.                            |
 | 
				
			||||||
| `textcat`       | [`TextCategorizer`](/api/textcategorizer)       | Assign text categories.                                                                   |
 | 
					| `textcat`       | [`TextCategorizer`](/api/textcategorizer)       | Assign text categories.                                                                   |
 | 
				
			||||||
 | 
					| `lemmatizer`    | [`Lemmatizer`](/api/lemmatizer)                 | Assign base forms to words.                                                               |
 | 
				
			||||||
| `morphologizer` | [`Morphologizer`](/api/morphologizer)           | Assign morphological features and coarse-grained POS tags.                                |
 | 
					| `morphologizer` | [`Morphologizer`](/api/morphologizer)           | Assign morphological features and coarse-grained POS tags.                                |
 | 
				
			||||||
| `senter`        | [`SentenceRecognizer`](/api/sentencerecognizer) | Assign sentence boundaries.                                                               |
 | 
					| `senter`        | [`SentenceRecognizer`](/api/sentencerecognizer) | Assign sentence boundaries.                                                               |
 | 
				
			||||||
| `sentencizer`   | [`Sentencizer`](/api/sentencizer)               | Add rule-based sentence segmentation without the dependency parse.                        |
 | 
					| `sentencizer`   | [`Sentencizer`](/api/sentencizer)               | Add rule-based sentence segmentation without the dependency parse.                        |
 | 
				
			||||||
| `tok2vec`       | [`Tok2Vec`](/api/tok2vec)                       |                                                                                           |
 | 
					| `tok2vec`       | [`Tok2Vec`](/api/tok2vec)                       |                                                                                           |
 | 
				
			||||||
| `transformer`   | [`Transformer`](/api/transformer)               | Assign the tokens and outputs of a transformer model.                                     |
 | 
					| `transformer`   | [`Transformer`](/api/transformer)               | Assign the tokens and outputs of a transformer model.                                     |
 | 
				
			||||||
 | 
					
 | 
				
			||||||
<!-- TODO: finish and update with more components -->
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
<!-- TODO: explain default config and factories -->
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
### Disabling and modifying pipeline components {#disabling}
 | 
					### Disabling and modifying pipeline components {#disabling}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
If you don't need a particular component of the pipeline – for example, the
 | 
					If you don't need a particular component of the pipeline – for example, the
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
		Loading…
	
		Reference in New Issue
	
	Block a user