mirror of
https://github.com/explosion/spaCy.git
synced 2024-11-14 21:57:15 +03:00
e597110d31
<!--- Provide a general summary of your changes in the title. --> ## Description The new website is implemented using [Gatsby](https://www.gatsbyjs.org) with [Remark](https://github.com/remarkjs/remark) and [MDX](https://mdxjs.com/). This allows authoring content in **straightforward Markdown** without the usual limitations. Standard elements can be overwritten with powerful [React](http://reactjs.org/) components and wherever Markdown syntax isn't enough, JSX components can be used. Hopefully, this update will also make it much easier to contribute to the docs. Once this PR is merged, I'll implement auto-deployment via [Netlify](https://netlify.com) on a specific branch (to avoid building the website on every PR). There's a bunch of other cool stuff that the new setup will allow us to do – including writing front-end tests, service workers, offline support, implementing a search and so on. This PR also includes various new docs pages and content. Resolves #3270. Resolves #3222. Resolves #2947. Resolves #2837. ### Types of change enhancement ## Checklist <!--- Before you submit the PR, go over this checklist and make sure you can tick off all the boxes. [] -> [x] --> - [x] I have submitted the spaCy Contributor Agreement. - [x] I ran the tests, and all new and existing tests passed. - [x] My changes don't require a change to the documentation, or if they do, I've added all required information.
79 lines
3.2 KiB
Markdown
79 lines
3.2 KiB
Markdown
---
|
|
title: SentenceSegmenter
|
|
tag: class
|
|
source: spacy/pipeline.pyx
|
|
---
|
|
|
|
A simple spaCy hook, to allow custom sentence boundary detection logic that
|
|
doesn't require the dependency parse. By default, sentence segmentation is
|
|
performed by the [`DependencyParser`](/api/dependencyparser), so the
|
|
`SentenceSegmenter` lets you implement a simpler, rule-based strategy that
|
|
doesn't require a statistical model to be loaded. The component is also
|
|
available via the string name `"sentencizer"`. After initialization, it is
|
|
typically added to the processing pipeline using
|
|
[`nlp.add_pipe`](/api/language#add_pipe).
|
|
|
|
## SentenceSegmenter.\_\_init\_\_ {#init tag="method"}
|
|
|
|
Initialize the sentence segmenter. To change the sentence boundary detection
|
|
strategy, pass a generator function `strategy` on initialization, or assign a
|
|
new strategy to the `.strategy` attribute. Sentence detection strategies should
|
|
be generators that take `Doc` objects and yield `Span` objects for each
|
|
sentence.
|
|
|
|
> #### Example
|
|
>
|
|
> ```python
|
|
> # Construction via create_pipe
|
|
> sentencizer = nlp.create_pipe("sentencizer")
|
|
>
|
|
> # Construction from class
|
|
> from spacy.pipeline import SentenceSegmenter
|
|
> sentencizer = SentenceSegmenter(nlp.vocab)
|
|
> ```
|
|
|
|
| Name | Type | Description |
|
|
| ----------- | ------------------- | ----------------------------------------------------------- |
|
|
| `vocab` | `Vocab` | The shared vocabulary. |
|
|
| `strategy` | unicode / callable | The segmentation strategy to use. Defaults to `"on_punct"`. |
|
|
| **RETURNS** | `SentenceSegmenter` | The newly constructed object. |
|
|
|
|
## SentenceSegmenter.\_\_call\_\_ {#call tag="method"}
|
|
|
|
Apply the sentence segmenter on a `Doc`. Typically, this happens automatically
|
|
after the component has been added to the pipeline using
|
|
[`nlp.add_pipe`](/api/language#add_pipe).
|
|
|
|
> #### Example
|
|
>
|
|
> ```python
|
|
> from spacy.lang.en import English
|
|
>
|
|
> nlp = English()
|
|
> sentencizer = nlp.create_pipe("sentencizer")
|
|
> nlp.add_pipe(sentencizer)
|
|
> doc = nlp(u"This is a sentence. This is another sentence.")
|
|
> assert list(doc.sents) == 2
|
|
> ```
|
|
|
|
| Name | Type | Description |
|
|
| ----------- | ----- | ------------------------------------------------------------ |
|
|
| `doc` | `Doc` | The `Doc` object to process, e.g. the `Doc` in the pipeline. |
|
|
| **RETURNS** | `Doc` | The modified `Doc` with added sentence boundaries. |
|
|
|
|
## SentenceSegmenter.split_on_punct {#split_on_punct tag="staticmethod"}
|
|
|
|
Split the `Doc` on punctuation characters `.`, `!` and `?`. This is the default
|
|
strategy used by the `SentenceSegmenter.`
|
|
|
|
| Name | Type | Description |
|
|
| ---------- | ------ | ------------------------------ |
|
|
| `doc` | `Doc` | The `Doc` object to process. |
|
|
| **YIELDS** | `Span` | The sentences in the document. |
|
|
|
|
## Attributes {#attributes}
|
|
|
|
| Name | Type | Description |
|
|
| ---------- | -------- | ------------------------------------------------------------------- |
|
|
| `strategy` | callable | The segmentation strategy. Can be overwritten after initialization. |
|