mirror of
				https://github.com/explosion/spaCy.git
				synced 2025-10-30 23:47:31 +03:00 
			
		
		
		
	Merge pull request #11886 from essenmitsosse/rename-mdx-files
Rename all MDX file to `.mdx`
This commit is contained in:
		
						commit
						3eed6801d7
					
				|  | @ -3,6 +3,6 @@ title: Library Architecture | ||||||
| next: /api/architectures | next: /api/architectures | ||||||
| --- | --- | ||||||
| 
 | 
 | ||||||
| import Architecture101 from 'usage/101/\_architecture.md' | import Architecture101 from 'usage/101/\_architecture.mdx' | ||||||
| 
 | 
 | ||||||
| <Architecture101 /> | <Architecture101 /> | ||||||
|  | @ -539,7 +539,7 @@ sit amet dignissim justo congue. | ||||||
| 
 | 
 | ||||||
| ## Markdown reference {#markdown} | ## Markdown reference {#markdown} | ||||||
| 
 | 
 | ||||||
| All page content and page meta lives in the `.md` files in the `/docs` | All page content and page meta lives in the `.mdx` files in the `/docs` | ||||||
| directory. The frontmatter block at the top of each file defines the page title | directory. The frontmatter block at the top of each file defines the page title | ||||||
| and other settings like the sidebar menu. | and other settings like the sidebar menu. | ||||||
| 
 | 
 | ||||||
|  | @ -69,7 +69,7 @@ pipeline, which is less accurate but much cheaper to run. | ||||||
| >   gold-standard segmentation and tokenization, from a pretty specific type of | >   gold-standard segmentation and tokenization, from a pretty specific type of | ||||||
| >   text (articles from a single newspaper, 1984-1989). | >   text (articles from a single newspaper, 1984-1989). | ||||||
| 
 | 
 | ||||||
| import Benchmarks from 'usage/\_benchmarks-models.md' | import Benchmarks from 'usage/\_benchmarks-models.mdx' | ||||||
| 
 | 
 | ||||||
| <Benchmarks /> | <Benchmarks /> | ||||||
| 
 | 
 | ||||||
|  | @ -28,7 +28,7 @@ annotations. | ||||||
| 
 | 
 | ||||||
| ## Part-of-speech tagging {#pos-tagging model="tagger, parser"} | ## Part-of-speech tagging {#pos-tagging model="tagger, parser"} | ||||||
| 
 | 
 | ||||||
| import PosDeps101 from 'usage/101/\_pos-deps.md' | import PosDeps101 from 'usage/101/\_pos-deps.mdx' | ||||||
| 
 | 
 | ||||||
| <PosDeps101 /> | <PosDeps101 /> | ||||||
| 
 | 
 | ||||||
|  | @ -536,7 +536,7 @@ with new examples. | ||||||
| 
 | 
 | ||||||
| ### Named Entity Recognition 101 {#named-entities-101} | ### Named Entity Recognition 101 {#named-entities-101} | ||||||
| 
 | 
 | ||||||
| import NER101 from 'usage/101/\_named-entities.md' | import NER101 from 'usage/101/\_named-entities.mdx' | ||||||
| 
 | 
 | ||||||
| <NER101 /> | <NER101 /> | ||||||
| 
 | 
 | ||||||
|  | @ -785,7 +785,7 @@ during tokenization. This is kind of a core principle of spaCy's `Doc` object: | ||||||
| 
 | 
 | ||||||
| </Infobox> | </Infobox> | ||||||
| 
 | 
 | ||||||
| import Tokenization101 from 'usage/101/\_tokenization.md' | import Tokenization101 from 'usage/101/\_tokenization.mdx' | ||||||
| 
 | 
 | ||||||
| <Tokenization101 /> | <Tokenization101 /> | ||||||
| 
 | 
 | ||||||
|  | @ -1868,7 +1868,7 @@ initialized before training. See the | ||||||
| 
 | 
 | ||||||
| ## Word vectors and semantic similarity {#vectors-similarity} | ## Word vectors and semantic similarity {#vectors-similarity} | ||||||
| 
 | 
 | ||||||
| import Vectors101 from 'usage/101/\_vectors-similarity.md' | import Vectors101 from 'usage/101/\_vectors-similarity.mdx' | ||||||
| 
 | 
 | ||||||
| <Vectors101 /> | <Vectors101 /> | ||||||
| 
 | 
 | ||||||
|  | @ -1998,7 +1998,7 @@ for word, vector in vector_data.items(): | ||||||
| 
 | 
 | ||||||
| ## Language Data {#language-data} | ## Language Data {#language-data} | ||||||
| 
 | 
 | ||||||
| import LanguageData101 from 'usage/101/\_language-data.md' | import LanguageData101 from 'usage/101/\_language-data.mdx' | ||||||
| 
 | 
 | ||||||
| <LanguageData101 /> | <LanguageData101 /> | ||||||
| 
 | 
 | ||||||
|  | @ -12,7 +12,7 @@ menu: | ||||||
|   - ['Plugins & Wrappers', 'plugins'] |   - ['Plugins & Wrappers', 'plugins'] | ||||||
| --- | --- | ||||||
| 
 | 
 | ||||||
| import Pipelines101 from 'usage/101/\_pipelines.md' | import Pipelines101 from 'usage/101/\_pipelines.mdx' | ||||||
| 
 | 
 | ||||||
| <Pipelines101 /> | <Pipelines101 /> | ||||||
| 
 | 
 | ||||||
|  | @ -10,7 +10,7 @@ menu: | ||||||
| 
 | 
 | ||||||
| ## Basics {#basics hidden="true"} | ## Basics {#basics hidden="true"} | ||||||
| 
 | 
 | ||||||
| import Serialization101 from 'usage/101/\_serialization.md' | import Serialization101 from 'usage/101/\_serialization.mdx' | ||||||
| 
 | 
 | ||||||
| <Serialization101 /> | <Serialization101 /> | ||||||
| 
 | 
 | ||||||
|  | @ -191,7 +191,7 @@ text with spaCy. | ||||||
| 
 | 
 | ||||||
| ### Tokenization {#annotations-token} | ### Tokenization {#annotations-token} | ||||||
| 
 | 
 | ||||||
| import Tokenization101 from 'usage/101/\_tokenization.md' | import Tokenization101 from 'usage/101/\_tokenization.mdx' | ||||||
| 
 | 
 | ||||||
| <Tokenization101 /> | <Tokenization101 /> | ||||||
| 
 | 
 | ||||||
|  | @ -207,7 +207,7 @@ language-specific data**, see the usage guides on | ||||||
| 
 | 
 | ||||||
| ### Part-of-speech tags and dependencies {#annotations-pos-deps model="parser"} | ### Part-of-speech tags and dependencies {#annotations-pos-deps model="parser"} | ||||||
| 
 | 
 | ||||||
| import PosDeps101 from 'usage/101/\_pos-deps.md' | import PosDeps101 from 'usage/101/\_pos-deps.mdx' | ||||||
| 
 | 
 | ||||||
| <PosDeps101 /> | <PosDeps101 /> | ||||||
| 
 | 
 | ||||||
|  | @ -222,7 +222,7 @@ how to **navigate and use the parse tree** effectively, see the usage guides on | ||||||
| 
 | 
 | ||||||
| ### Named Entities {#annotations-ner model="ner"} | ### Named Entities {#annotations-ner model="ner"} | ||||||
| 
 | 
 | ||||||
| import NER101 from 'usage/101/\_named-entities.md' | import NER101 from 'usage/101/\_named-entities.mdx' | ||||||
| 
 | 
 | ||||||
| <NER101 /> | <NER101 /> | ||||||
| 
 | 
 | ||||||
|  | @ -238,7 +238,7 @@ of a model, see the usage guides on | ||||||
| 
 | 
 | ||||||
| ### Word vectors and similarity {#vectors-similarity model="vectors"} | ### Word vectors and similarity {#vectors-similarity model="vectors"} | ||||||
| 
 | 
 | ||||||
| import Vectors101 from 'usage/101/\_vectors-similarity.md' | import Vectors101 from 'usage/101/\_vectors-similarity.mdx' | ||||||
| 
 | 
 | ||||||
| <Vectors101 /> | <Vectors101 /> | ||||||
| 
 | 
 | ||||||
|  | @ -252,7 +252,7 @@ To learn more about word vectors, how to **customize them** and how to load | ||||||
| 
 | 
 | ||||||
| ## Pipelines {#pipelines} | ## Pipelines {#pipelines} | ||||||
| 
 | 
 | ||||||
| import Pipelines101 from 'usage/101/\_pipelines.md' | import Pipelines101 from 'usage/101/\_pipelines.mdx' | ||||||
| 
 | 
 | ||||||
| <Pipelines101 /> | <Pipelines101 /> | ||||||
| 
 | 
 | ||||||
|  | @ -266,7 +266,7 @@ guide on [language processing pipelines](/usage/processing-pipelines). | ||||||
| 
 | 
 | ||||||
| ## Architecture {#architecture} | ## Architecture {#architecture} | ||||||
| 
 | 
 | ||||||
| import Architecture101 from 'usage/101/\_architecture.md' | import Architecture101 from 'usage/101/\_architecture.mdx' | ||||||
| 
 | 
 | ||||||
| <Architecture101 /> | <Architecture101 /> | ||||||
| 
 | 
 | ||||||
|  | @ -384,7 +384,7 @@ it. | ||||||
| 
 | 
 | ||||||
| ## Serialization {#serialization} | ## Serialization {#serialization} | ||||||
| 
 | 
 | ||||||
| import Serialization101 from 'usage/101/\_serialization.md' | import Serialization101 from 'usage/101/\_serialization.mdx' | ||||||
| 
 | 
 | ||||||
| <Serialization101 /> | <Serialization101 /> | ||||||
| 
 | 
 | ||||||
|  | @ -397,7 +397,7 @@ guide on [saving and loading](/usage/saving-loading#models). | ||||||
| 
 | 
 | ||||||
| ## Training {#training} | ## Training {#training} | ||||||
| 
 | 
 | ||||||
| import Training101 from 'usage/101/\_training.md' | import Training101 from 'usage/101/\_training.mdx' | ||||||
| 
 | 
 | ||||||
| <Training101 /> | <Training101 /> | ||||||
| 
 | 
 | ||||||
|  | @ -476,7 +476,7 @@ for trainable components. | ||||||
| 
 | 
 | ||||||
| ## Language data {#language-data} | ## Language data {#language-data} | ||||||
| 
 | 
 | ||||||
| import LanguageData101 from 'usage/101/\_language-data.md' | import LanguageData101 from 'usage/101/\_language-data.mdx' | ||||||
| 
 | 
 | ||||||
| <LanguageData101 /> | <LanguageData101 /> | ||||||
| 
 | 
 | ||||||
|  | @ -17,7 +17,7 @@ menu: | ||||||
| 
 | 
 | ||||||
| ## Introduction to training {#basics hidden="true"} | ## Introduction to training {#basics hidden="true"} | ||||||
| 
 | 
 | ||||||
| import Training101 from 'usage/101/\_training.md' | import Training101 from 'usage/101/\_training.mdx' | ||||||
| 
 | 
 | ||||||
| <Training101 /> | <Training101 /> | ||||||
| 
 | 
 | ||||||
|  | @ -88,7 +88,7 @@ giving you access to thousands of pretrained models for your pipelines. | ||||||
| 
 | 
 | ||||||
|  |  | ||||||
| 
 | 
 | ||||||
| import Benchmarks from 'usage/\_benchmarks-models.md' | import Benchmarks from 'usage/\_benchmarks-models.mdx' | ||||||
| 
 | 
 | ||||||
| <Benchmarks /> | <Benchmarks /> | ||||||
| 
 | 
 | ||||||
|  | @ -33,11 +33,11 @@ export function github(filepath, branch = DEFAULT_BRANCH) { | ||||||
| /** | /** | ||||||
|  * Get the source of a file in the documentation based on its slug |  * Get the source of a file in the documentation based on its slug | ||||||
|  * @param {string} slug - The slug, e.g. /api/doc. |  * @param {string} slug - The slug, e.g. /api/doc. | ||||||
|  * @param {boolean} [isIndex] - Whether the page is an index, e.g. /api/index.md |  * @param {boolean} [isIndex] - Whether the page is an index, e.g. /api/index.mdx | ||||||
|  * @param {string} [branch] - Optional branch on GitHub. Defaults to master. |  * @param {string} [branch] - Optional branch on GitHub. Defaults to master. | ||||||
|  */ |  */ | ||||||
| export function getCurrentSource(slug, isIndex = false, branch = DEFAULT_BRANCH) { | export function getCurrentSource(slug, isIndex = false, branch = DEFAULT_BRANCH) { | ||||||
|     const ext = isIndex ? '/index.md' : '.md' |     const ext = isIndex ? '/index.mdx' : '.mdx' | ||||||
|     return github(`website/docs${slug}${ext}`, branch) |     return github(`website/docs${slug}${ext}`, branch) | ||||||
| } | } | ||||||
| 
 | 
 | ||||||
|  |  | ||||||
|  | @ -27,7 +27,7 @@ import prodigyImage from '../../docs/images/prodigy_overview.jpg' | ||||||
| import projectsImage from '../../docs/images/projects.png' | import projectsImage from '../../docs/images/projects.png' | ||||||
| import tailoredPipelinesImage from '../../docs/images/spacy-tailored-pipelines_wide.png' | import tailoredPipelinesImage from '../../docs/images/spacy-tailored-pipelines_wide.png' | ||||||
| 
 | 
 | ||||||
| import Benchmarks from 'usage/_benchmarks-models.md' | import Benchmarks from '../../docs/usage/_benchmarks-models.mdx' | ||||||
| 
 | 
 | ||||||
| function getCodeExample(nightly) { | function getCodeExample(nightly) { | ||||||
|     return `# pip install -U ${nightly ? 'spacy-nightly --pre' : 'spacy'} |     return `# pip install -U ${nightly ? 'spacy-nightly --pre' : 'spacy'} | ||||||
|  |  | ||||||
		Loading…
	
		Reference in New Issue
	
	Block a user