diff --git a/website/_includes/_mixins.jade b/website/_includes/_mixins.jade index b7375e2e0..692b47887 100644 --- a/website/_includes/_mixins.jade +++ b/website/_includes/_mixins.jade @@ -181,7 +181,7 @@ mixin codepen(slug, height, default_tab) alt_file - [string] alternative file path used in footer and link button height - [integer] height of code preview in px -mixin github(repo, file, alt_file, height, language) +mixin github(repo, file, height, alt_file, language) - var branch = ALPHA ? "develop" : "master" - var height = height || 250 diff --git a/website/usage/_processing-pipelines/_custom-components.jade b/website/usage/_processing-pipelines/_custom-components.jade index ea3ea9b97..79cd77eef 100644 --- a/website/usage/_processing-pipelines/_custom-components.jade +++ b/website/usage/_processing-pipelines/_custom-components.jade @@ -234,7 +234,7 @@ p | when you customise spaCy's tokenization rules. When you call #[code nlp] | on a text, the custom pipeline component is applied to the #[code Doc] -+github("spacy", "examples/pipeline/custom_component_entities.py", false, 500) ++github("spacy", "examples/pipeline/custom_component_entities.py", 500) p | Wrapping this functionality in a @@ -255,7 +255,7 @@ p | #[code Token] – for example, the capital, latitude/longitude coordinates | and even the country flag. -+github("spacy", "examples/pipeline/custom_component_countries_api.py", false, 500) ++github("spacy", "examples/pipeline/custom_component_countries_api.py", 500) p | In this case, all data can be fetched on initialisation in one request. diff --git a/website/usage/_processing-pipelines/_multithreading.jade b/website/usage/_processing-pipelines/_multithreading.jade index a80768f38..4dff9c924 100644 --- a/website/usage/_processing-pipelines/_multithreading.jade +++ b/website/usage/_processing-pipelines/_multithreading.jade @@ -50,4 +50,4 @@ p | dataset and will be loaded automatically via Thinc's built-in dataset | loader. -+github("spacy", "examples/pipeline/multi_processing.py") ++github("spacy", "examples/pipeline/multi_processing.py", 500) diff --git a/website/usage/_training/_ner.jade b/website/usage/_training/_ner.jade index 12f92dbce..c1002ecdf 100644 --- a/website/usage/_training/_ner.jade +++ b/website/usage/_training/_ner.jade @@ -34,7 +34,7 @@ p | #[strong character offsets] and #[strong labels] of each entity contained | in the texts. - +github("spacy", "examples/training/train_ner.py") + +github("spacy", "examples/training/train_ner.py", 500) +h(4) Step by step guide @@ -88,7 +88,7 @@ p | recognizer over unlabelled sentences, and adding their annotations to the | training set. -+github("spacy", "examples/training/train_new_entity_type.py") ++github("spacy", "examples/training/train_new_entity_type.py", 500) +h(4) Step by step guide diff --git a/website/usage/_training/_tagger-parser.jade b/website/usage/_training/_tagger-parser.jade index d8388f4d7..f2fa4bab5 100644 --- a/website/usage/_training/_tagger-parser.jade +++ b/website/usage/_training/_tagger-parser.jade @@ -8,7 +8,7 @@ p | #[strong training examples] and the respective #[strong heads] and | #[strong dependency label] for each token of the example texts. -+github("spacy", "examples/training/train_parser.py") ++github("spacy", "examples/training/train_parser.py", 500) +h(4) Step by step guide @@ -61,7 +61,7 @@ p | #[strong custom tags], as well as a dictionary mapping those tags to the | #[+a("http://universaldependencies.github.io/docs/u/pos/index.html") Universal Dependencies scheme]. -+github("spacy", "examples/training/train_tagger.py") ++github("spacy", "examples/training/train_tagger.py", 500) +h(4) Step by step guide @@ -141,7 +141,7 @@ p | of relations: #[code ROOT], #[code PLACE], #[code QUALITY], | #[code ATTRIBUTE], #[code TIME] and #[code LOCATION]. -+github("spacy", "examples/training/train_intent_parser.py") ++github("spacy", "examples/training/train_intent_parser.py", 500) +h(4) Step by step guide diff --git a/website/usage/_training/_textcat.jade b/website/usage/_training/_textcat.jade index 5ccff7a84..b7b47c3ba 100644 --- a/website/usage/_training/_textcat.jade +++ b/website/usage/_training/_textcat.jade @@ -11,7 +11,7 @@ p | loader. Predictions are available via | #[+api("doc#attributes") #[code Doc.cats]]. -+github("spacy", "examples/training/train_textcat.py") ++github("spacy", "examples/training/train_textcat.py", 500) +h(4) Step by step guide diff --git a/website/usage/examples.jade b/website/usage/examples.jade index 5e415af8f..9ad800954 100644 --- a/website/usage/examples.jade +++ b/website/usage/examples.jade @@ -179,4 +179,4 @@ include ../_includes/_mixins | parameters, and was implemented using #[+a("https://keras.io") Keras] | and spaCy. - +github("spacy", "examples/keras_parikh_entailment/__main__.py", "examples/keras_parikh_entailment") + +github("spacy", "examples/keras_parikh_entailment/__main__.py", false, "examples/keras_parikh_entailment")