diff --git a/website/docs/images/prodigy_overview.jpg b/website/docs/images/prodigy_overview.jpg new file mode 100644 index 000000000..84326ccea Binary files /dev/null and b/website/docs/images/prodigy_overview.jpg differ diff --git a/website/src/components/landing.js b/website/src/components/landing.js index fb03d2845..e1d60b15d 100644 --- a/website/src/components/landing.js +++ b/website/src/components/landing.js @@ -37,8 +37,8 @@ export const LandingSubtitle = ({ children }) => ( ) export const LandingGrid = ({ cols = 3, blocks = false, children }) => ( - - + + {children} diff --git a/website/src/styles/landing.module.sass b/website/src/styles/landing.module.sass index c29c0fffb..fc5dcea98 100644 --- a/website/src/styles/landing.module.sass +++ b/website/src/styles/landing.module.sass @@ -76,7 +76,7 @@ .banner background: var(--color-theme) color: var(--color-back) - padding: 5rem + padding: 1rem 5rem margin-bottom: var(--spacing-md) background-size: cover @@ -128,14 +128,17 @@ padding-right: 2rem @include breakpoint(max, md) + .banner + padding: 1rem 3rem + .banner-content display: block .banner-text padding-top: 0 - .col - grid-column: 1 / span 2 + .grid + grid-template-columns: 1fr !important .banner-button margin-bottom: var(--spacing-sm) diff --git a/website/src/widgets/landing.js b/website/src/widgets/landing.js index 1f788877c..9cb40acfe 100644 --- a/website/src/widgets/landing.js +++ b/website/src/widgets/landing.js @@ -20,6 +20,7 @@ import Button from '../components/button' import Link from '../components/link' import courseImage from '../../docs/images/course.jpg' +import prodigyImage from '../../docs/images/prodigy_overview.jpg' import BenchmarksChoi from 'usage/_benchmarks-choi.md' @@ -147,6 +148,59 @@ const Landing = ({ data }) => { + + + spaCy v3.0 features all new transformer-based pipelines that + bring spaCy's accuracy right up to the current state-of-the-art + . You can use any pretrained transformer to train your own pipelines, and even + share one transformer between multiple components with{' '} + multi-task learning. Training is now fully configurable and + extensible, and you can define your own custom models using{' '} + PyTorch, TensorFlow and other frameworks. The + new spaCy projects system lets you describe whole{' '} + end-to-end workflows in a single file, giving you an easy path + from prototype to production, and making it easy to clone and adapt + best-practice projects for your own use cases. + + + + + Prodigy: Radically efficient machine teaching + +
+
+ Prodigy is an annotation tool so efficient that data scientists + can do the annotation themselves, enabling a new level of rapid iteration. + Whether you're working on entity recognition, intent detection or image + classification, Prodigy can help you train and evaluate your + models faster. +
+
+ + + + + + { 55 exercises featuring videos, slide decks, multiple-choice questions and interactive coding practice in the browser. - - Prodigy is an annotation tool so efficient that data scientists - can do the annotation themselves, enabling a new level of rapid iteration. - Whether you're working on entity recognition, intent detection or image - classification, Prodigy can help you train and evaluate your - models faster. Stream in your own examples or real-world data from live APIs, - update your model in real-time and chain models together to build more complex - systems. + Learn more from small training corpora by initializing your models with{' '} + knowledge from raw text. The new pretrain command teaches + spaCy's CNN model to predict words based on their context, producing + representations of words in contexts. If you've seen Google's BERT system or + fast.ai's ULMFiT, spaCy's pretraining is similar – but much more efficient. It's + still experimental, but users are already reporting good results, so give it a + try! - - - - - - - Learn more from small training corpora by initializing your models with{' '} - knowledge from raw text. The new pretrain command teaches spaCy's - CNN model to predict words based on their context, producing representations of - words in contexts. If you've seen Google's BERT system or fast.ai's ULMFiT, spaCy's - pretraining is similar – but much more efficient. It's still experimental, but users - are already reporting good results, so give it a try! - -

Benchmarks