2016-11-01 04:13:54 +03:00
< a href = "https://explosion.ai" > < img src = "https://explosion.ai/assets/img/logo.svg" width = "125" height = "125" align = "right" / > < / a >
2016-11-01 05:19:15 +03:00
# A decomposable attention model for Natural Language Inference
2016-11-01 04:13:54 +03:00
**by Matthew Honnibal, [@honnibal ](https://github.com/honnibal )**
2018-10-01 11:28:45 +03:00
**Updated for spaCy 2.0+ and Keras 2.2.2+ by John Stewart, [@free-variation ](https://github.com/free-variation )**
2017-11-06 15:58:34 +03:00
2016-11-01 05:30:10 +03:00
This directory contains an implementation of the entailment prediction model described
2017-11-06 15:58:34 +03:00
by [Parikh et al. (2016) ](https://arxiv.org/pdf/1606.01933.pdf ). The model is notable
2016-11-01 04:13:54 +03:00
for its competitive performance with very few parameters.
2017-11-06 15:58:34 +03:00
The model is implemented using [Keras ](https://keras.io/ ) and [spaCy ](https://spacy.io ).
Keras is used to build and train the network. spaCy is used to load
the [GloVe ](http://nlp.stanford.edu/projects/glove/ ) vectors, perform the
feature extraction, and help you apply the model at run-time. The following
demo code shows how the entailment model can be used at runtime, once the
hook is installed to customise the `.similarity()` method of spaCy's `Doc`
2016-11-01 04:13:54 +03:00
and `Span` objects:
```python
2018-10-01 11:28:45 +03:00
def demo(shape):
nlp = spacy.load('en_vectors_web_lg')
nlp.add_pipe(KerasSimilarityShim.load(nlp.path / 'similarity', nlp, shape[0]))
doc1 = nlp(u'The king of France is bald.')
doc2 = nlp(u'France has no king.')
print("Sentence 1:", doc1)
print("Sentence 2:", doc2)
entailment_type, confidence = doc1.similarity(doc2)
print("Entailment type:", entailment_type, "(Confidence:", confidence, ")")
2016-11-01 04:13:54 +03:00
```
2016-11-01 03:51:54 +03:00
2018-10-01 11:28:45 +03:00
Which gives the output `Entailment type: contradiction (Confidence: 0.60604566)` , showing that
the system has definite opinions about Betrand Russell's [famous conundrum ](https://users.drew.edu/jlenz/br-on-denoting.html )!
2016-11-01 03:51:54 +03:00
I'm working on a blog post to explain Parikh et al.'s model in more detail.
2018-10-01 11:28:45 +03:00
A [notebook ](https://github.com/free-variation/spaCy/blob/master/examples/notebooks/Decompositional%20Attention.ipynb ) is available that briefly explains this implementation.
2016-11-01 03:51:54 +03:00
I think it is a very interesting example of the attention mechanism, which
2016-11-01 05:05:17 +03:00
I didn't understand very well before working through this paper. There are
lots of ways to extend the model.
2016-11-01 03:51:54 +03:00
2016-11-01 05:05:17 +03:00
## What's where
2016-11-01 03:51:54 +03:00
2016-11-01 05:19:15 +03:00
| File | Description |
| --- | --- |
2017-11-06 15:58:34 +03:00
| `__main__.py` | The script that will be executed. Defines the CLI, the data reading, etc — all the boring stuff. |
2018-10-01 11:28:45 +03:00
| `spacy_hook.py` | Provides a class `KerasSimilarityShim` that lets you use an arbitrary function to customize spaCy's `doc.similarity()` method. Instead of the default average-of-vectors algorithm, when you call `doc1.similarity(doc2)` , you'll get the result of `your_model(doc1, doc2)` . |
2016-11-01 05:19:15 +03:00
| `keras_decomposable_attention.py` | Defines the neural network model. |
2016-11-01 03:51:54 +03:00
2016-11-01 05:05:17 +03:00
## Setting up
2016-11-01 03:51:54 +03:00
2017-11-06 15:58:34 +03:00
First, install [Keras ](https://keras.io/ ), [spaCy ](https://spacy.io ) and the spaCy
2016-11-01 05:19:15 +03:00
English models (about 1GB of data):
2016-11-01 03:51:54 +03:00
2016-11-01 04:13:54 +03:00
```bash
2018-10-01 11:28:45 +03:00
pip install keras
2017-01-15 00:12:42 +03:00
pip install spacy
2018-10-01 11:28:45 +03:00
python -m spacy download en_vectors_web_lg
2016-11-01 04:13:54 +03:00
```
2016-11-01 03:51:54 +03:00
2018-10-01 11:28:45 +03:00
You'll also want to get Keras working on your GPU, and you will need a backend, such as TensorFlow or Theano.
This will depend on your set up, so you're mostly on your own for this step. If you're using AWS, try the
2016-11-01 05:19:15 +03:00
[NVidia AMI ](https://aws.amazon.com/marketplace/pp/B00FYCDDTE ). It made things pretty easy.
2016-11-01 03:51:54 +03:00
2016-11-01 05:05:17 +03:00
Once you've installed the dependencies, you can run a small preliminary test of
2016-11-01 05:19:15 +03:00
the Keras model:
2016-11-01 03:51:54 +03:00
2016-11-01 04:13:54 +03:00
```bash
py.test keras_parikh_entailment/keras_decomposable_attention.py
```
2016-11-01 03:51:54 +03:00
2016-11-01 05:05:17 +03:00
This compiles the model and fits it with some dummy data. You should see that
both tests passed.
2016-11-01 03:51:54 +03:00
2016-11-01 05:19:15 +03:00
Finally, download the [Stanford Natural Language Inference corpus ](http://nlp.stanford.edu/projects/snli/ ).
2016-11-01 03:51:54 +03:00
2016-11-01 05:05:17 +03:00
## Running the example
You can run the `keras_parikh_entailment/` directory as a script, which executes the file
2018-10-01 11:28:45 +03:00
[`keras_parikh_entailment/__main__.py` ](__main__.py ). If you run the script without arguments
the usage is shown. Running it with `-h` explains the command line arguments.
The first thing you'll want to do is train the model:
2016-11-01 03:51:54 +03:00
2016-11-01 04:13:54 +03:00
```bash
2018-10-01 11:28:45 +03:00
python keras_parikh_entailment/ train -t < path to SNLI train JSON > -s < path to SNLI dev JSON >
2016-11-01 04:13:54 +03:00
```
2016-11-01 03:51:54 +03:00
Training takes about 300 epochs for full accuracy, and I haven't rerun the full
2016-11-01 04:13:54 +03:00
experiment since refactoring things to publish this example — please let me
2018-10-01 11:28:45 +03:00
know if I've broken something. You should get to at least 85% on the development data even after 10-15 epochs.
2016-11-01 03:51:54 +03:00
2016-11-01 05:05:17 +03:00
The other two modes demonstrate run-time usage. I never like relying on the accuracy printed
by `.fit()` methods. I never really feel confident until I've run a new process that loads
the model and starts making predictions, without access to the gold labels. I've therefore
2018-10-01 11:28:45 +03:00
included an `evaluate` mode.
```bash
python keras_parikh_entailment/ evaluate -s < path to SNLI train JSON >
```
Finally, there's also a little demo, which mostly exists to show
2016-11-01 05:05:17 +03:00
you how run-time usage will eventually look.
2016-11-01 03:51:54 +03:00
2018-10-01 11:28:45 +03:00
```bash
python keras_parikh_entailment/ demo
```
2016-11-01 05:05:17 +03:00
## Getting updates
2016-11-01 03:51:54 +03:00
2016-11-01 05:05:17 +03:00
We should have the blog post explaining the model ready before the end of the week. To get
2018-01-03 14:24:22 +03:00
notified when it's published, you can either follow me on [Twitter ](https://twitter.com/honnibal )
2016-11-01 05:19:15 +03:00
or subscribe to our [mailing list ](http://eepurl.com/ckUpQ5 ).