spaCy/website/docs/usage/101/_training.md

39 lines
2.4 KiB
Markdown
Raw Blame History

This file contains ambiguous Unicode characters

This file contains Unicode characters that might be confused with other characters. If you think that this is intentional, you can safely ignore this warning. Use the Escape button to reveal them.

spaCy's models are **statistical** and every "decision" they make for example,
which part-of-speech tag to assign, or whether a word is a named entity is a
**prediction**. This prediction is based on the examples the model has seen
during **training**. To train a model, you first need training data examples
of text, and the labels you want the model to predict. This could be a
part-of-speech tag, a named entity or any other information.
The model is then shown the unlabelled text and will make a prediction. Because
we know the correct answer, we can give the model feedback on its prediction in
the form of an **error gradient** of the **loss function** that calculates the
difference between the training example and the expected output. The greater the
difference, the more significant the gradient and the updates to our model.
> - **Training data:** Examples and their annotations.
> - **Text:** The input text the model should predict a label for.
> - **Label:** The label the model should predict.
> - **Gradient:** Gradient of the loss function calculating the difference
> between input and expected output.
![The training process](../../images/training.svg)
When training a model, we don't just want it to memorize our examples we want
it to come up with a theory that can be **generalized across other examples**.
After all, we don't just want the model to learn that this one instance of
"Amazon" right here is a company we want it to learn that "Amazon", in
contexts _like this_, is most likely a company. That's why the training data
should always be representative of the data we want to process. A model trained
on Wikipedia, where sentences in the first person are extremely rare, will
likely perform badly on Twitter. Similarly, a model trained on romantic novels
will likely perform badly on legal text.
This also means that in order to know how the model is performing, and whether
it's learning the right things, you don't only need **training data** you'll
also need **evaluation data**. If you only test the model with the data it was
trained on, you'll have no idea how well it's generalizing. If you want to train
a model from scratch, you usually need at least a few hundred examples for both
training and evaluation. To update an existing model, you can already achieve
decent results with very few examples as long as they're representative.