🛸 Use pretrained transformers like BERT, XLNet and GPT-2 in spaCy
🌙 This release is a pre-release and requires spaCy v3 (nightly).
Transformer
component for easy pipeline integration.TransformerListener
to share transformer weights between components.spacy-transformers
is installed in the same environment.Transformer
: Pipeline component API reference⚠️ This release requires downloading new models.
spacy-transformers
for spaCy v2.3transformers
versions to >=2.4.0,<2.9.0
transformers.AutoConfig
to support loading pretrained models from https://huggingface.co/models
pytokenizations
Thanks to @tamuhey for the pull request!
Bug fixes related to alignment and truncation:
Enhancement:
Thanks to @ZhuoruLin for the pull requests and helping us debug issues related to batching and truncation!
Update to newer version of transformers
.
This library is being rewritten for spaCy v3, in order to improve its flexibility and performance and to make it easier to stay up to date with new transformer models. See here for details: https://github.com/explosion/spacy-transformers/pull/173
Fix various alignment and preprocessing bugs.
importlib_metadata
to prevent conflict.doc.tensor
.Thanks to @ssavvi for the pull request!
⚠️ This release requires downloading new models. Also note the new model names that specify
trf
(transformers) instead ofpytt
(PyTorch transformers).
spacy-pytorch-transformers
to spacy-transformers
.spacy>=2.2.0
.transformers
.pytorch-transformers
to support DistilBERT.pytorch-transformers
v1.1.