v1.1.0
We expect another release within a week or two that will add support for RoBERTa (see #890), but this is a quick intermediate release now that XLNet support is stable/working.
Highlighted changes:
- Full support for XLNet and the whole-word-masking variants of BERT.
- Many small improvements to Google Cloud Platform/Kubernetes/Docker support.
- Add small but handy option to automatically delete checkpoints when a job finishes.
max_vals
is now used when computing warmup time with optimizers that use warmup.- New
auto
option for tokenizer chooses an appropriate tokenizer for any given input module. - Some internal changes to how
<SOS>
/<EOS>
/[SEP]
/[CLS]
tokens are handled during task preprocessing. This will require small changes to custom task code along the lines of what is seen in #845.
Dependency changes:
- AllenNLP 0.8.4 now required
- pytorch_transformers 1.0 now required when using BERT or XLNet.
Warnings:
- Upgrading to 1.1 will break existing checkpoints for BERT-based models.