Introducing Version 2.2.0!
- Text generation with training
- Word prediction training
- Saving/loading models
- Saving/loading preprocessed data
- You can now change the batch size when training and evaluating
- Dataclasses can now be used for all finetuning “arg” parameters
Introducing Version 2.1.0! You can now use any model type available on Hugging Face’s model distribution network for the implemented features. This includes BERT, ROBERTA, ALBERT XLNET and more.
You can also now perform token classification
Introducing Version 2.0.0!
We fully redesigned Happy Transformer from the ground up.
- Question answering training
- Multi label text classification training
- Single predictions for text classification
- Masked word prediction training
- Masked word prediction with multiple masks
Happy Transformer have been redesigned to promote scalability. Now it’s easier than ever to add new models and features, and we encourage you to create PRs to contribute to the project.