Models
The models submodule contains implementations of various algorithms that can be used in addition to external packages to evaluate and develop new natural language processing systems. A description of which algorithms are used in each scenario can be found on this table
Summary
The following table summarizes each submodule.
Submodule | Description |
---|---|
bert | This submodule includes the BERT-based models for sequence classification, token classification, and sequence encoding. |
gensen | This submodule includes a distributed Pytorch implementation based on Horovod of learning general purpose distributed sentence representations via large scale multi-task learning by refactoring https://github.com/Maluuba/gensen |
pretrained embeddings | This submodule provides utilities to download and extract pretrained word embeddings trained with Word2Vec, GloVe, fastText methods. |
pytorch_modules | This submodule provides Pytorch modules like Gated Recurrent Unit with peepholes. |
xlnet | This submodule includes the XLNet-based model for sequence classification. |