Multi-Label Classification using Transformer-based models: BERT and XLNet.
References:
Devlin, J., Et al., (2019), BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding.
Yang, Z., Et al., (2019), XLNet: Generalized Autoregressive Pretraining for Language Understanding.
Kebe, G. Y. , Matuszek, C. , Ferraro F., (2020), BERT vs. XLNet in Multilabel Text Classification.