Extreme Multi-label Text Classification: Difference between revisions

From statwiki
Jump to navigation Jump to search
No edit summary
Line 4: Line 4:
== Introduction ==
== Introduction ==


In the field of deep learning, there are times where the label set for text classification is so large that most models give poor results. The authors propose a new model called APLC-XLNet which fine tunes the generalized autoregressive pretrained model (XLNet) by using Adaptive Probabilistic Label Clusters (APLC) to calculate cross entropy loss. This method takes advantage of unbalanced label distributions by forming clusters to reduce training time. The authors experimented on five different datasets and achieved results far better than existing state-of-the-art models.


== Motivation ==
== Motivation ==

Revision as of 08:46, 9 November 2020

Presented By

Mohan Wu

Introduction

In the field of deep learning, there are times where the label set for text classification is so large that most models give poor results. The authors propose a new model called APLC-XLNet which fine tunes the generalized autoregressive pretrained model (XLNet) by using Adaptive Probabilistic Label Clusters (APLC) to calculate cross entropy loss. This method takes advantage of unbalanced label distributions by forming clusters to reduce training time. The authors experimented on five different datasets and achieved results far better than existing state-of-the-art models.

Motivation