Abstract
Universal Sentence Encoder (USE) has gained much popularity recently as a general-purpose sentence encoding technique. As the name suggests, USE is designed to be fairly general and has indeed been shown to achieve superior performances for many downstream NLP tasks. In this paper, we present an interesting “negative” result on USE in the context of zero-shot text classification, a challenging task, which has recently gained much attraction. More specifically, we found some interesting cases of zero-shot text classification, where topic based inference outperformed USE-based inference in terms of F1 score. Further investigation revealed that USE struggles to perform well on data-sets with a large number of labels with high semantic overlaps, while topic-based classification works well for the same.
| Original language | American English |
|---|---|
| Title of host publication | Proceedings of the 2nd Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics and the 12th International Joint Conference on Natural Language Processing |
| Editors | Yulan He, Heng Ji, Sujian Li, Yang Liu, Chua-Hui Chang |
| Publisher | Association for Computational Linguistics |
| Pages | 135-147 |
| Volume | 2 |
| State | Published - Nov 2022 |
| Externally published | Yes |