Exploring Universal Sentence Encoders for Zero-shot Text Classification

Souvika Sarkar, Dongji Feng, Shubhra Kanti Karmaker Santu

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

Universal Sentence Encoder (USE) has gained much popularity recently as a general-purpose sentence encoding technique. As the name suggests, USE is designed to be fairly general and has indeed been shown to achieve superior performances for many downstream NLP tasks. In this paper, we present an interesting “negative” result on USE in the context of zero-shot text classification, a challenging task, which has recently gained much attraction. More specifically, we found some interesting cases of zero-shot text classification, where topic based inference outperformed USE-based inference in terms of F1 score. Further investigation revealed that USE struggles to perform well on data-sets with a large number of labels with high semantic overlaps, while topic-based classification works well for the same.
Original languageAmerican English
Title of host publicationProceedings of the 2nd Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics and the 12th International Joint Conference on Natural Language Processing
EditorsYulan He, Heng Ji, Sujian Li, Yang Liu, Chua-Hui Chang
PublisherAssociation for Computational Linguistics
Pages135-147
Volume2
StatePublished - Nov 2022
Externally publishedYes

Cite this