Citation:
Attieh , J , Zewoudie , A , Vlassov , V , Flanagan , A & Bäckström , T 2023 , Optimizing the Performance of Text Classification Models by Improving the Isotropy of the Embeddings using a Joint Loss Function . in 17th International Conference on Document Analysis and Recognition (ICDAR 2023) . Lecture notes in computer science , Springer , International Conference on Document Analysis and Recognition , San Jose , California , United States , 21/08/2023 . https://doi.org/10.1007/978-3-031-41734-4_8
|
Abstract:
Recent studies show that the spatial distribution of the sen- tence representations generated from pre-trained language models is highly anisotropic. This results in a degradation in the performance of the models on the downstream task. Most methods improve the isotropy of the sentence embeddings by refining the corresponding contextual word representations, then deriving the sentence embeddings from these refined representations. In this study, we propose to improve the quality of the sentence embeddings extracted from the [CLS] token of the pre- trained language models by improving the isotropy of the embeddings. We add one feed-forward layer between the model and the downstream task layers, and we train it using a novel joint loss function. The pro- posed approach results in embeddings with better isotropy, that gener- alize better on the downstream task. Experimental results on 3 GLUE datasets with classification as the downstream task show that our pro- posed method is on par with the state-of-the-art, as it achieves perfor- mance gains of around 2–3% on the downstream tasks compared to the baseline.
|