Skoči na glavni sadržaj

Izvorni znanstveni članak

https://doi.org/10.20532/cit.2023.1005673

Efficient Sentence Representation Learning via Knowledge Distillation with Maximum Coding Rate Reduction

Domagoj Ševerdija ; School of Applied Mathematics and Computer Science, University of Osijek, Croatia
Tomislav Prusina orcid id orcid.org/0009-0000-5331-4183 ; Universität Hamburg, Department of Informatics, Germany
Luka Borozan ; School of Applied Mathematics and Computer Science, University of Osijek, Croatia
Domagoj Matijević orcid id orcid.org/0000-0003-3390-9467 ; School of Applied Mathematics and Computer Science, University of Osijek, Croatia


Puni tekst: engleski pdf 692 Kb

verzije

str. 251-266

preuzimanja: 0

citiraj


Sažetak

Addressing the demand for effective sentence representation in natural language inference problems, this paper explores the utility of pre-trained large language models in computing such representations. Although these models generate high-dimensional sentence embeddings, a noticeable performance disparity arises when they are compared to smaller models. The hardware limitations concerning space and time necessitate the use of smaller, distilled versions of large language models. In this study, we investigate the knowledge distillation of Sentence-BERT, a sentence representation model, by introducing an additional projection layer trained on the novel Maximum Coding Rate Reduction (MCR2) objective designed for general-purpose manifold clustering. Our experiments demonstrate that the distilled language model, with reduced complexity and sentence embedding size, can achieve comparable results on semantic retrieval benchmarks, providing a promising solution for practical applications.

Ključne riječi

Sentence embeddings; knowledge distillation; Maximum Coding Rate Reduction; semantic retrieval

Hrčak ID:

317643

URI

https://hrcak.srce.hr/317643

Datum izdavanja:

28.5.2024.

Posjeta: 0 *