Semantic Textual Similarity(STS) is one of the challenging problems in the field of natural language processing and plays an important role in many applications such as question answering, recommendation systems and information retrieval. For the STS tasks, previous work has primarily been based on encoder-only structural models, as researchers widely believe that such structure can better capture feature information compared to the other structural models. However, there has been a lack of systematic comparative studies to investigate the performance of different structural models on STS. To fill this gap, we systematically compared different structural models and explored methods for extracting sentence embeddings tailored to different structures and conducted extensive experiments across 7 text similarity datasets. The results show that, the decoder-only model LlaMA2 has shown superior overall performance on the SentEval benchmark without fine-tuning and as the parameters increased, the performance of the decoder-only model improves gradually. Additionally, for any model structure, without doing any projection, the intermediate layers of the model actually degrade the quality of sentence embeddings, affecting the model's performance on STS tasks.
ABSTRACT
PUBLICATION RECORD
- Publication year
2025
- Venue
Cybersecurity and Cyberforensics Conference
- Publication date
2025-07-28
- Fields of study
Not labeled
- Identifiers
- External record
- Source metadata
Semantic Scholar
CITATION MAP
EXTRACTION MAP
CLAIMS
- No claims are published for this paper.
CONCEPTS
- No concepts are published for this paper.
REFERENCES
Showing 1-15 of 15 references · Page 1 of 1
CITED BY
- No citing papers are available for this paper.
Showing 0-0 of 0 citing papers · Page 1 of 1