Representational Isomorphism and Alignment of Multilingual Large Language Models
| Authors | |
|---|---|
| Publication date | 2024 |
| Host editors |
|
| Book title | The 2024 Conference on Empirical Methods in Natural Language Processing : Findings of EMNLP 2024 |
| Book subtitle | EMNLP 2024 : November 12-16, 2024 |
| ISBN (electronic) |
|
| Event | 2024 Conference on Empirical Methods in Natural Language Processing |
| Pages (from-to) | 14074-14085 |
| Number of pages | 12 |
| Publisher | Kerrville, TX: Association for Computational Linguistics |
| Organisations |
|
| Abstract |
In this paper, we investigate the capability of Large Language Models (LLMs) to represent texts in multilingual contexts. Our findings show that sentence representations derived from LLMs exhibit a high degree of isomorphism across languages.This existing isomorphism can facilitate representational alignments in zero-shot and few-shot settings.Specifically, by applying a contrastive objective at the representation level with only a small number of translation pairs (e.g., 100), we substantially improve models’ performance on Semantic Textual Similarity (STS) tasks across languages. This representation-level approach proves to be more efficient and effective for semantic alignment than continued pretraining or instruction tuning. Interestingly, we also observe substantial STS improvements within individual languages, even without a monolingual objective specifically designed for this purpose.
|
| Document type | Conference contribution |
| Language | English |
| Related publication | Representational Isomorphism and Alignment of Multilingual Large Language Models |
| Published at | https://doi.org/10.18653/v1/2024.findings-emnlp.823 |
| Downloads |
2024.findings-emnlp.823
(Final published version)
|
| Permalink to this page | |