Papers
arxiv:2506.12450

Language Surgery in Multilingual Large Language Models

Published on Jun 14
· Submitted by joanitolopo on Jun 17
Authors:
,
,
,
,
,

Abstract

Research confirms natural representation alignment in large language models and introduces Inference-Time Language Control to enhance cross-lingual performance.

AI-generated summary

Large Language Models (LLMs) have demonstrated remarkable generalization capabilities across tasks and languages, revolutionizing natural language processing. This paper investigates the naturally emerging representation alignment in LLMs, particularly in the middle layers, and its implications for disentangling language-specific and language-agnostic information. We empirically confirm the existence of this alignment, analyze its behavior in comparison to explicitly designed alignment models, and demonstrate its potential for language-specific manipulation without semantic degradation. Building on these findings, we propose Inference-Time Language Control (ITLC), a novel method that leverages latent injection to enable precise cross-lingual language control and mitigate language confusion in LLMs. Our experiments highlight ITLC's strong cross-lingual control capabilities while preserving semantic integrity in target languages. Furthermore, we demonstrate its effectiveness in alleviating the cross-lingual language confusion problem, which persists even in current large-scale LLMs, leading to inconsistent language generation. This work advances our understanding of representation alignment in LLMs and introduces a practical solution for enhancing their cross-lingual performance.

Community

Paper author Paper submitter

We’re excited to share our latest work, "Language Surgery in Multilingual Large Language Models". We proposed a method, named Inference-Time Language Control (ITLC), designed to enhance cross-lingual language control and mitigate language confusion in Large Language Models (LLMs). ITLC leverages latent injection to enable precise manipulation of language-specific information during inference, while preserving semantic integrity. By exploiting representation alignment in LLMs’ middle layers, ITLC achieves zero-shot cross-lingual generation (10.70 average BLEU), mitigates language confusion (2.7x better LCPR, 4x better LPR), and allows language-specific manipulation without compromising meaning. Key contributions include confirming representation alignment via cosine similarity analysis and providing a practical solution for cross-lingual tasks. ITLC’s applications include enabling zero-shot cross-lingual generation and ensuring consistent language output.

📖 Paper: http://arxiv.org/abs/2506.12450
💻 Code: https://github.com/SEACrowd/itlc

This is an automated message from the Librarian Bot. I found the following papers similar to this paper.

The following papers were recommended by the Semantic Scholar API

Please give a thumbs up to this comment if you found it helpful!

If you want recommendations for any Paper on Hugging Face checkout this Space

You can directly ask Librarian Bot for paper recommendations by tagging it in a comment: @librarian-bot recommend

Sign up or log in to comment

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2506.12450 in a model README.md to link it from this page.

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2506.12450 in a dataset README.md to link it from this page.

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2506.12450 in a Space README.md to link it from this page.

Collections including this paper 0

No Collection including this paper

Add this paper to a collection to link it from this page.