Loading...
How a Bilingual LM Becomes Bilingual: Tracing Internal Representations with Sparse Autoencoders
Inaba, Tatsuro ; Kamoda, Go ; Inui, Kentaro ; Isonuma, Masaru ; Miyao, Yusuke ; Oseki, Yohei ; Takagi, Yu ; Heinzerling, Benjamin
Inaba, Tatsuro
Kamoda, Go
Inui, Kentaro
Isonuma, Masaru
Miyao, Yusuke
Oseki, Yohei
Takagi, Yu
Heinzerling, Benjamin
Files
Supervisor
Department
Natural Language Processing
Embargo End Date
Type
Conference proceeding
Date
License
http://creativecommons.org/licenses/by/4.0/
Language
Collections
Research Projects
Organizational Units
Journal Issue
Abstract
This study explores how bilingual language models develop complex internal representations.We employ sparse autoencoders to analyze internal representations of bilingual language models with a focus on the effects of training steps, layers, and model sizes.Our analysis shows that language models first learn languages separately, and then gradually form bilingual alignments, particularly in the mid layers. We also found that this bilingual tendency is stronger in larger models.Building on these findings, we demonstrate the critical role of bilingual representations in model performance by employing a novel method that integrates decomposed representations from a fully trained model into a mid-training model.Our results provide insights into how language models acquire bilingual capabilities.
Citation
T. Inaba, G. Kamoda, K. Inui, M. Isonuma, Y. Miyao, Y. Oseki, Y. Takagi, B. Heinzerling, "How a Bilingual LM Becomes Bilingual: Tracing Internal Representations with Sparse Autoencoders," 2025, pp. 13458-13470.
Source
Findings of the Association for Computational Linguistics: EMNLP 2025
Conference
Findings of the Association for Computational Linguistics: EMNLP 2025
Keywords
Subjects
Source
Findings of the Association for Computational Linguistics: EMNLP 2025
Publisher
Association for Computational Linguistics (ACL)
