Analyzing LLMs' Knowledge Boundary Cognition Across Languages Through the Lens of Internal Representations
Abstract
While understanding the knowledge boundaries of LLMs is crucial to prevent hallucination, research on knowledge boundaries of LLMs has predominantly focused on English. In this work, we present the first study to analyze how LLMs recognize knowledge boundaries across different languages by probing their internal representations when processing known and unknown questions in multiple languages. Our empirical studies reveal three key findings: 1) LLMs' perceptions of knowledge boundaries are encoded in the middle to middle-upper layers across different languages. 2) Language differences in knowledge boundary perception follow a linear structure, which motivates our proposal of a training-free alignment method that effectively transfers knowledge boundary perception ability across languages, thereby helping reduce hallucination risk in low-resource languages; 3) Fine-tuning on bilingual question pair translation further enhances LLMs' recognition of knowledge boundaries across languages. Given the absence of standard testbeds for cross-lingual knowledge boundary analysis, we construct a multilingual evaluation suite comprising three representative types of knowledge boundary data. Our code and datasets are publicly available at https://github.com/DAMO-NLP-SG/LLM-Multilingual-Knowledge-Boundaries.
Community
We present the first study to analyze how LLMs recognize knowledge boundaries across different languages by probing their internal representations when processing known and
unknown questions in multiple languages. And found that:
- Knowledge boundary perception is encoded in mid to mid-upper layers of LLMs.
- Cross-lingual knowledge boundary presents a linear structure, allowing training-free alignment.
- further bilingual translation SFT enhances knowledge boundary perception across languages, including those not directly trained on.
We also release 3 high-quality linguist verified multilingual knowledge boundary evaluation datasets.
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- High-Dimensional Interlingual Representations of Large Language Models (2025)
- Lost in Multilinguality: Dissecting Cross-lingual Factual Inconsistency in Transformer Language Models (2025)
- Uncovering inequalities in new knowledge learning by large language models across different languages (2025)
- Probing LLMs for Multilingual Discourse Generalization Through a Unified Label Set (2025)
- Language Models' Factuality Depends on the Language of Inquiry (2025)
- Can you map it to English? The Role of Cross-Lingual Alignment in Multilingual Performance of LLMs (2025)
- XCOMPS: A Multilingual Benchmark of Conceptual Minimal Pairs (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper