Projects per year
Abstract
Understanding how linguistic knowledge is encoded in language models is crucial for improving their generalisation capabilities. In this paper, we investigate the processing of morphosyntactic phenomena, by leveraging a recently proposed method for probing language models via Shapley Head Values (SHVs). Using the English language BLiMP dataset, we test our approach on two widely used models, BERT and RoBERTa, and compare how linguistic constructions such as anaphor agreement and filler-gap dependencies are handled. Through quantitative pruning and qualitative clustering analysis, we demonstrate that attention heads responsible for processing related linguistic phenomena cluster together. Our results show that SHV-based attributions reveal distinct patterns across both models, providing insights into how language models organize and process linguistic information. These findings support the hypothesis that language models learn subnetworks corresponding to linguistic theory, with potential implications for cross-linguistic model analysis and interpretability in Natural Language Processing (NLP).
Original language | English |
---|---|
Title of host publication | Findings of the Association for Computational Linguistics: NAACL 2025 |
Publisher | Association for Computational Linguistics |
Publication date | 29 Apr 2025 |
DOIs | |
Publication status | Accepted/In press - 23 Jan 2025 |
Event | The 2025 Annual Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics - Albuquerque, United States Duration: 29 Apr 2025 → 4 May 2025 https://2025.naacl.org/ |
Conference
Conference | The 2025 Annual Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics |
---|---|
Country/Territory | United States |
City | Albuquerque |
Period | 29/04/2025 → 04/05/2025 |
Internet address |
Fingerprint
Dive into the research topics of 'Linguistically Grounded Analysis of Language Models using Shapley Head Values'. Together they form a unique fingerprint.Projects
- 1 Active
-
Multilingual Modelling for Resource-Poor Languages
Bjerva, J. (PI), Lent, H. C. (Project Participant), Chen, Y. (Project Participant), Ploeger, E. (Project Participant), Fekete, M. R. (Project Participant) & Lavrinovics, E. (Project Participant)
01/09/2022 → 31/08/2025
Project: Research
-
Analysing Language Model Knowledge using Linguistic Theory
Fekete, M. R. (Lecturer)
20 Feb 2025Activity: Talks and presentations › Guest lecturers
-
1st Annual AAU NLP Symposium
Lavrinovics, E. (Organizer) & Bjerva, J. (Organizer)
3 Dec 2024Activity: Attending an event › Conference organisation or participation
-
Do Language Models Dream With Linguistics?
Fekete, M. R. (Lecturer)
2 Dec 2024Activity: Talks and presentations › Conference presentations