Evaluating Large Language Models on Linguistic Competence

Project Description

Recent LLMs have demonstrated remarkable capabilities across a range of tasks, yet their underlying linguistic competence remains less explored. As these models increasingly inform research and applications in the social sciences and beyond, a systematic evaluation of their linguistic abilities becomes critical. This project investigates the extent to which LLMs capture core aspects of linguistic knowledge, including syntax, semantics, pragmatics, and sociolinguistic variation. Drawing on methods from linguistics, cognitive science, and natural language processing, we design targeted evaluation tasks to probe specific linguistic phenomena, develop new benchmarks, and identify systematic strengths and weaknesses in current models. Our aim is to contribute to a more rigorous understanding of LLM capabilities and limitations, providing insights that are essential for both theoretical modeling and practical deployment.

Project Team

Publications

Contact person

Bolei Ma