Toward cultural interpretability: A linguistic anthropological framework for describing and evaluating large language models

This article proposes a new integration of linguistic anthropology and machine learning (ML) around convergent interests in both the underpinnings of language and making language technologies more socially responsible. While linguistic anthropology focuses on interpreting the cultural basis for huma...

Full description

Saved in:
Bibliographic Details
Main Authors: Graham M Jones, Shai Satran, Arvind Satyanarayan
Format: Article
Language:English
Published: SAGE Publishing 2025-03-01
Series:Big Data & Society
Online Access:https://doi.org/10.1177/20539517241303118
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1832576401385979904
author Graham M Jones
Shai Satran
Arvind Satyanarayan
author_facet Graham M Jones
Shai Satran
Arvind Satyanarayan
author_sort Graham M Jones
collection DOAJ
description This article proposes a new integration of linguistic anthropology and machine learning (ML) around convergent interests in both the underpinnings of language and making language technologies more socially responsible. While linguistic anthropology focuses on interpreting the cultural basis for human language use, the ML field of interpretability is concerned with uncovering the patterns that Large Language Models (LLMs) learn from human verbal behavior. Through the analysis of a conversation between a human user and an LLM-powered chatbot, we demonstrate the theoretical feasibility of a new, conjoint field of inquiry, cultural interpretability (CI). By focusing attention on the communicative competence involved in the way human users and AI chatbots coproduce meaning in the articulatory interface of human-computer interaction, CI emphasizes how the dynamic relationship between language and culture makes contextually sensitive, open-ended conversation possible. We suggest that, by examining how LLMs internally “represent” relationships between language and culture, CI can: (1) provide insight into long-standing linguistic anthropological questions about the patterning of those relationships; and (2) aid model developers and interface designers in improving value alignment between language models and stylistically diverse speakers and culturally diverse speech communities. Our discussion proposes three critical research axes: relativity, variation, and indexicality.
format Article
id doaj-art-46b81277166d4121b72dda7f970b5477
institution Kabale University
issn 2053-9517
language English
publishDate 2025-03-01
publisher SAGE Publishing
record_format Article
series Big Data & Society
spelling doaj-art-46b81277166d4121b72dda7f970b54772025-01-31T06:03:21ZengSAGE PublishingBig Data & Society2053-95172025-03-011210.1177/20539517241303118Toward cultural interpretability: A linguistic anthropological framework for describing and evaluating large language modelsGraham M Jones0Shai Satran1Arvind Satyanarayan2 Anthropology, Massachusetts Institute of Technology, Cambridge, MA, USA Sociology and Anthropology, Tel Aviv University, Tel Aviv, IL Computer Science and Artificial Intelligence (CSAIL), Massachusetts Institute of Technology, Cambridge, MA, USAThis article proposes a new integration of linguistic anthropology and machine learning (ML) around convergent interests in both the underpinnings of language and making language technologies more socially responsible. While linguistic anthropology focuses on interpreting the cultural basis for human language use, the ML field of interpretability is concerned with uncovering the patterns that Large Language Models (LLMs) learn from human verbal behavior. Through the analysis of a conversation between a human user and an LLM-powered chatbot, we demonstrate the theoretical feasibility of a new, conjoint field of inquiry, cultural interpretability (CI). By focusing attention on the communicative competence involved in the way human users and AI chatbots coproduce meaning in the articulatory interface of human-computer interaction, CI emphasizes how the dynamic relationship between language and culture makes contextually sensitive, open-ended conversation possible. We suggest that, by examining how LLMs internally “represent” relationships between language and culture, CI can: (1) provide insight into long-standing linguistic anthropological questions about the patterning of those relationships; and (2) aid model developers and interface designers in improving value alignment between language models and stylistically diverse speakers and culturally diverse speech communities. Our discussion proposes three critical research axes: relativity, variation, and indexicality.https://doi.org/10.1177/20539517241303118
spellingShingle Graham M Jones
Shai Satran
Arvind Satyanarayan
Toward cultural interpretability: A linguistic anthropological framework for describing and evaluating large language models
Big Data & Society
title Toward cultural interpretability: A linguistic anthropological framework for describing and evaluating large language models
title_full Toward cultural interpretability: A linguistic anthropological framework for describing and evaluating large language models
title_fullStr Toward cultural interpretability: A linguistic anthropological framework for describing and evaluating large language models
title_full_unstemmed Toward cultural interpretability: A linguistic anthropological framework for describing and evaluating large language models
title_short Toward cultural interpretability: A linguistic anthropological framework for describing and evaluating large language models
title_sort toward cultural interpretability a linguistic anthropological framework for describing and evaluating large language models
url https://doi.org/10.1177/20539517241303118
work_keys_str_mv AT grahammjones towardculturalinterpretabilityalinguisticanthropologicalframeworkfordescribingandevaluatinglargelanguagemodels
AT shaisatran towardculturalinterpretabilityalinguisticanthropologicalframeworkfordescribingandevaluatinglargelanguagemodels
AT arvindsatyanarayan towardculturalinterpretabilityalinguisticanthropologicalframeworkfordescribingandevaluatinglargelanguagemodels