Entry #1: Thinking beyond the anthropomorphic paradigm benefits LLM research

Table of contents

Lujain Ibrahim and Myra Cheng's position paper addresses a critical and growing trend in Artificial Intelligence: the pervasive use of human characteristics and analogies – anthropomorphism – to describe and develop Large Language Models (LLMs). The authors provide empirical evidence for this trend and argue that while such analogies can sometimes be productive, deliberately moving beyond them is essential for future progress in LLM research. This analysis evaluates the paper's key contributions, examines potential limitations in its arguments and scope, and identifies crucial areas requiring further investigation.
Strengths
Empirical Grounding: A significant strength is the paper's quantitative demonstration of the prevalence and recent growth of anthropomorphic terminology within computer science literature, particularly concerning LLMs. Using a modified "AnthroScore" metric on large datasets of research abstracts (arXiv and ACL Anthology, detailed in Section 3, Figures 2 & 3), the authors provide valuable empirical backing for their central claims.
Structured Conceptual Framework: The paper effectively moves beyond just critiquing language by identifying five core anthropomorphic assumptions that influence research methodologies throughout the LLM development lifecycle: Training approaches (Section 4.1), Alignment strategies (4.2), Capability measurement (4.3), Understanding model behavior (4.4), and User interaction paradigms (4.5). This framework offers a clear lens for analyzing the deeper impact of human-centric thinking.
Highlighting Viable Alternatives: For each identified assumption, the authors constructively point to specific, non-anthropomorphic research avenues currently being pursued. Examples cited include byte-level tokenization instead of word-based units, latent-space reasoning (4.1), alignment via precise specifications and control theory rather than mimicking human values (4.2), developing mechanistic evaluation methods beyond human benchmarks (4.3), reframing concepts like "hallucination" as statistical artifacts (4.4), and designing structured, non-conversational user interfaces (4.5). This demonstrates that moving beyond anthropomorphism is practically feasible.
Linking Language to Thought: The paper effectively argues that the choice of terminology is not merely stylistic but actively reflects and shapes underlying conceptualizations of LLMs (Abstract, Introduction, page 1), drawing on established ideas about language and cognition (e.g., Lakoff & Johnson, 2008). This validates the focus on linguistic patterns as indicators of potentially limiting assumptions.
Weaknesses
Understated Critical Force: While identifying potentially problematic assumptions, the paper's critique often feels tentative (e.g., assumptions "may be limiting"). It carefully documents potential issues but stops short of forcefully arguing that certain anthropomorphic premises are fundamentally flawed or lead to significant scientific errors or harms. This measured tone, while perhaps diplomatic, can understate the potential negative consequences of building research on potentially unsound analogies.
Insufficient Analysis of High-Risk Areas: The paper's own data reveals that anthropomorphic framing is most prevalent in crucial fields like Safety/Ethics and Interpretability (Figure 4). This is a striking finding, suggesting the areas tasked with managing risk and ensuring understanding might be most affected by potentially misleading analogies. However, the paper doesn't fully explore the specific implications of this concentration, using the data mainly to support the general point about prevalence rather than deeply analyzing why these fields are susceptible and the specific dangers this poses.
Lack of Sharp Distinction Between Communication and Methodology: The paper acknowledges that anthropomorphism can be useful, sometimes citing its role in productive research (page 1). However, it could more sharply distinguish between using anthropomorphism as a potentially harmless communicative shortcut versus embedding it as a foundational principle in scientific hypotheses, experimental design, evaluation metrics, or theoretical models. This distinction is critical for determining when anthropomorphism becomes genuinely problematic.
Hesitation to Fundamentally Challenge Premises: The argument is framed around "moving beyond our default reliance" (page 1) and exploring alternatives. It largely avoids directly confronting whether certain widely used anthropomorphic premises (e.g., attributing human-like reasoning processes based on output) are simply incorrect representations of LLM mechanisms. Presenting non-anthropomorphic approaches as helpful alternatives differs significantly from arguing they are necessary correctives based on the technology's actual nature.
Unexplored
The Crucial Role of User Psychology: The paper focuses predominantly on researcher assumptions and methodologies. It significantly under-explores the user's side of the interaction – how diverse individuals perceive LLMs, form mental models, interpret outputs, and how factors like technical literacy, cognitive biases, and personal worldview (Umwelt, or subjective experience) shape the interaction's success and perceived safety. Understanding this psychological dimension is vital, as the same LLM output can lead to vastly different interpretations and consequences depending on the user.
Need for a Cohesive Non-Anthropomorphic Framework: While pointing to various alternative techniques, the paper doesn't synthesize or advocate for a comprehensive, alternative theoretical framework for understanding LLMs grounded primarily in computational, statistical, and information-theoretic principles. Developing such a unifying framework could provide a more robust foundation for research and development than the current collection of disparate non-anthropomorphic approaches.
Understanding the Drivers of Anthropomorphism: The paper documents the trend but delves little into the underlying reasons for its persistence and growth. A deeper analysis of the sociological, psychological, institutional (e.g., funding incentives, publication pressures), and pragmatic factors driving researchers towards anthropomorphic framing could reveal crucial levers for encouraging alternative perspectives.
Developing Holistic Diagnostic Approaches: Future work should emphasize diagnostic methods that analyze interaction problems systemically. Instead of defaulting to "fixing the model" when users report issues, methods are needed to explicitly investigate the interplay between the model's output, the user's interpretation and expectations, and the interface design. This requires gathering contextual information from the user, not just analyzing the model's log files, to correctly identify the problem's origin.
Conclusion
Ibrahim and Cheng offer a valuable contribution by empirically highlighting the prevalence of anthropomorphism in LLM research and structuring a critique around core underlying assumptions. The paper successfully maps the terrain and points towards promising alternative research directions. However, its impact is potentially limited by a cautious critical stance that may understate the risks associated with flawed analogies, particularly in high-stakes areas like AI safety. Furthermore, a significant gap remains in addressing the crucial role of user psychology and interpretation. While a commendable step, truly advancing the field requires not only acknowledging alternatives but potentially building a more robust, mechanistically grounded, and psychologically informed foundation for understanding and developing these powerful technologies.
Subscribe to my newsletter
Read articles from Gerard Sans directly inside your inbox. Subscribe to the newsletter, and don't miss out.
Written by

Gerard Sans
Gerard Sans
I help developers succeed in Artificial Intelligence and Web3; Former AWS Amplify Developer Advocate. I am very excited about the future of the Web and JavaScript. Always happy Computer Science Engineer and humble Google Developer Expert. I love sharing my knowledge by speaking, training and writing about cool technologies. I love running communities and meetups such as Web3 London, GraphQL London, GraphQL San Francisco, mentoring students and giving back to the community.