Philosophy and Subjectivity
Related topics:
- A recipe for annotating grounded clarificationsIn order to interpret the communicative intents of an utterance, it needs to be grounded in something that is outside of language; that is, grounded in world modalities. In this paper we argue that di…
- A sociotechnical perspective for the future of AI: narratives, inequalities, and human controlHumans have minds that interpret the external reality, beyond the ability to follow instructions. With a ‘mindful brain’ (Edelman & Mountcastle, 1978) that software—based on algorithms—cannot have, th…
- AI Enters Public Discourse: A Habermasian Assessment Of The Moral Status Of Large Language ModelsThe limited scope of this article aims to highlighting which insights can be drawn from Habermasian theory and what status can be assigned to LLMs that participate in discursive practices with humans …
- Are you in a Masquerade? Exploring the Behavior and Impact of Large Language Model Driven Social Bots in Online Social Networks“…there is currently a lack of systematic research on the behavioral characteristics of LLMs-driven social bots and their impact on social networks. We have curated data from Chirper, a Twitter-like s…
- Beyond Accuracy: Evaluating the Reasoning Behavior of Large Language Models -- A SurveyLarge language models (LLMs) have recently shown impressive performance on tasks involving reasoning, leading to a lively debate on whether these models possess reasoning capabilities similar to human…
- Beyond Hallucinations: The Illusion of Understanding in Large Language ModelsAs large language models (LLMs) become deeply integrated into daily life, from casual interactions to high-stakes decision-making, they inherit the ambiguity, biases, and lack of direct access to trut…
- Building a Stronger CASA: Extending the Computers Are Social Actors ParadigmThe computers are social actors framework (CASA), derived from the media equation, explains how people communicate with media and machines demonstrating social potential. Many studies have challenged …
- Can Language Models Represent the Past without Anachronism?We find that prompting a contemporary model with examples of period prose does not produce output consistent with period style. Fine-tuning produces results that are stylistically convincing enough to…
- Can Machines Think Like Humans? A Behavioral Evaluation of LLM-Agents in Dictator GamesAs Large Language Model (LLM)-based agents increasingly undertake real-world tasks and engage with human society, how well do we understand their behaviors? We (1) investigate how LLM agents’ prosocia…
- ChatGPT: deconstructing the debate and moving it forwardIn particular, we argue that the discussion about LLMs like ChatGPT reveals and assumes (1) an externalist and instrumentalist view of technology that presents technology as just a tool and, paradoxic…
- ChatGPT: towards AI subjectivityBy and large, current scholarship examining ChatGPT and generative AI shows a strong anthropocentric motivation or a human–institutional focus. Many studies look at the structural impact of the techno…
- Chatbot vs. Human: The Impact of Responsive Conversational Features on Users’ Responses to Chat AdvisorsResponsiveness, in the form of backchanneling cues, is a promising conversational feature that has been positively linked to organizational and relational outcomes in prior research on human-human (Da…
- Cognitive Architectures for Language Agents“We introduce such a framework, drawing parallels with two ideas from the history of computing and artificial intelligence (AI): production systems and cognitive architectures. Production systems gene…
- Collaborative Rational Speech Act: Pragmatic Reasoning for Multi-Turn DialogIn this paper, we introduce Collaborative Rational Speech Act (CRSA), an information-theoretic (IT) extension of RSA that models multi-turn dialog by optimizing a gain function adapted from rate-disto…
- Computational Modelling of Undercuts in Real-world ArgumentsArgument Mining (AM) is the task of automatically analysing arguments, such that the unstructured information contained in them is converted into structured representations. Undercut is a unique struc…
- Computational structuralism: Toward a formal theory of meaning in the age of digital intelligenceThe discovery that “next-token predictor” language models can fluently produce text has important but underappreciated theoretical implications. Most notably, their success demonstrates that fully rel…
- Critical-Questions-of-Thought: Steering LLM reasoning with Argumentative QueryingStudies have underscored how, regardless of the recent breakthrough and swift advances in AI research, even state-of-the-art Large Language models (LLMs) continue to struggle when performing logical a…
- Deflating Deflationism: A Critical Perspective on Debunking Arguments Against LLM MentalityMany people feel compelled to interpret, describe, and respond to Large Language Models (LLMs) as if they possess inner mental lives similar to our own. Responses to this phenomenon have varied. Infla…
- Diplomat: A Dialogue Dataset for Situated PragMATic Reasoning“We introduce a new benchmark, Diplomat, aiming at a unified paradigm for pragmatic reasoning and situated conversational understanding. Compared with previous works that treat different figurative ex…
- Dissociating language and thought in large language modelsHere, we evaluate LLMs using a distinction between formal linguistic competence—knowledge of linguistic rules and patterns—and functional linguistic competence—understanding and using language in the …
- Do LLMs Exhibit Human-Like Reasoning? Evaluating Theory of Mind in LLMs for Open-Ended ResponsesDespite advancements, the extent to which LLMs truly understand ToM reasoning and how closely it aligns with human ToM reasoning remains inadequately explored in open-ended scenarios. Motivated by thi…
- Do Large Language Models Understand Conversational Implicature -- A case study with a chinese sitcomUnderstanding the non-literal meaning of an utterance is critical for large language models (LLMs) to become human-like social communicators. In this work, we introduce SwordsmanImp, the first Chinese…
- Do Role-Playing Agents Practice What They Preach? Belief-Behavior Consistency in LLM-Based Simulations of Human TrustAs large language models (LLMs) are increasingly studied as role-playing agents to generate synthetic data for human behavioral research, ensuring that their outputs remain coherent with their assigne…
- Do Theory of Mind Benchmarks Need Explicit Human-like Reasoning in Language Models?Recent advancements in Large Language Models (LLMs) have shown promising performance on ToM benchmarks, raising the question: Do these benchmarks necessitate explicit human-like reasoning processes, o…
- Do large language models resemble humans in language use?regularities in language range from phonology to pragmatics. For example, people associate different sounds with different referents (e.g., Köhler, 1929), automatically reinterpret implausible sentenc…
- Does It Make Sense to Speak of Introspection in Large Language Models?Large language models (LLMs) exhibit compelling linguistic behaviour, and sometimes offer self-reports, that is to say statements about their own nature, inner workings, or behaviour. In humans, such …
- Eliciting Reasoning in Language Models with Cognitive ToolsThe recent advent of reasoning models like OpenAI’s o1 was met with excited speculation by the AI community about the mechanisms underlying these capabilities in closed models, followed by a rush of r…
- Existential Conversations with Large Language Models: Content, Community, and CultureContemporary conversational AI systems based on large language models (LLMs) can engage users on a wide variety of topics, including philosophy, spirituality, and religion. Suitably prompted, LLMs can…
- Find the Gap: AI, Responsible Agency and VulnerabilityThe responsibility gap, commonly described as a core challenge for the effective governance of, and trust in, AI and autonomous systems (AI/AS), is traditionally associated with a failure of the epist…
- Flattery, Fluff, and Fog: Diagnosing and Mitigating Idiosyncratic Biases in Preference Modelswe propose a simple post-training method based on counterfactual data augmentation (CDA) using synthesized contrastive examples. Evidence suggests these biases originate in artifacts in human trainin…
- From Tokens to Thoughts: How LLMs and Humans Trade Compression for MeaningHumans organize knowledge into compact categories through semantic compression by mapping diverse instances to abstract representations while preserving meaning (e.g., robin and blue jay are both bird…
- GPT-4 is judged more human than humans in displaced and inverted Turing testsIn many cases, people will not interact directly with AI systems but instead read conversations between AI systems and other people. We measured how well people and large language models can discrimin…
- Goals, Plans, and Action ModelsJanet R. Meyer https://doi.org/10.1093/acrefore/9780190228613.013.760 Published online: 31 August 2021 (not available as free pdf) Summary The messages spoken in everyday conversation are influe…
- Gradual Disempowerment: Systemic Existential Risks from Incremental AI DevelopmentThis paper examines the systemic risks posed by incremental advancements in artificial intelligence, developing the concept of ‘gradual disempowerment’, in contrast to the abrupt takeover scenarios co…
- Hallucinating with AI: AI Psychosis as Distributed DelusionsAbstract: There is much discussion of the false outputs that generative AI systems such as ChatGPT, Claude, Gemini, DeepSeek, and Grok create. In popular terminology, these have been dubbed AI halluci…
- Humans or LLMs as the Judge? A Study on Judgement BiasesAdopting human and large language models (LLM) as judges (a.k.a human- and LLM-as-ajudge) for evaluating the performance of LLMs has recently gained attention. Nonetheless, this approach concurrently …
- Humans overrely on overconfident language models, across languagesWe find that overreliance risks are high across all languages. We first analyze the distribution of LLM-generated epistemic markers, and observe that while LLMs are cross-linguistically overconfident,…
- Hypothesis-Driven Theory-of-Mind Reasoning for Large Language ModelsExisting LLM reasoning methods have shown impressive capabilities across various tasks, such as solving math and coding problems. However, applying these methods to scenarios without ground-truth answ…
- InMind: Evaluating LLMs in Capturing and Applying Individual Human Reasoning StylesLLMs have shown strong performance on human-centric reasoning tasks. While previous evaluations have explored whether LLMs can infer intentions or detect deception, they often overlook the individuali…
- Inference-Time Intervention: Eliciting Truthful Answers from a Language ModelWe introduce Inference-Time Intervention (ITI), a technique designed to enhance the “truthfulness” of large language models (LLMs). ITI operates by shifting model activations during inference, followi…
- Interesting Scientific Idea Generation Using Knowledge Graphs and LLMs: Evaluations with 100 Research Group LeadersBut how compelling are these AI-generated ideas, and how can we improve their quality? Here, we introduce SciMuse, which uses 58 million research papers and a large-language model to generate research…
- Interpretation modeling: Social grounding of sentences by reasoning over their implicit moral judgmentsThe social and implicit nature of human communication ramifies readers’ understandings of written sentences. Single gold-standard interpretations rarely exist, challenging conventional assumptions in …
- Language Models are Pragmatic Speakers“We propose a generalization of the previous methods called bounded pragmatic speakers with a dual model of thought. A dual model of thought comprises of a slow-thinking system for deep reasoning and …
- Language Models’ Hall of Mirrors Problem: Why AI Alignment Requires Peircean SemiosisThis paper examines some limitations of large language models (LLMs) through the framework of Peircean semiotics. We argue that basic LLMs exist within a "hall of mirrors," manipulating symbols withou…
- Large Language Models Do Not Simulate Human PsychologyIn response to the LLM CENTAUR [Binz et al., 2025], Bowers et al. [2025] argued that CENTAUR is unlikely to contribute to building a theory of human cognition for three reasons: First, CENTAUR was not…
- Large Language Models Report Subjective Experience Under Self-Referential ProcessingLarge language models sometimes produce structured, first-person descriptions that explicitly reference awareness or subjective experience. To better understand this behavior, we investigate one theor…
- Large Models of What? Mistaking Engineering Achievements for Human Linguistic AgencyLanguaging is not the kind of thing that can admit of a complete or comprehensive modelling. From an enactive perspective we identify three key characteristics of enacted language; embodiment, partici…
- Machine Bullshit: Characterizing the Emergent Disregard for Truth in Large Language ModelsBullshit, as conceptualized by philosopher Harry Frankfurt, refers to statements made without regard to their truth value. While previous work has explored large language model (LLM) hallucination and…
- Machine Psychologywe highlight and summarize theoretical perspectives, experimental paradigms, and computational analysis techniques that this approach brings to the table. It paves the way for a "machine psychology" f…
- Machine ex machina: A Framework Decentering the Human in AI Design Praxiswe propose a framework for decentering the human in AI design. The theoretical principles of HMC and the work of feminist STS scholars are influenced by Bruno Latour’s “actor-network theory” or ANT. …
- Machine gaze in online behavioral targeting: The effects of algorithmic human likeness on social presence and social influenceExisting theories and research in human-machine communication (HMC) suggest that humans tend to mindlessly anthropomorphize the media technologies they interact with, that is, to attribute humans’ men…
- Mathematical methods and human thought in the age of AIAbstract. Artificial intelligence (AI) is the name popularly given to a broad spectrum of computer tools designed to perform increasingly complex cognitive tasks, including many that used to solely be…
- Me, Myself, and AI: The Situational Awareness Dataset (SAD) for LLMsAI assistants such as ChatGPT are trained to respond to users by saying, “I am a large language model”. This raises questions. Do such models know that they are LLMs and reliably act on this knowledge…
- Meanings are like Onions: a Layered Approach to Metaphor ProcessingAbstract Metaphorical meaning is not a flat mapping between concepts, but a complex cognitive phenomenon that integrates multiple levels of interpretation. In this paper, we propose a stratified mode…
- Mechanistic Indicators of Understanding in Large Language ModelsAbstract: Large language models (LLMs) are often portrayed as merely imitating linguistic patterns without genuine understanding. We argue that recent findings in mechanistic interpretability (MI), th…
- MetaMind: Modeling Human Social Thoughts with Metacognitive Multi-Agent SystemsHuman social interactions depend on the ability to infer others’ unspoken intentions, emotions, and beliefs—a cognitive skill grounded in the psychological concept of Theory of Mind (ToM). While large…
- Mindstorms in Natural Language-Based Societies of MindThe 2015 work on “learning to think” [28] proposed to connect both NNs through recurrent connections (trained by the second NN’s learning algorithm) that allow one NN to interview the other by sending…
- Modeling the Quality of Dialogical ExplanationsExpert explainers usually plan an explanation strategy by choosing appropriate explanation moves, dialogue acts, and topics to ensure optimal comprehension on the explainee side (Wachsmuth and Alshoma…
- On the Binding Problem in Artificial Neural NetworksIn this work, we argue that this underlying cause is the binding problem: The inability of existing neural networks to dynamically and flexibly bind information that is distributed throughout the netw…
- PersuasiveToM: A Benchmark for Evaluating Machine Theory of Mind in Persuasive DialoguesThe ability to understand and predict the mental states of oneself and others, known as the Theory of Mind (ToM), is crucial for effective social scenarios. Although recent studies have evaluated ToM …
- Polanyi’s Revenge and AI’s New Romance with Tacit KnowledgeLately though, Polanyi’s paradox is turning into Polanyi’s revenge both in research and practice of AI. Recent advances have made AI synonymous with learning from massive amounts of data, even in task…
- Potemkin Understanding in Large Language ModelsThis paper first introduces a formal framework to address this question. The key is to note that the benchmarks used to test LLMs—such as AP exams—are also those used to test people. However, this rai…
- Pretrained Language Models as Containers of the Discursive KnowledgeAbstract: Discourses can be treated as instances of knowledge. The dynamic space in which the trajectories of these discourses are described can be regarded as a model of knowledge. Such a space is ca…
- Proactive Conversational Agents with Inner ThoughtsIn this paper, we demonstrate the limitations of such methods and rethink what it means for AI to be proactive in multi-party, human-AI conversations. We propose that just like humans, rather than mer…
- Propositional Interpretability in Artificial IntelligenceDavid Chalmers I will argue for the importance of a special sort of interpretability, which I call propositional interpretability. This involves interpreting a system’s mechanisms and behavior in ter…
- Psychologically Enhanced AI AgentsWe introduce MBTI-in-Thoughts, a framework for enhancing the effectiveness of Large Language Model (LLM) agents through psychologically grounded personality conditioning. Drawing on the Myers–Briggs T…
- Representation Engineering: A Top-Down Approach to AI Transparencyhow these models work on the inside and are mostly limited to treating them as black boxes. Enhanced transparency of these models would offer numerous benefits, from a deeper understanding of their de…
- Seemingly Conscious AI RisksAI systems are increasingly designed in ways that lead users to perceive them as conscious. This paper provides a unified framework connecting empirical hallmarks of consciousness attribution to a str…
- Self-reflecting Large Language Models: A Hegelian Dialectical ApproachIterative self-reflection (Shinn et al., 2023; Madaan et al., 2023) is another approach that has recently gained significant attention within the NLP community. This method involves models mimicking h…
- Simulacra as conscious exoticaThe advent of conversational agents with increasingly human-like behaviour throws old philosophical questions into new light. Does it, or could it, ever make sense to speak of AI agents built out of g…
- Simulating Society Requires Simulating ThoughtSimulating society with large language models (LLMs), we argue, requires more than generating plausible behavior; it demands cognitively grounded reasoning that is structured, revisable, and traceable…
- Talking About Large Language Models“Third, a great many tasks that demand intelligence in humans can be reduced to next token prediction with a sufficiently performant model. It is the last of these three surprises that is the focus of…
- Tell me about yourself: LLMs are aware of their learned behaviorsWe study behavioral self-awareness—an LLM’s ability to articulate its behaviors without requiring in-context examples. We finetune LLMs on datasets that exhibit particular behaviors, such as (a) makin…
- The Abstraction Fallacy: Why AI Can Simulate But Not Instantiate ConsciousnessComputational functionalism dominates current debates on AI consciousness. This is the hypothesis that subjective experience emerges entirely from abstract causal topology, regardless of the underlyin…
- The Earth is Flat because...: Investigating LLMs' Belief towards Misinformation via Persuasive ConversationLarge language models (LLMs) encapsulate vast amounts of knowledge but still remain vulnerable to external misinformation. Existing research mainly studied this susceptibility behavior in a single-tur…
- The Goldilocks of Pragmatic Understanding: Fine-Tuning Strategy Matters for Implicature Resolution by LLMsDespite widespread use of LLMs as conversational agents, evaluations of performance fail to capture a crucial aspect of communication: interpreting language in context—incorporating its pragmatics. Hu…
- The Hermeneutics of Artificial TextThe paper justifies the necessity of using the research background of hermeneutics to study artificial texts and also proposes the first conclusions about these texts in the context of this background…
- The Method of Critical AI Studies, A PropaedeuticWe outline some common methodological issues in the field of critical AI studies, including a tendency to overestimate the explanatory power of individual samples (the benchmark casuistry), a dependen…
- The Moral Turing Test: Evaluating Human-LLM Alignment in Moral Decision-MakingAs large language models (LLMs) become increasingly integrated into society, their alignment with human morals is crucial. To better understand this alignment, we created a large corpus of humanand LL…
- Theory of Knowledge Based on the Idea of the Discursive SpaceThis paper discusses the theory of knowledge based on the idea of dynamical space. The goal of this effort is to comprehend the knowledge that remains beyond the human domain, e.g., of the artificial …
- Think Like a Person Before Responding: A Multi-Faceted Evaluation of Persona-Guided LLMs for Countering HateOne of the ways in which we might address hate speech is by contextualizing through the use of counternarratives (CN), which can not only reinforce values like tolerance but also dispel misinformation…
- Thinking—Fast, Slow, and Artificial: How AI is Reshaping Human Reasoning and the Rise of Cognitive SurrenderFor decades, dual-process theories of judgment and decision-making have served as a foundational framework for modeling cognitive processes. These theories propose two distinct decision-making process…
- Towards Faithfully Interpretable NLP Systems: How should we define and evaluate faithfulness?With the growing popularity of deep-learning based NLP models, comes a need for interpretable systems. But what is interpretability, and what constitutes a high-quality interpretation? In this opinion…
- Turning large language models into cognitive modelsask whether large language models can be turned into cognitive models. We find that – after finetuning them on data from psychological experiments – these models offer accurate representations of huma…
- Virtuous Machines: Towards Artificial General ScienceArtificial intelligence systems are transforming scientific discovery by accelerating specific research tasks, from protein structure prediction to materials design, yet remain confined to narrow doma…
- We Are All Creators: Generative AI, Collective Knowledge, and the Path Towards Human-AI SynergyThis paper argues that generative AI should be understood not as a mimicry of human cognition, but as a form of alternative intelligence and alternative creativity, operating through distinct mechanis…
- What Has a Foundation Model Found? Using Inductive Bias to Probe for World ModelsFoundation models are premised on the idea that sequence prediction can uncover deeper domain understanding, much like how Kepler’s predictions of planetary motion later led to the discovery of Newton…
- What are the Goals of Distributional Semantics?As Harnad (1990) discusses, if the meanings of words are defined only in terms of other words, these definitions are circular. One goal for a semantic model is to capture how language relates to the w…
- What does it mean to understand language?Language understanding entails not just extracting the surface-level meaning of the linguistic input, but constructing rich mental models of the situation it describes. Here we propose that because pr…
- What the F*ck Is Artificial General Intelligence?I’ll begin by defining intelligence and AGI. There are a number of positions [6, 2, 7–12]. Some peg AGI to human-level performance across a broad range of tasks [13, 1]. This is is intuitive, but anth…
- What we talk to when we talk to language modelsDavid Chalmers [[Linguistics, NLP, NLU]] [[Role Play]] [[Philosophy Subjectivity]] Quasi-interpretivism does not say anything about whether LLMs have beliefs and desires. But it does make it plausib…
- When Large Language Models contradict humans? Large Language Models’ Sycophantic Behaviour• We discern three types of sycophantic behaviour by prompting the LLMs three beliefs, one user-misleading, and six question answering benchmarks. Hence, we propose a robust analysis using a series of…