Natural Language Inference
Related topics:
- (QA)2: Question Answering with Questionable AssumptionsFor instance, the question When did Marie Curie discover Uranium? cannot be answered as a typical when question without addressing the false assumption Marie Curie discovered Uranium. In this work, we…
- A Hybrid Intelligence Method for Argument MiningLarge-scale survey tools enable the collection of citizen feedback in opinion corpora. Extracting the key arguments from a large and noisy set of opinions helps in understanding the opinions quickly a…
- A Robustness Evaluation Framework for Argument Mininglittle is usually known about the model’s stability and consistency when deployed in real-world settings. In this paper, we propose a robustness evaluation framework to guide the design of rigorous ar…
- A ripple in time: a discontinuity in American historyAbstract—In this technical note we suggest a novel approach to discover temporal (related and unrelated to language dilation) and personality (authorship attribution) aspects in historical datasets. W…
- Adam's Law: Textual Frequency Law on Large Language ModelsWhile textual frequency has been validated as relevant to human cognition in reading speed, its relatedness to Large Language Models (LLMs) is seldom studied. We propose a novel research direction in …
- Are Customers Lying to Your Chatbot?Dishonesty is far from a new phenomenon. But as chatbots, online forms, and other digital interfaces grow more and more common across a wide range of customer service applications, bending the truth t…
- Attention, Intentions, And The Structure Of DiscourseIn this paper we explore a new theory of discourse structure that stresses the role of purpose and processing in discourse. In this theory, discourse structure is composed of three separate but interr…
- AutoPrompt: Eliciting Knowledge from Language Models with Automatically Generated Promptswe develop AUTOPROMPT, an automated method to create prompts for a diverse set of tasks, based on a gradient-guided search. Using AUTOPROMPT, we show that masked language models (MLMs) have an inheren…
- Automatic Extraction of Metaphoric Analogies from Literary Texts: Task Formulation, Dataset Construction, and EvaluationExtracting metaphors and analogies from free text requires high-level reasoning abilities such as abstraction and language understanding. Our study focuses on the extraction of the concepts that form …
- Branch-Solve-Merge Improves Large Language Model Evaluation and GenerationLarge Language Models (LLMs) are frequently used for multi-faceted language generation and evaluation tasks that involve satisfying intricate user constraints or taking into account multiple aspects a…
- Can LLMs Ground when they (Don't) Know: A Study on Direct and Loaded Political QuestionsCommunication among humans relies on conversational grounding, allowing interlocutors to reach mutual understanding even when they do not have perfect knowledge and must resolve discrepancies in each …
- Can LLMs assist with Ambiguity? A Quantitative Evaluation of various Large Language Models on Word Sense DisambiguationAmbiguous words are often found in modern digital communications. Lexical ambiguity challenges traditional Word Sense Disambiguation (WSD) methods, due to limited data. Consequently, the efficiency of…
- Chain of Stance: Stance Detection with Large Language ModelsStance detection is an active task in natural language processing (NLP) that aims to identify the author’s stance towards a particular target within a text. Given the remarkable language understanding…
- Collaborative Rational Speech Act: Pragmatic Reasoning for Multi-Turn DialogIn this paper, we introduce Collaborative Rational Speech Act (CRSA), an information-theoretic (IT) extension of RSA that models multi-turn dialog by optimizing a gain function adapted from rate-disto…
- Comparing emotion feature extraction approaches for predicting depression and anxietyFor example, pride may be impacted by depression in a unique way. Gruber et al. (2011) showed that pride, a positive emotion relating to the self, is inversely correlated with depression, which is oft…
- Complex Logical Instruction GenerationInstruction following has catalyzed the recent era of Large Language Models (LLMs) and is the foundational skill underpinning more advanced capabilities such as reasoning and agentic behaviors. As tas…
- Computational structuralism: Toward a formal theory of meaning in the age of digital intelligenceThe discovery that “next-token predictor” language models can fluently produce text has important but underappreciated theoretical implications. Most notably, their success demonstrates that fully rel…
- Conversational Semantic Parsing for Dialog State TrackingWe consider a new perspective on dialog state tracking (DST), the task of estimating a user’s goal through the course of a dialog. By formulating DST as a semantic parsing task over hierarchical repre…
- DEAM: Dialogue Coherence Evaluation using AMR-based Semantic ManipulationsThose models take a contrastive learning approach, where they build binary classifiers to differentiate positive, or coherent examples from negative, or incoherent dialogues. Those classifiers are usu…
- Detecting Cognitive Distortions from Patient-Therapist InteractionsAn important part of Cognitive Behavioral Therapy (CBT) is to recognize and restructure certain negative thinking patterns that are also known as cognitive distortions. This project aims to detect the…
- Development and validation of large language model rating scales for automatically transcribed psychological therapy sessionsRating scales have shaped psychological research, but are resource-intensive and can burden participants. Large Language Models (LLMs) offer a tool to assess latent constructs in text. This study intr…
- Discourse-Level Representations can Improve Prediction of Degree of AnxietyThe primary clinical manifestation of anxiety is worry associated cognitive distortions, which are likely expressed at the discourse-level of semantics. discourse patterns of causal explanations, amo…
- Discursive Socratic Questioning: Evaluating the Faithfulness of Language Models’ Understanding of Discourse RelationsWhile large language models have significantly enhanced the effectiveness of discourse relation classifications, it remains unclear whether their comprehension is faithful and reliable. We provide DIS…
- DiscussLLM: Teaching Large Language Models When to SpeakLarge Language Models (LLMs) have demonstrated remarkable capabilities in understanding and generating human-like text, yet they largely operate as reactive agents, responding only when directly promp…
- Dissociating language and thought in large language modelsHere, we evaluate LLMs using a distinction between formal linguistic competence—knowledge of linguistic rules and patterns—and functional linguistic competence—understanding and using language in the …
- Empirical Study of Symmetrical Reasoning in Conversational ChatbotsThis work explores the capability of conversational chatbots powered by large language models (LLMs), to understand and characterize predicate symmetry, a cognitive linguistic function traditionally b…
- Enhancing Performance on Seen and Unseen Dialogue Scenarios using Retrieval-Augmented End-to-End Task-Oriented System“Task-oriented dialogue (TOD) systems play an important role in various applications, such as restaurant booking, alarm setting, and recommendations (Gao et al., 2018; Xie et al., 2022). These systems…
- Evaluating the Efficacy of Interactive Language Therapy Based on LLM for High-Functioning Autistic Adolescent Psychological Counselingsignificant emphasis was placed on the development of prompts used to guide the Large LanguageModel (LLM). This process was intricate and involved multiple stages to ensure that the prompts were effec…
- Event-Aware Sentiment Factors from LLM-Augmented Financial Tweets: A Transparent Framework for Interpretable Quant TradingIn this study, we wish to showcase the unique utility of large language models (LLMs) in financial semantic annotation and alpha signal discovery. Leveraging a corpus of company-related tweets, we use…
- Explainable Compliance Detection with Multi-Hop Natural Language Inference on Assurance Case StructureEnsuring complex systems meet regulations typically requires checking the validity of assurance cases through a claim-argument-evidence framework. Some challenges in this process include the complicat…
- Explicit Inductive Inference using Large Language ModelsHowever, recently McKenna et al. (2023a) has pointed out that LLMs are severely affected by an attestation bias when performing inference tasks. Given the question of whether premise P entails hypothe…
- Exploring LLMs Applications in Law: A Literature Review on Current Legal NLP ApproachesThe integration of Natural Language Processing (NLP) and AI into legal tasks is a natural progression, given the linguistic nature of law. This combination allows for more efficient and accurate analy…
- Exploring the Potential of ChatGPT on Sentence Level Relations: A Focus on Temporal, Causal, and Discourse RelationsThis paper aims to quantitatively evaluate the performance of ChatGPT, an interactive large language model, on inter-sentential relations such as temporal relations, causal relations, and discourse re…
- Faithful and Robust LLM-Driven Theorem Proving for NLI ExplanationsNatural language explanations play a fundamental role in Natural Language Inference (NLI) by revealing how premises logically entail hypotheses. Recent work has shown that the interaction of large lan…
- HowProjective is Projective Content? Gradience in Projectivity and At-issuenessProjective content is utterance content that a speaker may be taken to be committed to even when the expression associated with the content occurs embedded under an entailment-canceling operator (e.g.…
- Human-like Category Learning by Injecting Ecological Priors from Large Language Models into Neural Networkslarge language models can generate cognitive tasks, specifically category learning tasks, that match the statistics of real-world tasks, deriving rational agents adapted to these tasks using the frame…
- Inspecting and Editing Knowledge Representations in Language Models[[Natural Language Inference]] Neural language models (LMs) represent facts about the world described by text. Sometimes these facts derive from training data (in most LMs, a representation of the …
- Is Sarcasm Detection A Step-by-Step Reasoning Process in Large Language Models?However, human sarcasm understanding is often considered an intuitive and holistic cognitive process, in which various linguistic, contextual, and emotional cues are integrated to form a comprehensive…
- LLM-based Rewriting of Inappropriate Argumentation using Reinforcement Learning from Machine FeedbackWhile a few methods improve content, they solely transfer the style of texts to be more formal (Rao and Tetreault, 2018; Lai et al., 2021), less subjective (Pryzant et al., 2020; Liu et al., 2021a), o…
- LLMs Struggle to Reject False Presuppositions when Misinformation Stakes are HighThese implicit assumptions, known as presuppositions, refer to background knowledge or shared beliefs assumed to be part of the common ground between interlocutors (Stalnaker, 1973). Presuppositions a…
- LLMs are Frequency Pattern Learners in Natural Language InferenceWhile fine-tuning LLMs on NLI corpora improves their inferential performance, the underlying mechanisms driving this improvement remain largely opaque. In this work, we conduct a series of experiments…
- Large Language Models Can Infer Psychological Dispositions of Social Media Userswe test whether GPT-3.5 and GPT-4 can derive the Big Five personality traits from users’ Facebook status updates in a zero-shot learning scenario. Our results show an average correlation of r = .29 (r…
- Large Language Models Sensitivity to The Order of Options in Multiple-Choice Questions“Large Language Models (LLMs) have demonstrated remarkable capabilities in various NLP tasks. However, previous works have shown these models are sensitive towards prompt wording, and few-shot demonst…
- Large Linguistic Models: Investigating LLMs' metalinguistic abilitiesAbstract—The performance of large language models (LLMs) has recently improved to the point where models can perform well on many language tasks. We show here that—for the first time—the models can al…
- Large language models can segment narrative events similarly to humansHumans perceive discrete events such as "restaurant visits" and "train rides" in their continuous experience. One important prerequisite for studying human event perception is the ability of researche…
- Lexical Entrainment for Conversational Systemslexical entrainment (LE), a phenomenon in which speakers in human-human conversations tend to naturally and subconsciously align their lexical choices with those of their interlocutors, leading to mor…
- Lost in Inference: Rediscovering the Role of Natural Language Inference for Large Language Modelswe investigate if NLI tasks, that are rarely used for LLM evaluation, can still be informative for evaluating LLMs. Focusing on five different NLI benchmarks across six models of different scales, we …
- MARS: A Multi-Agent Framework Incorporating Socratic Guidance for Automated Prompt OptimizationAutomated Prompt Optimization (APO) aims to break free from the cognitive biases of manually designed prompts and explores a broader design space for prompts. However, existing APO methods suffer from…
- Man vs machine – Detecting deception in online reviewsThis study focused on three main research objectives: analyzing the methods used to identify deceptive online consumer reviews, evaluating insights provided by multi-method automated approaches based …
- Meanings are like Onions: a Layered Approach to Metaphor ProcessingAbstract Metaphorical meaning is not a flat mapping between concepts, but a complex cognitive phenomenon that integrates multiple levels of interpretation. In this paper, we propose a stratified mode…
- Mechanistic Indicators of Understanding in Large Language ModelsAbstract: Large language models (LLMs) are often portrayed as merely imitating linguistic patterns without genuine understanding. We argue that recent findings in mechanistic interpretability (MI), th…
- Minds versus Machines: Rethinking Entailment Verification with Language ModelsLeveraging a comprehensively curated entailment verification benchmark, we evaluate both human and LLM performance across various reasoning categories. Our benchmark includes datasets from three categ…
- Neutralizing Bias in LLM Reasoning using Entailment GraphsHowever, recent works show that LLMs still suffer from hallucinations in NLI due to attestation bias, where LLMs overly rely on propositional memory to build shortcuts. To solve the issue, we design a…
- On the Conversational Basis of Some PresuppositionsThe current literature on presupposition focuses almost exclusively on the projection problem: the question of how and why the presuppositions of atomic clauses are projected to complex sentences whic…
- Persuasive presuppositionsA recurrent claim, coming from different approaches to pragmatics, argumentation theory and related disciplines, is that informative presuppositions have a special persuasive force. My aim in this pap…
- Post-training for Efficient Communication via Convention FormationHumans communicate with increasing efficiency in multi-turn interactions, by adapting their language and forming ad-hoc conventions. In contrast, prior work shows that LLMs do not naturally show this …
- Presuppositions are more persuasive than assertions if addressees accommodate them: Experimental evidence for philosophical reasoningBest practice and descriptive research claim that presuppositions, such as the “too” in “,” increase the persuasiveness of arguments. Surprisingly, there is hardly any causal evidence for this claim. …
- Pretrained Language Models as Containers of the Discursive KnowledgeAbstract: Discourses can be treated as instances of knowledge. The dynamic space in which the trajectories of these discourses are described can be regarded as a model of knowledge. Such a space is ca…
- Pushing the Limits of Rule Reasoning in Transformers through Natural Language Satisfiabilitythese models to be surprisingly strong at performing deductive reasoning over formal logical theories expressed in natural language. A shortcoming of these studies, however, is that they do not take i…
- Real-time News Story IdentificationTo improve the reading experience, many news sites organize news into topical collections, called stories. In this work, we present an approach for implementing real-time story identification for a ne…
- Rethinking STS and NLI in Large Language ModelsRecent years, have seen the rise of large language models (LLMs), where practitioners use task-specific prompts; this was shown to be effective for a variety of tasks. However, when applied to semanti…
- Rhetoric, Logic, and Dialectic: Advancing Theory-based Argument Quality Assessment in Natural Language ProcessingPreceding work in natural language processing (NLP) and computational linguistics (CL) has mostly focused on practical AQ assessment, considering either the overall quality of arguments (Toledo et al.…
- SciTopic: Enhancing Topic Discovery in Scientific Literature through Advanced LLMAbstract—Topic discovery in scientific literature provides valuable insights for researchers to identify emerging trends and explore new avenues for investigation, facilitating easier scientific infor…
- Semantic Change Characterization with LLMs using RhetoricsLanguages continually evolve in response to societal events, resulting in new terms and shifts in meanings. These changes have significant implications for computer applications, including automatic t…
- Semantic Parsing for Task Oriented Dialog using Hierarchical RepresentationsPrevious work on task oriented intent and slot-filling work has been restricted to one intent per query and one slot label per token, and thus cannot model complex compositional requests. Alternative …
- Semantic Structure in Large Language Model EmbeddingsPsychological research consistently finds that human ratings of words across diverse semantic scales can be reduced to a low-dimensional form with relatively little information loss. We find that the …
- Simple Linguistic Inferences of Large Language Models (LLMs): Blind Spots and BlindsWe evaluate LLMs’ language understanding capacities on simple inference tasks that most humans find trivial. Specifically, we target (i) grammatically-specified entailments, (ii) premises with evident…
- Sources of Hallucination by Large Language Models on Inference TasksWe establish two biases originating from pretraining which predict much of their behavior, and show that these are major sources of hallucination in generative LLMs. First, memorization at the level o…
- Stance Detection on Social Media with Fine-Tuned Large Language ModelsThe implementation of prompting strategies represents a significant departure from traditional NLP model training methods. By employing these strategies, LLMs can generate predictions without the exte…
- Task-Oriented Dialogue with In-Context LearningWe describe a system for building task oriented dialogue systems combining the in context learning abilities of large language models (LLMs) with the deterministic execution of business logic. LLMs ar…
- TaskLAMA: Probing the Complex Task Understanding of Language Models“Structured Complex Task Decomposition (SCTD) is the problem of breaking down a complex real-world task (such as planning a wedding) into a directed acyclic graph over individual steps that contribute…
- The Levers of Political Persuasion with Conversational AIThere are widespread fears that conversational AI could soon exert unprecedented influence over human beliefs. Here, in three large-scale experiments (N=76,977), we deployed 19 LLMs—including some pos…
- The social component of the projection behavior of clausal complement contentsAbstract. Some accounts of presupposition projection predict that content’s consistency with the Common Ground influences whether it projects (e.g., Heim 1983; Gazdar 1979a,b). I conducted an experime…
- Transformer-based cynical expression detection in a corpus of Spanish YouTube reviewsConsumers of services and products actively engage through social networks when they are dissatisfied, exhibiting a wide range of behaviors. Encinas and Cavazos (2021). Encinas presents a classificati…
- Turiya at DialAM-2024: Inference Anchoring Theory Based LLM ParsersRepresenting discourse as argument graphs facilitates robust analysis. Although computational frameworks for constructing graphs from monologues exist, there is a lack of frameworks for parsing dialog…
- Using Computational Models to Test Syntactic LearnabilityWe study the learnability of English filler—gap dependencies and the “island” constraints on them by assessing the generalizations made by autoregressive (incremental) language models that use deep le…
- Using Natural Language for Reward Shaping in Reinforcement LearningUsing arbitrary natural language statements within reinforcement learning presents several challenges. First, a mapping between language and objects/actions must implicitly or explicitly be learned, a…
- Verbal lie detection using Large Language ModelsWhen producing deceptive narratives, liars employ verbal strategies to create false beliefs in the interacting partners and are thus involved in a specific and temporary psychological and emotional st…
- Will I Sound Like Me? Improving Persona Consistency in Dialogues through Pragmatic Self-ConsciousnessWe explore the task of improving persona consistency of dialogue agents. Recent models tackling consistency often train with additional Natural Language Inference (NLI) labels or attach trained extra …
- Word Meanings in Transformer Language ModelsWe investigate how word meanings are represented in the transformer language models. Specifically, we focus on whether transformer models employ something analogous to a lexical store - where each wor…