[2603.04276] Causality Elicitation from Large Language Models

[2603.04276] Causality Elicitation from Large Language Models

arXiv - AI 3 min read

About this article

Abstract page for arXiv paper 2603.04276: Causality Elicitation from Large Language Models

Computer Science > Machine Learning arXiv:2603.04276 (cs) [Submitted on 4 Mar 2026] Title:Causality Elicitation from Large Language Models Authors:Takashi Kameyama, Masahiro Kato, Yasuko Hio, Yasushi Takano, Naoto Minakawa View a PDF of the paper titled Causality Elicitation from Large Language Models, by Takashi Kameyama and Masahiro Kato and Yasuko Hio and Yasushi Takano and Naoto Minakawa View PDF HTML (experimental) Abstract:Large language models (LLMs) are trained on enormous amounts of data and encode knowledge in their parameters. We propose a pipeline to elicit causal relationships from LLMs. Specifically, (i) we sample many documents from LLMs on a given topic, (ii) we extract an event list from from each document, (iii) we group events that appear across documents into canonical events, (iv) we construct a binary indicator vector for each document over canonical events, and (v) we estimate candidate causal graphs using causal discovery methods. Our approach does not guarantee real-world causality. Rather, it provides a framework for presenting the set of causal hypotheses that LLMs can plausibly assume, as an inspectable set of variables and candidate graphs. Subjects: Machine Learning (cs.LG); Artificial Intelligence (cs.AI); Computation and Language (cs.CL); Econometrics (econ.EM) Cite as: arXiv:2603.04276 [cs.LG]   (or arXiv:2603.04276v1 [cs.LG] for this version)   https://doi.org/10.48550/arXiv.2603.04276 Focus to learn more arXiv-issued DOI via DataCite (pen...

Originally published on March 05, 2026. Curated by AI News.

Related Articles

Claude Mythos and Project Glasswing: why an AI superhacker has the tech world on alert
Llms

Claude Mythos and Project Glasswing: why an AI superhacker has the tech world on alert

A new AI model could automate the process of searching for cybersecurity bugs and flaws – for better or worse.

AI Tools & Products · 5 min ·
Gemini could take a 'proactive' approach with leaked 'Your Day' feature
Llms

Gemini could take a 'proactive' approach with leaked 'Your Day' feature

This feature could leverage your apps in a way that might feel familiar.

AI Tools & Products · 5 min ·
I ditched my paper planner for Gemini Live — and it solved the one professional problem I couldn't fix
Llms

I ditched my paper planner for Gemini Live — and it solved the one professional problem I couldn't fix

Can Gemini Live replace a physical planner? Tom's Guide AI Editor Amanda Caswell ditched her notebook for Google’s voice AI. Here’s how i...

AI Tools & Products · 8 min ·
Anthropic is facing a wave of user backlash over reports of performance issues with its Claude AI chatbot
Llms

Anthropic is facing a wave of user backlash over reports of performance issues with its Claude AI chatbot

"Claude has regressed to the point [that] it cannot be trusted to perform complex engineering," one developer wrote.

AI Tools & Products · 12 min ·
More in Llms: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime