[2602.18662] Large Causal Models for Temporal Causal Discovery

[2602.18662] Large Causal Models for Temporal Causal Discovery

arXiv - Machine Learning 3 min read Article

Summary

This paper presents Large Causal Models (LCMs) designed for temporal causal discovery, addressing limitations of traditional dataset-specific approaches by enabling multi-dataset pretraining.

Why It Matters

The development of LCMs represents a significant advancement in causal discovery methodologies, particularly for temporal data. By overcoming the constraints of previous models, LCMs can enhance the accuracy and efficiency of causal inference in diverse applications, making them valuable for researchers and practitioners in machine learning and data science.

Key Takeaways

  • LCMs provide a scalable framework for temporal causal discovery.
  • They outperform traditional models in accuracy and efficiency, especially in out-of-distribution scenarios.
  • The integration of synthetic data with realistic datasets enhances model generalization.
  • LCMs enable fast, single-pass inference, improving practical applications.
  • Extensive experiments validate LCMs' effectiveness across various benchmarks.

Computer Science > Machine Learning arXiv:2602.18662 (cs) [Submitted on 20 Feb 2026] Title:Large Causal Models for Temporal Causal Discovery Authors:Nikolaos Kougioulis, Nikolaos Gkorgkolis, MingXue Wang, Bora Caglayan, Dario Simionato, Andrea Tonon, Ioannis Tsamardinos View a PDF of the paper titled Large Causal Models for Temporal Causal Discovery, by Nikolaos Kougioulis and 6 other authors View PDF HTML (experimental) Abstract:Causal discovery for both cross-sectional and temporal data has traditionally followed a dataset-specific paradigm, where a new model is fitted for each individual dataset. Such an approach limits the potential of multi-dataset pretraining. The concept of large causal models (LCMs) envisions a class of pre-trained neural architectures specifically designed for temporal causal discovery. Prior approaches are constrained to small variable counts, degrade with larger inputs, and rely heavily on synthetic data, limiting generalization. We propose a principled framework for LCMs, combining diverse synthetic generators with realistic time-series datasets, allowing learning at scale. Extensive experiments on synthetic, semi-synthetic and realistic benchmarks show that LCMs scale effectively to higher variable counts and deeper architectures while maintaining strong performance. Trained models achieve competitive or superior accuracy compared to classical and neural baselines, particularly in out-of-distribution settings, while enabling fast, single-pass ...

Related Articles

[2604.01989] Attention at Rest Stays at Rest: Breaking Visual Inertia for Cognitive Hallucination Mitigation
Llms

[2604.01989] Attention at Rest Stays at Rest: Breaking Visual Inertia for Cognitive Hallucination Mitigation

Abstract page for arXiv paper 2604.01989: Attention at Rest Stays at Rest: Breaking Visual Inertia for Cognitive Hallucination Mitigation

arXiv - AI · 4 min ·
[2604.01447] Better Rigs, Not Bigger Networks: A Body Model Ablation for Gaussian Avatars
Machine Learning

[2604.01447] Better Rigs, Not Bigger Networks: A Body Model Ablation for Gaussian Avatars

Abstract page for arXiv paper 2604.01447: Better Rigs, Not Bigger Networks: A Body Model Ablation for Gaussian Avatars

arXiv - AI · 3 min ·
[2603.24326] Boosting Document Parsing Efficiency and Performance with Coarse-to-Fine Visual Processing
Llms

[2603.24326] Boosting Document Parsing Efficiency and Performance with Coarse-to-Fine Visual Processing

Abstract page for arXiv paper 2603.24326: Boosting Document Parsing Efficiency and Performance with Coarse-to-Fine Visual Processing

arXiv - AI · 4 min ·
[2603.18545] CoDA: Exploring Chain-of-Distribution Attacks and Post-Hoc Token-Space Repair for Medical Vision-Language Models
Llms

[2603.18545] CoDA: Exploring Chain-of-Distribution Attacks and Post-Hoc Token-Space Repair for Medical Vision-Language Models

Abstract page for arXiv paper 2603.18545: CoDA: Exploring Chain-of-Distribution Attacks and Post-Hoc Token-Space Repair for Medical Visio...

arXiv - AI · 4 min ·
More in Machine Learning: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime