Papers
arxiv:2602.18662

Large Causal Models for Temporal Causal Discovery

Published on Feb 20
· Submitted by
Nikolas Kougioulis
on Feb 24
Authors:
,
,
,
,
,
,

Abstract

Large causal models combine synthetic and real time-series data to enable scalable temporal causal discovery with improved generalization and fast inference.

AI-generated summary

Causal discovery for both cross-sectional and temporal data has traditionally followed a dataset-specific paradigm, where a new model is fitted for each individual dataset. Such an approach limits the potential of multi-dataset pretraining. The concept of large causal models (LCMs) envisions a class of pre-trained neural architectures specifically designed for temporal causal discovery. Prior approaches are constrained to small variable counts, degrade with larger inputs, and rely heavily on synthetic data, limiting generalization. We propose a principled framework for LCMs, combining diverse synthetic generators with realistic time-series datasets, allowing learning at scale. Extensive experiments on synthetic, semi-synthetic and realistic benchmarks show that LCMs scale effectively to higher variable counts and deeper architectures while maintaining strong performance. Trained models achieve competitive or superior accuracy compared to classical and neural baselines, particularly in out-of-distribution settings, while enabling fast, single-pass inference. Results demonstrate LCMs as a promising foundation-model paradigm for temporal causal discovery. Experiments and model weights are available at https://github.com/kougioulis/LCM-paper/.

Community

Paper submitter

This paper introduces Large Causal Models, foundation models for Causal Discovery (CD) on time-series data following a supervised learning scheme. It expands prior proof-of-concept approaches by combining hundends of thousands of synthetic and real training data to enable scalable performance (deeper models and higher input dimensions) with improved generalization and faster inference compared to classical CD approaches.

Sign up or log in to comment

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2602.18662 in a model README.md to link it from this page.

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2602.18662 in a dataset README.md to link it from this page.

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2602.18662 in a Space README.md to link it from this page.

Collections including this paper 0

No Collection including this paper

Add this paper to a collection to link it from this page.