Discovering a system’s causal relationships and structure is a crucial yet challenging problem in scientific disciplines ranging from medicine and biology to economics. While researchers typically adopt the graphical formalism of causal Bayesian networks (CBNs) to induce a graph structure that best describes these relationships, such unsupervised score-based approaches can quickly lead to prohibitively heavy computation burdens.
A research team from DeepMind, Mila – University of Montreal and Google Brain challenges the conventional causal induction approach in their new paper Learning to Induce Causal Structure, proposing a neural network architecture that learns the graph structure of observational and/or interventional data via supervised training on synthetic graphs. The team’s proposed Causal Structure Induction via Attention (CSIvA) method effectively makes causal induction a black-box problem and generalizes favourably to new synthetic and naturalistic graphs.

The team summarizes their main contributions as:
- We tackle causal structure induction with a supervised approach (CSIvA) that maps datasets composed of both observational and interventional samples to structures.
- We introduce a variant of a transformer architecture whose attention mechanism is structured to discover relationships among variables across samples.
- We show that CSIvA generalizes to novel structures, whether or not training and test distributions match. Most importantly, training on synthetic data transfers effectively to naturalistic CBNs.

The team first trains their model on synthetic data generated using different CBNs to capture dataset and graph structure relationships, then leverages these mappings to induce the structures underlying datasets of interest. This novel variant of the conventional transformer architecture takes observational and interventional samples corresponding to the same CBN pairs as inputs, and outputs the predicted CBN graph structure.
The proposed method’s mapping between datasets and graph structures is achieved by leveraging a transformer-based attention mechanism that alternates between attending to different variables in the graph and different samples from a variable, analyzing the data and computing a distribution of candidate graphs. The decoder mechanism used to output the CBN graph structures is an autoregressive generative model that operates on the inferred structure.
The resulting CSIvA can thus be viewed as a meta-learning approach, as the model itself learns the relationships between datasets and their underlying structures. This design also enables the model to generalize well to data from naturalistic CBNs.

In their empirical study, the team compared their supervised learning paradigm to unsupervised baseline methods that included non-linear ICP (Heinze-Deml et al., 2018a), and DAG-GNN (Yu et al., 2019) on classic benchmarks such as the Sachs (Sachs et al., 2005) and Asia (Lauritzen & Spiegelhalter, 1988) datasets. In the tests, CSIvA outperformed all baselines, indicating its ability to effectively induce causal structures in realistic real-world CBNs.
Overall, the results show the proposed CSIvA generalizes well to out-of-distribution data even when trained only on synthetic data, taking a significant step forward in causal graph structure inference.
The paper Learning to Induce Causal Structure is on arXiv.
Author: Hecate He | Editor: Michael Sarazen

We know you don’t want to miss any news or research breakthroughs. Subscribe to our popular newsletter Synced Global AI Weekly to get weekly AI updates.

NY Sapphire exceeded all my expectations! The performers were fantastic, and the ambiance was top-notch. We booked a table in advance and received outstanding service throughout the night. If you’re looking for an unforgettable night out, check out sapphire new york!
Wolf Winner Casino is an online casino for players from Australia. The casino offers unique promotions, VIP programs and regular draws, making every visit unforgettable. The modern interface and mobile adaptation allow you to play anywhere and anytime, and customer support works 24/7. Get ready for new emotions and victories along with many others.
wowo thganksss
The advancements in causal structure induction presented by the DeepMind team are truly fascinating. Their approach of utilizing supervised learning on synthetic data to improve generalization for real-world applications is impressive. This kind of innovation can significantly benefit various fields, including those here in Hungary, where understanding complex relationships in data can lead to better decision-making. For anyone interested in exploring more about machine learning and its applications, I recommend checking out the roulettino app. It offers great insights and resources that can be especially useful for residents in Hungary looking to enhance their in AI and data science.
The advancements in causal structure induction, particularly the CSIvA method proposed by DeepMind and its collaborators, are quite fascinating. It’s impressive how this approach could potentially streamline complex analyses in various fields, including medicine and economics. For those of us in Greece, it’s essential to keep up with such innovations, as they could be instrumental in enhancing our local research and industry practices. Speaking of enhancements, I found an interesting opportunity for bonuses specifically for residents here: spin mama bonus. This could be a great way to leverage our local resources while benefiting from global advancements. Embracing such technologies can significantly impact our capacities and growth in different sectors.