Hostname: page-component-78c5997874-t5tsf Total loading time: 0 Render date: 2024-11-16T15:18:39.023Z Has data issue: false hasContentIssue false

Beyond integrative experiment design: Systematic experimentation guided by causal discovery AI

Published online by Cambridge University Press:  05 February 2024

Erich Kummerfeld*
Affiliation:
Institute for Health Informatics, University of Minnesota, Minneapolis, MN, USA erichk@umn.edu; https://erichkummerfeld.com/
Bryan Andrews
Affiliation:
Department of Psychiatry and Behavioral Sciences, University of Minnesota, Minneapolis, MN, USA andr1017@umn.edu
*
*Corresponding author.

Abstract

Integrative experiment design is a needed improvement over ad hoc experiments, but the specific proposed method has limitations. We urge a further break with tradition through the use of an enormous untapped resource: Decades of causal discovery artificial intelligence (AI) literature on optimizing the design of systematic experimentation.

Type
Open Peer Commentary
Copyright
Copyright © The Author(s), 2024. Published by Cambridge University Press

Almaatouq et al. propose a break from tradition to accelerate scientific progress, and we applaud them for it. However, we urge an even further shift to incorporate theory and methods from causal discovery, a subfield of machine learning with decades of research on artificial intelligence (AI)-guided causal learning and experiment design. Causal discovery has not been well leveraged in the experimental sciences perhaps because it also breaks from tradition – statistical tradition.

Causal discovery contains a growing collection of methods for learning multivariate structural causal models (Pearl, Reference Pearl2000; Spirtes et al., Reference Spirtes, Glymour, Scheines, Heckerman, Meek, Cooper and Richardson2000). Design spaces can be represented as a substructure of a larger structural causal model (illustrated in Fig. 1), making causal discovery closely aligned with research cartography. It is not surprising then that some of the challenges faced by integrative experiment design might be overcome with causal discovery. We focus on three such challenges: Practical application and scalability, confined inferential scope, and unknown causal factors.

Figure 1. (a) Hypothetical design space with three binary dimensions: Veteran status, rural status, and sex. Different experiment outcomes are colored red, green, blue, and yellow. Note that in this hypothetical example, rural status makes no difference to the outcome of the experiment, while each of the four combinations of veteran status and sex produce different outcomes. (b) A causal model that would correspond to the example design space. The structure of the causal model is shown on the left, and the two causal dependency tables are shown on the right: One for veteran status, which depends on sex and rural, and the other for outcome. The table for outcome is shown with rural included, to make the comparison with the design space clear, but in a normal causal model rural would not be included in this table as no arrow points directly from rural to outcome in the model structure.

Regarding the practical application of design spaces, causal discovery can learn entire causal models from nonexperimental data alone, but the direction of causal relationships can be difficult to identify (Hoyer, Janzing, Mooij, Peters, & Schölkopf, Reference Hoyer, Janzing, Mooij, Peters and Schölkopf2008; Peters, Janzing, & Schölkopf, Reference Peters, Janzing and Schölkopf2011; Peters et al., Reference Peters, Mooij, Janzing and Schölkopf2014; Shimizu, Hoyer, Hyvärinen, & Kerminen, Reference Shimizu, Hoyer, Hyvärinen and Kerminen2006; Shimizu et al., Reference Shimizu, Inazumi, Sogawa, Hyvarinen, Kawahara, Washio and Bollen2011; Spirtes et al., Reference Spirtes, Glymour, Scheines, Heckerman, Meek, Cooper and Richardson2000). Causal discovery can be applied to experimental data to resolve this limitation. Multiple methods are capable of combining datasets with: Both experimental and observational samples, samples with nonidentical variables, and samples from different contexts and populations (Bareinboim & Pearl, Reference Bareinboim and Pearl2016; Huang et al., Reference Huang, Zhang, Zhang, Ramsey, Sanchez-Romero, Glymour and Schölkopf2020; Mooij, Magliacane, & Claassen, Reference Mooij, Magliacane and Claassen2020; Peters, Bühlmann, & Meinshausen, Reference Peters, Bühlmann and Meinshausen2016). Incorporating these methods could enable increased flexibility when dealing with practical study design challenges.

Scalability is another practical issue: The size of these spaces makes complete search infeasible. Causal discovery methods can scale to large numbers of variables, however. Even a million variables is possible (Ramsey, Glymour, Sanchez-Romero, & Glymour, Reference Ramsey, Glymour, Sanchez-Romero and Glymour2017), but this applies to sparse models. In sparse models, each variable is directly related to only a small number of other variables. When variables have large numbers of interacting causes, causal discovery also suffers scalability problems (Spirtes et al., Reference Spirtes, Glymour, Scheines, Heckerman, Meek, Cooper and Richardson2000). However, such situations may not be common in reality. Like how linear and Gaussian modeling are surprisingly effective, sparse models often capture the important elements of a causal system. As alternatives, the active learning methods Almaatouq et al. point to could be used, and active learning causal discovery methods also exist (Ghassami, Salehkaleybar, Kiyavash, & Bareinboim, Reference Ghassami, Salehkaleybar, Kiyavash, Bareinboim, Dy and Krause2018; Hyttinen, Eberhardt, & Hoyer, Reference Hyttinen, Eberhardt and Hoyer2013a; Lindgren, Kocaoglu, Dimakis, & Vishwanath, Reference Lindgren, Kocaoglu, Dimakis and Vishwanath2018).

Confined inferential scope limits the kinds of information that can be learned. For example, let X, Y, and Z be variables. Some study designs allow researchers to learn that X causes Z and Y causes Z, but prevent researchers from learning whether X mediates the effect of Y on Z. In a pair of papers, Mayo-Wilson (Reference Mayo-Wilson2011, Reference Mayo-Wilson2014) proved: (1) certain causal facts cannot be learned from a system of experiments that each only investigate a single exposure–outcome pair, (2) the proportion of unlearnable facts approaches 100% as the complexity of the system increases, and (3) overcoming this requires that each experiment measures more variables than an exposure–outcome pair. By focusing on a single experiment under different conditions, Almaatouq et al. are at risk of being confined to a space of causal facts not much greater than the ad hoc experimentation they are trying to break away from.

Researchers ought to simultaneously measure as many relevant variables as possible. This happens naturally when planning to use causal discovery methods. Most causal discovery methods treat all variables equally, with no labeled outcome variable. It is normal in causal discovery to cast a wide net and use measurements from a larger number of variables, and then simultaneously model them with an algorithm. There is a growing body of papers applying this approach, including some in the social and behavioral sciences (Bronstein, Everaert, Kummerfeld, Haynos, & Vinogradov, Reference Bronstein, Everaert, Kummerfeld, Haynos and Vinogradov2022a; Bronstein, Kummerfeld, MacDonald, & Vinogradov, Reference Bronstein, Kummerfeld, MacDonald and Vinogradov2022b; Shen, Ma, Vemuri, & Simon, Reference Shen, Ma, Vemuri and Simon2020; Stevenson et al., Reference Stevenson, Kummerfeld, Merrill, Blevins, Abrantes, Kushner and Lim2022).

Unknown causal factors are ubiquitous in science and, unbeknownst to the researcher, can modify the context under which the data were collected. This commonly manifests as latent confounding. In the integrative experimental design paradigm it would occur as a failure to fully specify the design space. Research cartography could possibly solve this, but it is unclear how.

In contrast, causal discovery offers multiple solutions to unknown causal factors. Many causal discovery algorithms are only correct assuming “causal sufficiency”: That there are no unknown causal factors causing two or more measured variables. However there are also many papers developing theory and methods without assuming causal sufficiency (Chen et al., Reference Chen, Zhang, Cai, Huang, Ramsey, Hao and Glymour2021; Hyttinen, Hoyer, Eberhardt, & Jarvisalo, Reference Hyttinen, Hoyer, Eberhardt and Jarvisalo2013b; Ogarrio, Spirtes, & Ramsey, Reference Ogarrio, Spirtes and Ramsey2016; Spirtes et al., Reference Spirtes, Glymour, Scheines, Heckerman, Meek, Cooper and Richardson2000; Zhang, Reference Zhang2008). In many cases the presence or absence of unknown causal factors can be identified from measured data, and there are even causal discovery methods designed to learn the causal relationships among them (Huang, Low, Xie, Glymour, & Zhang, Reference Huang, Low, Xie, Glymour and Zhang2022; Kummerfeld & Ramsey, Reference Kummerfeld and Ramsey2016; Xie et al., Reference Xie, Huang, Chen, He, Geng, Zhang, Chaudhuri, Jegelka, Song, Szepesvari, Niu and Sabato2022).

Unfortunately, causal discovery has had limited application in the experimental sciences. We hope this commentary helps to raise awareness of these resources. Almaatouq et al. make it clear that there is a demand for these research products in the social and behavioral sciences. There is a serious barrier to the adoption and use of causal discovery: Much of it is buried and scattered among journals covering relatively unapplied topics such as theoretical machine learning and philosophy of science. We expect that in the future causal discovery will gain presence in journals on experimental methods and design or topics such as behavioral and brain sciences.

Financial support

E. K. was supported by funding through Grant No. NCRR 1UL1TR002494-01 and B. A. was supported by T32 DA037183. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

Competing interest

None.

References

Bareinboim, E., & Pearl, J. (2016). Causal inference and the data-fusion problem. Proceedings of the National Academy of Sciences of the United States of America, 113(27), 73457352.CrossRefGoogle ScholarPubMed
Bronstein, M. V., Everaert, J., Kummerfeld, E., Haynos, A. F., & Vinogradov, S. (2022a). Biased and inflexible interpretations of ambiguous social situations: Associations with eating disorder symptoms and socioemotional functioning. International Journal of Eating Disorders, 55(4), 518529. https://doi.org/10.1002/eat.23688CrossRefGoogle ScholarPubMed
Bronstein, M. V., Kummerfeld, E., MacDonald, A III, & Vinogradov, S. (2022b). Willingness to vaccinate against SARS-CoV-2: The role of reasoning biases and conspiracist ideation. Vaccine, 40(2), 213222.CrossRefGoogle ScholarPubMed
Chen, W., Zhang, K., Cai, R., Huang, B., Ramsey, J., Hao, Z., & Glymour, C. (2021). FRITL: A hybrid method for causal discovery in the presence of latent confounders. arXiv [cs.LG]. arXiv. http://arxiv.org/abs/2103.14238Google Scholar
Ghassami, A., Salehkaleybar, S., Kiyavash, N., & Bareinboim, E. (2018). Budgeted experiment design for causal structure learning. In Dy, J. & Krause, A. (Eds.), Proceedings of the 35th international conference on machine learning (Vol. 80, pp. 17241733). PMLR.Google Scholar
Hoyer, P., Janzing, D., Mooij, J. M., Peters, J., & Schölkopf, B. (2008). Nonlinear causal discovery with additive noise models. Advances in Neural Information Processing Systems, 21, 689696. https://proceedings.neurips.cc/paper/2008/hash/f7664060cc52bc6f3d620bcedc94a4b6-Abstract.htmlGoogle Scholar
Huang, B., Low, C. J. H., Xie, F., Glymour, C., & Zhang, K. (2022). Latent hierarchical causal structure discovery with rank constraints. Advances in Neural Information Processing Systems, 35, 55495561.Google Scholar
Huang, B., Zhang, K., Zhang, J., Ramsey, J., Sanchez-Romero, R., Glymour, C., & Schölkopf, B. (2020). Causal discovery from heterogeneous/nonstationary data. Journal of Machine Learning Research: JMLR, 21(1), 34823534.Google Scholar
Hyttinen, A., Eberhardt, F., & Hoyer, P. O. (2013a). Experiment selection for causal discovery. Journal of Machine Learning Research: JMLR, 14, 30413071.Google Scholar
Hyttinen, A., Hoyer, P. O., Eberhardt, F., & Jarvisalo, M. (2013b). Discovering cyclic causal models with latent variables: A general SAT-based procedure. In Proceedings of the Twenty-Ninth Conference on Uncertainty in Artificial Intelligence (pp. 301310).Google Scholar
Kummerfeld, E., & Ramsey, J. (2016). Causal Clustering for 1-Factor Measurement Models. In KDD '16: Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (pp. 16551664). https://doi.org/10.1145/2939672.2939838CrossRefGoogle ScholarPubMed
Lindgren, E., Kocaoglu, M., Dimakis, A. G., & Vishwanath, S. (2018). Experimental design for cost-aware learning of causal graphs. Advances in Neural Information Processing Systems, 31, 52845294. https://proceedings.neurips.cc/paper/2018/hash/ba3e9b6a519cfddc560b5d53210df1bd-Abstract.htmlGoogle Scholar
Mayo-Wilson, C. (2011). The problem of piecemeal induction. Philosophy of Science, 78(5), 864874.CrossRefGoogle Scholar
Mayo-Wilson, C. (2014). The limits of piecemeal causal inference. The British Journal for the Philosophy of Science, 65(2), 213249.CrossRefGoogle Scholar
Mooij, J. M., Magliacane, S., & Claassen, T. (2020). Joint causal inference from multiple contexts. Journal of Machine Learning Research: JMLR, 21(1), 39194026.Google Scholar
Ogarrio, J. M., Spirtes, P., & Ramsey, J. (2016). A hybrid causal search algorithm for latent variable models. JMLR Workshop and Conference Proceedings, 52, 368379.Google ScholarPubMed
Pearl, J. (2000). Causality: Models, reasoning and inference (Vol. 29). Springer.Google Scholar
Peters, J., Bühlmann, P., & Meinshausen, N. (2016). Causal inference by using invariant prediction: Identification and confidence intervals. Journal of the Royal Statistical Society. Series B, Statistical Methodology, 78(5), 9471012.CrossRefGoogle Scholar
Peters, J., Janzing, D., & Schölkopf, B. (2011). Causal inference on discrete data using additive noise models. IEEE Transactions on Pattern Analysis and Machine Intelligence, 33(12), 24362450.CrossRefGoogle ScholarPubMed
Peters, J., Mooij, J., Janzing, D., & Schölkopf, B. (2014). Causal discovery with continuous additive noise models. Journal of Machine Learning Research, 15, 20092053. https://www.jmlr.org/papers/volume15/peters14a/peters14a.pdfGoogle Scholar
Ramsey, J., Glymour, M., Sanchez-Romero, R., & Glymour, C. (2017). A million variables and more: The fast greedy equivalence search algorithm for learning high-dimensional graphical causal models, with an application to functional magnetic resonance images. International Journal of Data Science and Analytics, 3(2), 121129.CrossRefGoogle ScholarPubMed
Shen, X., Ma, S., Vemuri, P., & Simon, G., & Alzheimer's Disease Neuroimaging Initiative. (2020). Challenges and opportunities with causal discovery algorithms: Application to Alzheimer's pathophysiology. Scientific Reports, 10(1), 2975.CrossRefGoogle ScholarPubMed
Shimizu, S., Hoyer, P. O., Hyvärinen, A., & Kerminen, A. (2006). A linear non-Gaussian acyclic model for causal discovery. Journal of Machine Learning Research: JMLR, 7(Oct), 20032030.Google Scholar
Shimizu, S., Inazumi, T., Sogawa, Y., Hyvarinen, A., Kawahara, Y., Washio, T., … Bollen, K. (2011). DirectLiNGAM: A direct method for learning a linear non-Gaussian structural equation model. The Journal of Machine Learning Research, 12, 12251248. https://www.jmlr.org/papers/volume12/shimizu11a/shimizu11a.pdfGoogle Scholar
Spirtes, P., Glymour, C. N., Scheines, R., Heckerman, D., Meek, C., Cooper, G., & Richardson, T. (2000). Causation, prediction, and search. MIT Press.Google Scholar
Stevenson, B. L., Kummerfeld, E., Merrill, J. E., Blevins, C., Abrantes, A. M., Kushner, M. G., & Lim, K. O. (2022). Quantifying heterogeneity in mood–alcohol relationships with idiographic causal models. Alcoholism, Clinical and Experimental Research, 46(10), 19131924.CrossRefGoogle ScholarPubMed
Xie, F., Huang, B., Chen, Z., He, Y., Geng, Z., & Zhang, K. (2022). Identification of linear non-Gaussian latent hierarchical structure. In Chaudhuri, K., Jegelka, S., Song, L., Szepesvari, C., Niu, G., & Sabato, S. (Eds.), Proceedings of the 39th international conference on machine learning (Vol. 162, pp. 2437024387). PMLR.Google Scholar
Zhang, J. (2008). On the completeness of orientation rules for causal discovery in the presence of latent confounders and selection bias. Artificial Intelligence, 172(16), 18731896.CrossRefGoogle Scholar
Figure 0

Figure 1. (a) Hypothetical design space with three binary dimensions: Veteran status, rural status, and sex. Different experiment outcomes are colored red, green, blue, and yellow. Note that in this hypothetical example, rural status makes no difference to the outcome of the experiment, while each of the four combinations of veteran status and sex produce different outcomes. (b) A causal model that would correspond to the example design space. The structure of the causal model is shown on the left, and the two causal dependency tables are shown on the right: One for veteran status, which depends on sex and rural, and the other for outcome. The table for outcome is shown with rural included, to make the comparison with the design space clear, but in a normal causal model rural would not be included in this table as no arrow points directly from rural to outcome in the model structure.