There are increasing contrasts between most scientific and scholarly enterprises and much of the data science, machine learning, and artificial intelligence protocols. The former is often focused on discovering the mechanisms that produce various outcomes that are observable. What causal mechanisms underlie the process that produce metastatic cancers? What social forces produce widespread lack of trust in social institutions? How does the loss of biodiversity affect social inequalities?
These are often “why” or “how” questions versus “what” questions. Many of the tasks of machine learning and artificial intelligence systems are interest is prediction of an outcome. GPT-4 predicts the next word in a sequence of words based on learning from massive quantities of text. Simpler algorithms predict the likelihood of recidivism of an incarcerated person. Others predict a product we might be interest in buying given various expressions of interest in purchases.
Early uses of new models made popular in data science took as their starting point an existing set of data. The perceived job of the data scientist was to use all the data possible to predict a given outcome of interest. But such a task of prediction does not require the analyst to seek understanding of the causal mechanisms of an outcome of interest. The goal of accurate prediction can often be attained merely by sophisticated analytic techniques. That is, this can happen many times successfully, until it doesn’t happen. Changed circumstances, new populations that experience influences the training data did not, can lead to model breakdown.
It is interesting that many of the social sciences that use data to discover causal mechanisms, often do not attempt a prediction step. Instead, they are interested in the relative importance of alternative causal factors. The debate in fields is not about the accuracy of prediction but about whether the scholars have missed important factors that may influence the outcome or whether they have misunderstood the manner of influence on the outcome of some factor.
Of course, some academic efforts move from the identification of causal factors and estimation of their relative impact to prediction. This is common, for example, in simulation of policy options for interventions in some social outcome. How much would the child tax credit change the number of children in poverty? How much would increasing the number of police reduce the number of burglaries in a city?
Sometimes models similar to those that have been built to understand causal mechanisms are used in such predictions. But such models tend to be much more heavily scrutinized with regard to their applicability, because their designs are open and transparent.
The system of models that underlying many of the prediction tools arising in artificial intelligence are much more complex, because the sole goal of prediction accuracy doesn’t require the scrutiny of causal assertions. The application of models with billions of parameters yields the result that even the inventors of the system cannot explain its performance on a particular outcome. Transparency is lost through complexity.
Prediction models formed with disregard for the causal mechanisms of an outcome are doomed to failure at some point. The real challenge is whether humanity is capable of building general purpose artificial intelligence systems that are aware of such causal mechanisms – in the extreme, the identification of causal mechanisms for all possible outcomes.
Address

ICC 650
Box 571014
37th & O St, N.W.
Washington, D.C. 20057
Contact
Phone: (202) 687.6400
Email: provost@georgetown.edu
Office of the ProvostBox 571014 650 ICC37th and O Streets, N.W., Washington D.C. 20057Phone: (202) 687.6400Fax: (202) 687.5103provost@georgetown.edu
Connect with us via:
I agree with your perspective on the contrasting aims of traditional scientific inquiry and AI/machine learning models. Indeed, the lack of interpretability and transparency in complex machine learning models can become a significant issue, especially when these models are applied in critical real-world contexts.
Your point about the eventual failure of prediction models that disregard causal mechanisms resonates with me. As we advance further into the era of AI and big data, it will be crucial to develop hybrid models that incorporate our understanding of causal mechanisms while leveraging the power of machine learning for prediction. This could possibly lead us to more robust, reliable, and transparent AI systems. It’s an exciting yet challenging frontier for future research and development.
This post elegantly brings to light the crucial distinction between explanation and prediction in both traditional scientific studies and the evolving field of AI and data science. It underlines the essential understanding that accurate prediction does not necessarily equate to a deep understanding of the causal mechanisms at play.
The focus on causal inference in many scientific disciplines provides a solid framework for identifying and understanding the interplay of various factors. However, as the post points out, many of these studies do not necessarily emphasize prediction. On the other hand, AI and machine learning models excel at prediction based on patterns in vast amounts of data, but they often fall short when it comes to providing the ‘why’ behind these predictions.
I agree with the point that models solely fixated on prediction, without considering causal mechanisms, are ultimately bound to fail in some situations. This is especially the case when circumstances change or when new, unencountered influences come into play.
Digiwings Agency
When you come to a fork in the road. Take it ! Yogi. Also correlation is not causation .
Uwe, “two roads diverged in a yellow wood . . .” — Robert Frost
There is nothing (except for will, effort, time, expertise, resources, etc.) to prevent AT from making a stab at (just as other research-based decision-technologies make a stab at) conducting comprehensive policy analysis that takes causal mechanisms into consideration (including talking normative/value judgments of relative stakeholders into consideration):
comprehensive policy analysis
descriptive analysis
predictive analysis
normative analysis
prescriptive analysis
Hmm I guess correlations don’t mean causation? Does that sorta sum it up? Good discussion. All that can be counted doesn’t always count ! And sometimes you can’t count what needs to be counted ! My first year Med stats course summary !
“It’s hard to make predictions, especially about the future.”
– Yogi Berra
Yogisms. Yup “ when you come to the fork in the road …. Take it! “