Search This Blog

Thursday 30 November 2017

Philosophy of Science: Mathematical Modelling, Assumptions and Causal Realism

Q: How can mathematical modelling succeed when it inevitably involves false assumptions?

We happen to know from the history of science that models with false assumptions can be used to make systematically correct predictions with well-understood conditions of application. Kepler's laws of planetary motion are a canonical example [Colyvan & Ginzburg [2003]]; a more dramatic illustration is provided by the very considerable improvements in the power of metereological and climate-modelling since the 1950s. 1 The very general explanation for this very general fact is that the dynamics---the progression of state changes---of even complex systems seem to causally depend to a large degree on changes in only a small number of mathematically isolable variables which we can track with mathematical equations. This convenience given us by nature does not, however, entail scientific convenience; instead, it has the consequence that truly successful model-building in higher-level sciences will invariably require the identification of the critical causal variables---that is, the right theory or understanding. This methodological truth sits very uncomfortably with the fact that, precisely because of the large number of variables in play in the large-system modelling in the higher-level sciences, there seems to be very little hope for decisive empirical disconfirmation or falsification of models (embodying theories) which are at least vaguely appropriate for the target system. 2

My thesis in this essay is that these considerations ought to lead one to the view that assumptions do, in fact, really matter: more specifically, that the laissez-faire instrumentalist attitude towards assumptions within scientific theories (which has historically held most sway within the discipline of economics) is deeply misguided. I suggest that the history of science (including recent science, and recent developments within ecology) gives strong weight to the view that models which do not embody the causal structure of their target system either don't have predictive success at all, or eventually badly misfire, on account of being attuned only to some subset of the states of the system, rather than locking into the causal structure which determines the state changes of the system. I submit that the key source of instrumentalist error is the failure to recognise that there are different kinds of modelling assumptions, with different scientific significance, and that models which are not `realistic' in terms of their modelling of causal structure cannot succeed (at least in the long-run). I argue that ecology is an excellent case study for this thesis, by attempting to show that causal realism and the absence of highly unrealistic domain assumptions [Musgrave [1981]] is a feature of successful and commonly used models in ecology.

In his paper “Models and Fictions", the philosopher of science Peter Godfrey-Smith makes much of the fact that the scientific enterprise depends on mathematical fiction in the course of asking how it is that fictions in general can tell us about the world [Godfrey-Smith [2009]]. Left unexplored by Godfrey-Smith in the short paper is the fact that in the realm of science we call “fundamental physics", the history and current practice of the science gives us reason to think that our fundamental reality may be perfectly describable using mathematics.3 Godfrey-Smith does himself observe that one of the viable explanations for the problem

1 More dramatic since, although the amount of chemical and atmospheric `variables' explicitly encoded in climate models has increased over the last few decades and climate models are complicated enough to require state-of-the-art computing power to run, even our best climate models are still (perforce) attempts only to capture the `core' causal structure of the dynamics of

the chaotic target systems [Lucarini [2013]].

2There always exist escape clauses in the form of phrases like “exogenous shocks" and “disturbing factors". To clarify, I do not mean to convey a nihilistic, defeatist or relativist attitude towards the possibility of rational rejection of theories within the higher-level sciences (leading to genuine scientific progress); it will become clear that I believe, in fact, that models which have patently causally unreal assumptions (models which clearly embody the wrong theory) should be dismissed out of hand regardless of any apparent early empirical confirmation. Unfortunately, in the social sciences, there really does seem to be very little hope at all that such norms could be acceded to by a majority, because of the role of ideology and other social

complications.

3The theories of “fundamental physics" really do seem to be profoundly non- fictional, not embodying any assumptions known to be false in the way that higher-level sciences must. The General Theory of Relativity does not model spacetime as if it were a four-dimensional `pseudo-Riemannian' manifold; instead, we have no other grasp on what spacetime is. Similarly, Quantum Electrodynamics (the relativistic quantum field theory of electrodynamics, which Feynmann called the “jewel of

of how fictional mathematical models can tell us about the world is that many of the features of “reality" we assume away to design a good mathematical model are just ornaments for the fundamental universal mathematical patterns inherent in nature---such that a good mathematical model of a system, making the right kind of assumptions, simply cuts through the frippery. 4 Whatever the merits of Platonism, I claim that we have very good reason to believe that a very humanly useful approximation of the `causal architecture' of even highly complex or chaotic natural systems is isolable by means of mathematical equations encoding the dominant (numerically abstractable) dependencies of the system. I claim this in view of our knowledge that the necessary presence of the ceteris paribus clause and the “idealisation" in the canonical models of science as we move beyond fundamental physics into non-fundamental physics (and beyond non-fundamental physics into the rest) are not handicaps to very impressive predictions. As I noted in the introduction, it is not just the familiar tales of early Enlightenment physics that bear this out; one of the less appreciated major scientific success stories of the second half of the 20th Century is the constant leaps and bounds we have made in computer-modelling of the earth's weather and climate since the process started with the primitive computers of the 1950s. A 2015 paper in Nature entitled “The Quiet Revolution of Numerical Weather Prediction" gives us some of the key facts testifying to the magnitude this achievement: “Forecast skill in the range from 3 to 10 days ahead has been increasing by about one day per decade" whilst (as an example of some of the progress in longer-term forecasting that has been made) “tropical sea surface temperature variability following the El Nino/Southern Oscillation phenomenon can be predicted 3-4 months ahead" [Bauer et al. 2015, 47].

Unlike Godfrey-Smith (assuming I interpret him correctly), I reject the position that there is any kind of philosophically interesting analogy to be had between novels or films or other obvious fictions, and successful mathematical models of complex systems used in progressive science. I think instead that understanding the different kinds of assumptions in play in mathematical modelling within the various `higher' realms of the scientific enterprise can help us demarcate pseudoscientific modelling---that is, mathematical models that are more akin to novels or films---from mathematical models of complex systems that, whilst necessarily fictional in the sense that they only trace some of the core emergent dependencies or patterns inherent in the target system, are realistic in the sense that, by foregrounding some of the right emergent dependencies (i.e. by embodying a good theory of the target system), have locked into the “deep structural forces" of the target system that determine its evolution through time. As we'll see, I think that this can be illustrated by case studies: by drawing a contrast between---on the one hand---weather/climate modelling and successful modelling in ecology (population dynamics) (and various work on the margins of economics (complexity economics) and the new field of cliodynamics), and---on the other hand---DSGE-modelling in macroeconomics, whose predictive failures can be explained by the fact that the entire enterprise instantiates false domain assumptions (fundamentally embodying the mistaken assumption that the macroeconomy is an equilibrium system).

Before we can analyse these case studies, it is necessary to introduce an analytical tool: a taxonomy of assumptions in scientific theories due to the philosopher Alan Musgrave (contained in his brief discussion of Milton Friedman's very confused philosophy of instrumentalism from the in influential essay “Methodology of Positive Economics"). Musgrave claims that there are three kinds of assumptions worth distinguishing: negligibility assumptions, domain assumptions and heuristic assumptions. The third is less relevant for our investigations, but I shall quickly describe the first two. A negligibility assumption is an assumption that a certain phenomenon not known to be negligible in the real world (or known to be not negligible at least some of the time) is negligible. Galileo tested his theory about the acceleration of falling bodies in a vacuum in the real world with the assumption that (in his experimental cases) air-resistance was of no consequence (correct, in his experimental cases, though clearly not in general). In the complex-systems-level sciences,

physics" [Feyman 1985, 4]) has been confirmed by predictions at an astonishing level of precision---levels of precision that give us reason to think that the equations we have come up really do describe the fundamental dynamics about as well as one could imagine describing them.

4 Godfrey-Smith connects this explanation to the doctrine of Structuralism in philosophy of mathematics, the essence of which is given away by the title of Michael Resnik's 1997 book, Mathematics as the Science of Patterns.


we might consider calling the equivalent assumption-category the relative insignificance assumption, simply because in modelling population dynamics in ecology, financial crises in economics, or the rise and fall of social cohesion in history, a scientist will be forced to leave out plenty of variables which they know to make a super-negligible difference much or all of the time in the real world. This, of course, does not mean that they cannot do proper science: as I'll argue when I introduce the case studies, the scientist maintains a claim to doing science for a complex system if they can generate system state-changes that mirror the dynamics of the target system from the simplest possible model embodying their given theory, because this is a good indication that they have struck causal realism. The second kind of assumption, the domain assumption, is of a rather more general kind: it can be thought of as any kind of positive assumption that specifies the world embodied by the model (as opposed to the real world). More helpfully, an unrealistic domain assumption is an aspect of the mathematical structure of a model that is most definitely at odds with the target system it is supposed to model. I see the core instrumentalist mistake as the elision of the fact that any successful predictions made by a model which embodies too many unrealistic domain assumptions will be entirely accidental (and the addition of parameters and extra variables, bells and whistles, to bootstrap predictive power will just make the situation worse when there is a significant shift in the state of the system).

We now have the framework necessary for our case studies, beginning with ecology itself, or, more precisely, “population dynamics". In his 2002 treatise on the methodology, theory and models of mathematical ecology, Complex Population Dynamics, the ecologist and `cliodynamicist' Peter Turchin argues that ecology has become a mature, predictive science, and he attempts to demonstrate this via an analysis of a series of models. Relevant to this enterprise, the book also contains (and embodies) an expression of Turchin's modelling philosophy for complex systems science, a philosophy he has applied in his research in both ecology and on human societies in his development of the Structural-Demographic Theory of the internal cycles of empires. I think this modelling philosophy gives us the key to understanding what good complex-systems modelling should look like. Turchin's methodological algorithm goes as follows.

First, if there are two or more theories postulating fundamentally distinct (though not necessarily completely incompatible) causal mechanisms for a certain explanandum, then these theories must be given a maximally simple mathematical expression and their predictions/retrodictions compared with the data (maximally being the key word, of course, since `simple' in this context does not mean that they are easy-to-solve differential equations like the models introduced to beginning ecology students, or that they are as blatantly unrealistic as such models are). 5 Two miscellaneous examples of such explananda due to Turchin himself (from history and ecology respectively) are:

a)      The `saw-toothed' cohesion-cycles observed throughout historical societies, for which the two general competing theories are the “Malthusian theory" and the “Structural-Demographic (`elite overproduction') theory" (he has written several books explicating the latter).

b)   The fluctuations in “southern pine beetle" population size, for which the dominant explanation in the 1980s was climactic (exogenous), but which now appears (thanks to analysis of time-series data) to be explicable mostly “by second-order endogenous factors" [Turchin 2003, 164].

Ideally, this procedure should at least allow for a pretty confident falsification of very bad theories---since, if the fit for the simple version is far off the mark, then it seems a very strong bet that the theory is causally unrealistic, and if a relatively simple model generates dynamics that broadly mirror those in the target system (of course this requires a variety of quantitative measures determining the average period, period strength etc), then one has strong evidence that it is a theory worth pursuing. Second, when it comes to deciding between models of varying complexity embodying the same theory (i.e. once one has a strong theory), the procedure is again empirical. As Turchin explains in Complex Population Dynamics, “The question of whether to include population structure or not can be resolved by constructing and contrasting with data

5 It should also be noted that whilst the above principle was fairly simple to state, this conceals the sophisticated statis-tical techniques and intelligent quantitative-measure choice that are needed to execute it properly (“We can then attempt to distinguish between the competing models by doing parallel time-series analyses on the data and on model outputs, to obtain quantitative measures of their relative success at matching the patterns in the data."[165]).

two versions of the model: one that explicitly incorporates individual variability [fewer relative-insgificance assumptions], and the other that averages over this variability with judiciously chosen functional forms [more relative-insignificance assumptions]." [Turchin 2003, 65] The long and short of the philosophy is this: (a) that we should add extra parameters and include more specific variables iff these lead to a model which is a better fit to the facts, and (b) that what is of most importance fundamentally is causal realism. Crucially, this methodological algorithm should actively discourage the production of models with clearly false domain assumptions, on account of the strongly counter-instrumentalist first step.

The highly advanced science of metereology may, on the surface, seem to impugn the first principle of this methodological philosophy, inasmuch as there are no serious climate models in contemporary use that could be described as `simple' (or even which are solvable by hand at all). However, this is superficial, since one critical factor that distinguishes metereology from ecology is that, from its very inception in the 1950s, the modellers had the key `theory' in place: fluid dynamics. As the aforementioned Nature article notes, “The Navier-Stokes and mass continuity equations (including the effect of the Earth’s rotation), together with the first law of thermodynamics and the ideal gas law, represent the full set of prognostic equations upon which the change in space and time of wind, pressure, density and temperature is described in the atmosphere" [Bauer et al. 2015, 48], meaning that the first principle described by Turchin was irrelevant. As for Turchin's second principle, I believe it is in play; it just so happens that adding complexity to models and taking advantage of all the computing power available has been the way to generate better predictive power in metereology. 6

The contrastive case study is the discipline of macroeconomics. I claim that the predictive uselessness of even the most celebrated macroeconomic models, particularly (but by no means solely) when it comes to major recessions and downturns, and particularly since 2008 [Edge et al. [2010], Keen [2011]] ---a state of a airs which has led to foundational questioning by some elder statesmen of the field [Blanchard [2016], Romer [2016] ]---has evolved from a historically rooted excessive tolerance of unrealistic domain assumptions (i.e. a philosophy of instrumentalism), which is anathema to Turchin's methodology. Economics evolved as a mathematical science before the advent of complex-systems mathematics, much like ecology, but the mainstream of the discipline did not respond to the complex-systems revolution at all. Whilst there was much upheaval within economics in the 1970s and 1980s (with the rise of Friedman and Lucas and the turn to microfounded macroeconomics), this only pushed the discipline in a more unrealistic direction, in terms of the fundamental assumptions about the economy, with the paradigm of the `representative consumer' (nothing like a real person) and `representative firm' (with a cost structure unlike any real firm) optimising towards an infinite horizon [Keen [2011], Lavoie [2015]]. More fundamentally, the central paradigm of mainstream neoclassical macroeconomics is the model of the economy as a system that does return to equilibrium absent “exogenous shocks", which is not path-dependent and which functions essentially like a machine [Keen [2011], Lavoie [2015]]. This is fundamentally opposed to reality, which is why the only predictive successes that DSGE models ever did generate were the result of tailoring models to recent data, so-called “overfitting". This kind of methodological practice which may generate very impressive near-term successes but it is also likely to result in massive scientific crisis when the target system state changes in a dramatic way (as in the GFC), because the overfitted model that was in use up to this state change is now doubly useless: not only has it been shown decisively that it does not embody the causal structure of the target system, there is nothing anyone can do about it, because how are we to know what exactly went wrong in such a highly parametrised, massively complicated, baroque model?
A key difference between ecology and economics can be observed in the way and in which the role of assumptions changes as one progresses in the subject. In ecology, the elementary models taught to students---the three central `laws' (as Turchin calls them [Turchin [2001]]) of the exponential growth model,



6 Again, we should add a detail here: metereologists and climate scientists don't generate their predictions just from running one perfect model. Because of the path-dependence of chaotic systems and our extremely imperfect knowledge about all the potentially relevant atmospheric data, it is necessary to use what is called “ensemble-modelling" (comparing similar models with different initial conditions, etc) to generate reliable results.

the logistic growth model and the Lotka-Volterra model---are not used precisely because of their very bad domain assumptions: Turchin calls the most complicated of the three, the Lotka-Volterra equations, “a horribly unrealistic model for real resource-consumer systems [...] so bad that, to my knowledge, there has been no successful application of it to any actual population system, whether in the field or laboratory" [Turchin 2001, 21]. But in economics, unlike ecology, the higher-level models, although of course far more complicated than the elementary models (their mastery requiring years of graduate training), are just as profoundly unrealistic as the elementary ones.

In summary, and in answer to the question, mathematical models in ecology (population dynamics) serve their purpose when they are built in such a way that their relative-insignificance assumptions are orthogonal to the key causal structure of the target system, enabling the models to generate solid predictions even as the state of the target system changes (because they are calibrated to the drivers of the system's state-changes). This fact helps us see how and in what sense assumptions matter in complex-systems science: the chief task has to be to find the right theory, and this task will inevitably result in the mulching of mathematical models that instantiate completely unrealistic domain assumptions. The scourge of `overfitting' can be avoided by focussing on arriving at a model that can generate all the states of the target system, not just generate some neat predictions in a period of equilibrium, and this is equivalent to a causally realistic model.


























References

Bauer, Peter, Thorpe, Alan, & Brunet, Gilbert. 2015. “The quiet revolution of numerical weather prediction".

Nature, 525(7567), 47{55.

Blanchard, Olivier. 2016. “Do DSGE Models Have a Future?" Peterson Institute of International Economics, 11{16.

Colyvan, Mark, & Ginzburg, Lev R. 2003. “Laws of nature and laws of ecology". Oikos, 101(3), 649{653.

Edge, Rochelle M., Gurkaynak, Refet S., Reis, Ricardo, & A., Sims Christopher. 2010. “How Useful Are Estimated DSGE Model Forecasts for Central Bankers?" [with Comments and Discussion]. Brookings Papers on Economic Activity, 209{259.

Feyman, Richard P. 1985. QED: The Strange Theory of Light and Matter. Princeton University Press.

Godfrey-Smith, Peter. 2009. “Models and Fictions in Science". Philosophical Studies, 143(1), 101{116.

Keen, Steve. 2011. Debunking economics: the naked emperor dethroned? Rev. and expanded edn. London:

Zed Books Ltd.

Lavoie, Marc. 2015. Post-Keynesian Economics: New Foundations. Edward Elgar.

Lucarini, Valerio. 2013. “Modelling Complexity: The Case of Climate Science". Pages 229{253 of: Ghde, Ulrich, Hartmann, Stephen, & Wolf, Jrn H. (eds), Models, Simulations, and the Reduction of Complexity. De Gruyter.

Musgrave, Alan. 1981. “Unreal Assumptions in Economic Theory: The F-Twist Untwisted". Kyklos, 34(3), 377{387.

Romer, C. D. 2016. The Trouble with Macroeconomics. Commons Memorial Lecture of the Omicron Delta Epsilon Society.

Turchin, Peter. 2001. “Does population ecology have general laws?" Oikos, 94(1), 17{26.

Turchin, Peter. 2003. Complex Population Dynamics: A Theoretical/Empirical Synthesis. Princeton Uni-versity Press.

No comments:

Post a Comment