Models in Science
Models are of central importance in many scientific contexts. The centrality of models such as the billiard ball model of a gas, the Bohr model of the atom, the MIT bag model of the nucleon, the Gaussian-chain model of a polymer, the Lorenz model of the atmosphere, the Lotka-Volterra model of predator-prey interaction, the double helix model of DNA, agent-based and evolutionary models in the social sciences, and general equilibrium models of markets in their respective domains are cases in point. Scientists spend a great deal of time building, testing, comparing and revising models, and much journal space is dedicated to introducing, applying and interpreting these valuable tools. In short, models are one of the principal instruments of modern science.
Philosophers are acknowledging the importance of models with increasing attention and are probing the assorted roles that models play in scientific practice. The result has been an incredible proliferation of model-types in the philosophical literature. Probing models, phenomenological models, computational models, developmental models, explanatory models, impoverished models, testing models, idealized models, theoretical models, scale models, heuristic models, caricature models, didactic models, fantasy models, toy models, imaginary models, mathematical models, substitute models, iconic models, formal models, analogue models and instrumental models are but some of the notions that are used to categorize models. While at first glance this abundance is overwhelming, it can quickly be brought under control by recognizing that these notions pertain to different problems that arise in connection with models. For example, models raise questions in semantics (what is the representational function that models perform?), ontology (what kind of things are models?), epistemology (how do we learn with models?), and, of course, in general philosophy of science (how do models relate to theory?; what are the implications of a model based approach to science for the debates over scientific realism, reductionism, explanation and laws of nature?).
- 1. Semantics: Models and Representation
- 2. Ontology: What Are Models?
- 3. Epistemology: Learning with Models
- 4. Models and Theory
- 5. Models and Other Debates in the Philosophy of Science
- 6. Conclusion
- Academic Tools
- Other Internet Resources
- Related Entries
Models can perform two fundamentally different representational functions. On the one hand, a model can be a representation of a selected part of the world (the ‘target system’). Depending on the nature of the target, such models are either models of phenomena or models of data. On the other hand, a model can represent a theory in the sense that it interprets the laws and axioms of that theory. These two notions are not mutually exclusive as scientific models can be representations in both senses at the same time.
Many scientific models represent a phenomenon, where ‘phenomenon’ is used as an umbrella term covering all relatively stable and general features of the world that are interesting from a scientific point of view. Empiricists like van Fraassen (1980) only allow for observables to qualify as such, while realists like Bogen and Woodward (1988) do not impose any such restrictions. The billiard ball model of a gas, the Bohr model of the atom, the double helix model of DNA, the scale model of a bridge, the Mundell-Fleming model of an open economy, or the Lorenz model of the atmosphere are well-known examples for models of this kind.
A first step towards a discussion of the issue of scientific representation is to realize that there is no such thing as the problem of scientific representation. Rather, there are different but related problems. It is not yet clear what specific set of questions a theory of representation has to come to terms with, but whatever list of questions one might put on the agenda of a theory of scientific representation, there are two problems that will occupy center stage in the discussion (Frigg 2006). The first problem is to explain in virtue of what a model is a representation of something else. To appreciate the thrust of this question we have to anticipate a position as regards the ontology of models (which we discuss in the next section). It is now common to construe models as non-linguistic entities rather than as descriptions. This approach has wide-ranging consequences. If we understand models as descriptions, the above question would be reduced to the time-honored problem of how language relates to reality and there would not be any problems over and above those already discussed in the philosophy of language. However, if we understand models as non-linguistic entities, we are faced with the new question of what it is for an object (that is not a word or a sentence) to scientifically represent a phenomenon.
Somewhat surprisingly, until recently this question has not attracted much attention in twentieth century philosophy of science, despite the fact that the corresponding problems in the philosophy of mind and in aesthetics have been discussed extensively for decades (there is a substantial body of literature dealing with the question of what it means for a mental state to represent a certain state of affairs; and the question of how a configuration of flat marks on a canvass can depict something beyond this canvass has puzzled aestheticians for a long time). However, some recent publications address this and other closely related problems (Bailer-Jones 2003, Contessa 2007, Elgin 2010, Frigg 2006, 2010c, Knuuttila 2009, Morrison 2009, Giere 2004, Suárez 2003, 2004, 2009, Suárez and Solé 2006, Thomson-Jones 2010, Toon 2010, 2011, 2012, van Fraassen 2004), while others dismiss it as a non-issue (Callender and Cohen 2006, 2008 Teller 2001).
The second problem is concerned with representational styles. It is a commonplace that one can represent the same subject matter in different ways. This pluralism does not seem to be a prerogative of the fine arts as the representations used in the sciences are not all of one kind either. Weizsäcker's liquid drop model represents the nucleus of an atom in a manner very different from the shell model, and a scale model of the wing of an air plane represents the wing in a way that is different from how a mathematical model of its shape does. What representational styles are there in the sciences?
Although this question is not explicitly addressed in the literature on the so-called semantic view of theories, some answers seem to emerge from its understanding of models. One version of the semantic view, one that builds on a mathematical notion of models (see Sec. 2), posits that a model and its target have to be isomorphic (van Fraassen 1980; Suppes 2002) or partially isomorphic (Da Costa and French 2003) to each other. Formal requirements weaker than these have been discussed by Mundy (1986) and Swoyer (1991). Another version of the semantic view drops formal requirements in favor of similarity (Giere 1988 and 2004, Teller 2001). This approach enjoys the advantage over the isomorphism view that it is less restrictive and also can account for cases of inexact and simplifying models. However, as Giere points out, this account remains empty as long as no relevant respects and degrees of similarity are specified. The specification of such respects and degrees depends on the problem at hand and the larger scientific context and cannot be made on the basis of purely philosophical considerations (Teller 2001).
Further notions that can be understood as addressing the issue of representational styles have been introduced in the literature on models. Among them, scale models, idealized models, analogical models and phenomenological models play an important role. These categories are not mutually exclusive; for instance, some scale models would also qualify as idealized models and it is not clear where exactly to draw the line between idealized and analogue models.
Scale models. Some models are basically down-sized or enlarged copies of their target systems (Black 1962). Typical examples are wooden cars or model bridges. The leading intuition is that a scale model is a naturalistic replica or a truthful mirror image of the target; for this reason scale models are sometimes also referred to as ‘true models’ (Achinstein 1968, Ch. 7). However, there is no such thing as a perfectly faithful scale model; faithfulness is always restricted to some respects. The wooden model of the car, for instance, provides a faithful portrayal of the car's shape but not its material. Scale models seem to be a special case of a broader category of representations that Peirce dubbed icons: representations that stand for something else because they closely resemble it (Peirce 1931–1958 Vol. 3, Para. 362). This raises the question of what criteria a model has to satisfy in order to qualify as an icon. Although we seem to have strong intuitions about how to answer this question in particular cases, no theory of iconicity for models has been formulated yet.
Idealized models. An idealization is a deliberate simplification of something complicated with the objective of making it more tractable. Frictionless planes, point masses, infinite velocities, isolated systems, omniscient agents, and markets in perfect equilibrium are but some well-know examples. Philosophical debates over idealization have focused on two general kinds of idealizations: so-called Aristotelian and Galilean idealizations.
Aristotelian idealization amounts to ‘stripping away’, in our imagination, all properties from a concrete object that we believe are not relevant to the problem at hand. This allows us to focus on a limited set of properties in isolation. An example is a classical mechanics model of the planetary system, describing the planets as objects only having shape and mass, disregarding all other properties. Other labels for this kind of idealization include ‘abstraction’ (Cartwright 1989, Ch. 5), ‘negligibility assumptions’ (Musgrave 1981) and ‘method of isolation’ (Mäki 1994).
Galilean idealizations are ones that involve deliberate distortions. Physicists build models consisting of point masses moving on frictionless planes, economists assume that agents are omniscient, biologists study isolated populations, and so on. It was characteristic of Galileo's approach to science to use simplifications of this sort whenever a situation was too complicated to tackle. For this reason it is common to refer to this sort of idealizations as ‘Galilean idealizations’ (McMullin 1985); another common label is ‘distorted models’.
Galilean idealizations are beset with riddles. What does a model involving distortions of this kind tell us about reality? How can we test its accuracy? In reply to these questions Laymon (1991) has put forward a theory which understands idealizations as ideal limits: imagine a series of experimental refinements of the actual situation which approach the postulated limit and then require that the closer the properties of a system come to the ideal limit, the closer its behavior has to come to the behavior of the ideal limit (monotonicity). But these conditions need not always hold and it is not clear how to understand situations in which no ideal limit exists. We can, at least in principle, produce a series of table tops that are ever more slippery but we cannot possibly produce a series of systems in which Planck's constant approaches zero. This raises the question of whether one can always make an idealized model more realistic by de-idealizing it. We will come back to this issue in section 5.1.
Galilean and Aristotelian idealizations are not mutually exclusive. On the contrary, they often come together. Consider again the mechanical model of the planetary system: the model only takes into account a narrow set of properties and distorts these, for instance by describing planets as ideal spheres with a rotation-symmetric mass distribution.
Models that involve substantial Galilean as well as Aristotelian idealizations are sometimes referred to as ‘caricatures’ (Gibbard and Varian 1978). Caricature models isolate a small number of salient characteristics of a system and distort them into an extreme case. A classical example is Ackerlof's (1970) model of the car market, which explains the difference in price between new and used cars solely in terms of asymmetric information, thereby disregarding all other factors that may influence prices of cars. However, it is controversial whether such highly idealized models can still be regarded as informative representations of their target systems (for a discussion of caricature models, in particular in economics, see Reiss 2006).
At this point we would like to mention a notion that seems to be closely related to idealization, namely approximation. Although the terms are sometimes used interchangeably, there seems to be a clear difference between the two. Approximations are introduced in a mathematical context. One mathematical item is an approximation of another one if it is close to it in some relevant sense. What this item is may vary. Sometimes we want to approximate one curve with another one. This happens when we expand a function into a power series and only keep the first two or three terms. In other situations we approximate an equation by another one by letting a control parameter tend towards zero (Redhead 1980). The salient point is that the issue of physical interpretation need not arise. Unlike Galilean idealization, which involves a distortion of a real system, approximation is a purely formal matter. This, of course, does not imply that there cannot be interesting relations between approximations and idealization. For instance, an approximation can be justified by pointing out that it is the ‘mathematical pendant’ to an acceptable idealization (e.g. when we neglect a dissipative term in an equation because we make the idealizing assumption that the system is frictionless).
Analogical models. Standard examples of analogical models include the hydraulic model of an economic system, the billiard ball model of a gas, the computer model of the mind or the liquid drop model of the nucleus. At the most basic level, two things are analogous if there are certain relevant similarities between them. Hesse (1963) distinguishes different types of analogies according to the kinds of similarity relations in which two objects enter. A simple type of analogy is one that is based on shared properties. There is an analogy between the earth and the moon based on the fact that both are large, solid, opaque, spherical bodies, receiving heat and light from the sun, revolving around their axes, and gravitating towards other bodies. But sameness of properties is not a necessary condition. An analogy between two objects can also be based on relevant similarities between their properties. In this more liberal sense we can say that there is an analogy between sound and light because echoes are similar to reflections, loudness to brightness, pitch to color, detectability by the ear to detectability by the eye, and so on.
Analogies can also be based on the sameness or resemblance of relations between parts of two systems rather than on their monadic properties. It is this sense that some politicians assert that the relation of a father to his children is analogous to the relation of the state to the citizens. The analogies mentioned so far have been what Hesse calls ‘material analogies’. We obtain a more formal notion of analogy when we abstract from the concrete features the systems possess and only focus on their formal set-up. What the analogue model then shares with its target is not a set of features, but the same pattern of abstract relationships (i.e. the same structure, where structure is understood in the formal sense). This notion of analogy is closely related to what Hesse calls ‘formal analogy’. Two items are related by formal analogy if they are both interpretations of the same formal calculus. For instance, there is a formal analogy between a swinging pendulum and an oscillating electric circuit because they are both described by the same mathematical equation.
A further distinction due to Hesse is the one between positive, negative and neutral analogies. The positive analogy between two items consists in the properties or relations they share (both gas molecules and billiard balls have mass), the negative analogy in the ones they do not share (billiard balls are colored, gas molecules are not). The neutral analogy comprises the properties of which it is not known yet whether they belong to the positive or the negative analogy. Neutral analogies play an important role in scientific research because they give rise to questions and suggest new hypotheses. In this vein, various authors have emphasized the heuristic role that analogies play in theory construction and in creative thought (Bailer-Jones and Bailer-Jones 2002; Hesse 1974, Holyoak and Thagard 1995, Kroes 1989, Psillos 1995, and the essays collected in Hellman 1988).
Phenomenological models. Phenomenological models have been defined in different, though related, ways. A traditional definition takes them to be models that only represent observable properties of their targets and refrain from postulating hidden mechanisms and the like. Another approach, due to McMullin (1968), defines phenomenological models as models that are independent of theories. This, however, seems to be too strong. Many phenomenological models, while failing to be derivable from a theory, incorporate principles and laws associated with theories. The liquid drop model of the atomic nucleus, for instance, portrays the nucleus as a liquid drop and describes it as having several properties (surface tension and charge, among others) originating in different theories (hydrodynamics and electrodynamics, respectively). Certain aspects of these theories—though usually not the complete theory—are then used to determine both the static and dynamical properties of the nucleus.
Concluding remarks. Each of these notions is still somewhat vague, suffers from internal problems, and much work needs to be done to tighten them. But more pressing than these issues is the question of how the different notions relate to each other. Are analogies fundamentally different from idealizations, or do they occupy different areas on a continuous scale? How do icons differ from idealizations and analogies? At present we do not know how to answer these questions. What we need is a systematic account of the different ways in which models can relate to reality and of how these ways compare to each other.
Another kind of representational models are so-called ‘models of data’ (Suppes 1962). A model of data is a corrected, rectified, regimented, and in many instances idealized version of the data we gain from immediate observation, the so-called raw data. Characteristically, one first eliminates errors (e.g. removes points from the record that are due to faulty observation) and then present the data in a ‘neat’ way, for instance by drawing a smooth curve through a set of points. These two steps are commonly referred to as ‘data reduction’ and ‘curve fitting’. When we investigate the trajectory of a certain planet, for instance, we first eliminate points that are fallacious from the observation records and then fit a smooth curve to the remaining ones. Models of data play a crucial role in confirming theories because it is the model of data and not the often messy and complex raw data that we compare to a theoretical prediction.
The construction of a data model can be extremely complicated. It requires sophisticated statistical techniques and raises serious methodological as well as philosophical questions. How do we decide which points on the record need to be removed? And given a clean set of data, what curve do we fit to it? The first question has been dealt with mainly within the context of the philosophy of experiment (see for instance Galison 1997 and Staley 2004). At the heart of the latter question lies the so-called curve fitting problem, which is that the data themselves do not indicate what form the fitted curve should take. Traditional discussions of theory choice suggest that this issue is settled by background theory, considerations of simplicity, prior probabilities, or a combination of these. Forster and Sober (1994) point out that this formulation of the curve fitting problem is a slight overstatement because there is a theorem in statistics due to Akaike which shows (given certain assumptions) that the data themselves underwrite (though not determine) an inference concerning the curve's shape if we assume that the fitted curve has to be chosen such that it strikes a balance between simplicity and goodness of fit in a way that maximizes predictive accuracy. Further discussions of data models can be found in Chin and Brewer (1994), Harris (2003), Laymon (1982) and Mayo (1996).
In modern logic, a model is a structure that makes all sentences of a theory true, where a theory is taken to be a (usually deductively closed) set of sentences in a formal language (see Bell and Machover 1977 or Hodges 1997 for details). The structure is a ‘model’ in the sense that it is what the theory represents. As a simple example consider Euclidean geometry, which consists of axioms—e.g. ‘any two points can be joined by a straight line’—and the theorems that can be derived from these axioms. Any structure of which all these statements are true is a model of Euclidean geometry.
A structure S = <U, O, R> is a composite entity consisting of (i) a non-empty set U of individuals called the domain (or universe) of S, (ii) an indexed set O (i.e. an ordered list) of operations on U (which may be empty), and (iii) a non-empty indexed set R of relations on U. It is important to note that nothing about what the objects are matters for the definition of a structure—they are mere dummies. Similarly, operations and functions are specified purely extensionally; that is, n-place relations are defined as classes of n-tuples, and functions taking n arguments are defined as classes of (n+1)-tuples. If all sentences of a theory are true when its symbols are interpreted as referring to either objects, relations, or functions of a structure S, then S is a model of this theory.
Many models in science carry over from logic the idea of being the interpretation of an abstract calculus. This is particularly pertinent in physics, where general laws—such as Newton's equation of motion—lie at the heart of a theory. These laws are applied to a particular system—e.g. a pendulum—by choosing a special force function, making assumptions about the mass distribution of the pendulum etc. The resulting model then is an interpretation (or realization) of the general law.
There is a variety of things that are commonly referred to as models: physical objects, fictional objects, set-theoretic structures, descriptions, equations, or combinations of some of these. However, these categories are neither mutually exclusive nor jointly exhaustive. Where one draws the line between, say, fictional objects and set-theoretical structures may well depend on one's metaphysical convictions, and some models may fall into yet another class of things. What models are is, of course, an interesting question in its own right, but, as briefly indicated in the last section, it has also important implications for semantics and, as we will see below, for epistemology.
Some models are straightforward physical objects. These are commonly referred to as ‘material models’. The class of material models comprises anything that is a physical entity and that serves as a scientific representation of something else. Among the members of this class we find stock examples like wooden models of bridges, planes, or ships, Watson and Crick's metal model of DNA (Schaffner 1969) and Phillips' hydraulic model of the economy (Morgan and Boumans 2004). More cutting edge cases of material models are so-called model organisms: organisms used in the life sciences as stand-ins for other organisms (Ankeny 2009, Ankeny and Leonelli 2012 and Leonelli 2010).
Material models do not give rise to any ontological difficulties over and above the well-known quibbles in connection with objects, which metaphysicians deal with (e.g. the nature of properties, the identity of objects, parts and wholes, and so on).
Many models are not material models. The Bohr model of the atom, a frictionless pendulum, or isolated populations, for instance, are in the scientist's mind rather than in the laboratory and they do not have to be physically realized and experimented upon to perform their representational function. It seems natural to view them as fictional entities. This position can be traced back to the German neo-Kantian Vaihinger (1911), who emphasized the importance of fictions for scientific reasoning. Giere has recently advocated the view that models are abstract entities (1988, 81). It is not entirely clear what Giere means by ‘abstract entities’, but his discussion of mechanical models seems to suggest that he uses the term to designate fictional entities.
This view squares well with scientific practice, where scientists often talk about models as if they were objects, as well as with philosophical views that see the manipulation of models as an essential part of the process of scientific investigation (Morgan 1999). It is natural to assume that one can manipulate something only if it exists. Furthermore, models often have more properties than we explicitly attribute to them when we construct them, which is why they are interesting vehicles of research. A view that regards models as objects can easily explain this without further ado: when we introduce a model we use an identifying description, but the object itself is not exhaustively characterized by this description. Research then simply amounts to finding out more about the object thus identified.
The drawback of this suggestion is that fictional entities are notoriously beset with ontological riddles. This has led many philosophers to argue that there are no such things as fictional entities and that apparent ontological commitments to them must be renounced. The most influential of these deflationary accounts goes back to Quine (1953). Building on Russell's discussion of definite descriptions, Quine argues that it is an illusion that we refer to fictional entities when we talk about them. Instead, we can dispose of these alleged objects by turning the terms that refer to them into predicates and analyse sentences like ‘Pegasus does not exist’ as ‘nothing pegasizes’. By eliminating the troublesome term we eschew the ontological commitment they seem to carry. This has resulted in a lack of interest in fictional entities, in particular among philosophers of science. In a programmatic essay Fine (1993) draws attention to this neglect and submits that Quinean skepticism notwithstanding fictions play an important role in scientific reasoning. However, Fine does not offer a systematic account of fictions and of how they are put to use in science.
The issue of how to understand fictions in science has been the subject matter of a recent debate in the philosophy of modeling. Barberousse and Ludwig (2009), Contessa (2010), Frigg (2010a, 2010b), Godfrey-Smith (2006, 2009), Leng (2010) and Toon (2010) develop views that see models as fictions of some kind. Giere (2009) denies that his earlier work should be understood in this way and argues against viewing models as fictions. Magnani (2012), Pincock (2012, Ch.4) and Teller (2009) second Giere's anti-fictionalism and argue that models should not be regarded as fictions. Weisberg (2012) argues for a middle position which sees models as playing a heuristic role but denies that they form part of a scientific model.
An influential point of view takes models to be set-theoretic structures. This position can be traced back to Suppes (1960) and is now, with slight variants, held by most proponents of the semantic view of theories. Needless to say, there are differences between different versions of the semantic view (van Fraassen, for instance, emphasizes that models are state-space structures); a survey of the different positions can be found in Suppe (1989, Ch. 1). However, on all these accounts models are structures of one sort or another (Da Costa and French 2000). As models of this kind are often closely tied to mathematized sciences, they are sometimes also referred to as ‘mathematical models’. (For a discussion of such models in biology see Lloyd 1984 and 1994.)
This view of models has been criticized on different grounds. One pervasive criticism is that many types of models that play an important role in science are not structures and cannot be accommodated within the structuralist view of models, which can neither account for how these models are constructed nor for how they work in the context of investigation (Cartwright 1999, Downes 1992, Morrison 1999). Another charge held against the set-theoretic approach is that it is not possible to explain how structures represent a target system which forms part of the physical world without making assumptions that go beyond what the approach can afford (Frigg 2006).
A time-honored position has it that what scientists display in scientific papers and textbooks when they present a model are more or less stylized descriptions of the relevant target systems (Achinstein 1968, Black 1962).
This view has not been subject to explicit criticism. However, some of the criticisms that have been marshaled against the syntactic view of theories equally threaten a linguistic understanding of models. First, it is a commonplace that we can describe the same thing in different ways. But if we identify a model with its description, then each new description yields a new model, which seems to be counterintuitive. One can translate a description into other languages (formal or natural), but one would not say that one hereby obtains a different model. Second, models have different properties than descriptions. On the one hand, we say that the model of the solar system consists of spheres orbiting around a big mass or that the population in the model is isolated from its environment, but it does not seem to make sense to say this about a description. On the other hand, descriptions have properties that models do not have. A description can be written in English, consist of 517 words, be printed in red ink, and so on. None of this makes any sense when said about a model. The descriptivist faces the challenge to either make a case that these arguments are mistaken or to show how to get around these difficulties.
Another group of things that is habitually referred to as ‘models’, in particular in economics, is equations (which are then also termed ‘mathematical models’). The Black-Scholes model of the stock market or the Mundell-Fleming model of an open economy are cases in point.
The problem with this suggestion is that equations are syntactic items and as such they face objections similar to the ones put forward against descriptions. First, one can describe the same situation using different co-ordinates and as a result obtain different equations; but we do not seem to obtain a different model. Second, the model has properties different from the equation. An oscillator is three-dimensional but the equation describing its motion is not. Equally, an equation may be inhomogeneous but the system it describes is not.
The proposals discussed so far have tacitly assumed that a model belongs to one particular class of objects. But this assumption is not necessary. It might be the case that models are a mixture of elements belonging to different ontological categories. In this vein Morgan (2001) suggests that models involve structural as well as narrative elements (‘stories’, as she calls them).
Models are vehicles for learning about the world. Significant parts of scientific investigation are carried out on models rather than on reality itself because by studying a model we can discover features of and ascertain facts about the system the model stands for; in brief, models allow for surrogative reasoning (Swoyer 1991). For instance, we study the nature of the hydrogen atom, the dynamics of populations, or the behavior of polymers by studying their respective models. This cognitive function of models has been widely acknowledged in the literature, and some even suggest that models give rise to a new style of reasoning, so-called ‘model based reasoning’ (Magnani and Nersessian 2002, Magnani, Nersessian and Thagard 1999). This leaves us with the question of how learning with a model is possible.
Hughes (1997) provides a general framework for discussing this question. According to his so-called DDI account, learning takes place in three stages: denotation, demonstration, and interpretation. We begin by establishing a representation relation (‘denotation’) between the model and the target. Then we investigate the features of the model in order to demonstrate certain theoretical claims about its internal constitution or mechanism; i.e. we learn about the model (‘demonstration’). Finally, these findings have to be converted into claims about the target system; Hughes refers to this step as ‘interpretation’. It is the latter two notions that are at stake here.
Learning about a model happens at two places, in the construction and the manipulation of the model (Morgan 1999). There are no fixed rules or recipes for model building and so the very activity of figuring out what fits together and how affords an opportunity to learn about the model. Once the model is built, we do not learn about its properties by looking at it; we have to use and manipulate the model in order to elicit its secrets.
Depending on what kind of model we are dealing with, building and manipulating a model amounts to different activities demanding a different methodology. Material models seem to be unproblematic as they are used in common experimental contexts (e.g. we put the model of a car in the wind tunnel and measure its air resistance). Hence, as far as learning about the model is concerned, material models do not give raise to questions that go beyond the questions of experimentation more generally.
Not so with fictional models. What constraints are there to the construction of fictional models and how do we manipulate them? The natural response seems to be that we answer these questions by performing a thought experiment. Different authors (e.g. Brown 1991, Gendler 2000, Norton 1991, Reiss 2003, Sorensen 1992) have explored this line of argument but they have reached very different and often conflicting conclusions as to how thought experiments are performed and what the status of their outcomes is (for details see the entry on thought experiments).
An important class of models is of mathematical nature. In some cases it is possible to derive results or solve equations analytically. But quite often this is not the case. It is at this point where the invention of the computer had a great impact, as it allows us to solve equations which are otherwise intractable by making a computer simulation. Many parts of current research in both the natural and social sciences rely on computer simulations. The formation and development of stars and galaxies, the detailed dynamics of high-energy heavy ion reactions, aspects of the intricate process of the evolution of life as well as the outbreak of wars, the progression of an economy, decision procedures in an organization and moral behavior are explored with computer simulations, to mention only a few examples (Hegselmann et al. 1996, Skyrms 1996).
What is a simulation? Simulations characteristically are used in connection with dynamic models, i.e. models that involve time. The aim of a simulation is to solve the equations of motion of such a model, which is designed to represent the time-evolution of its target system. So one can say that a simulation imitates a (usually real) process by another process (Hartmann 1996, Humphreys 2004).
It has been claimed that computer simulations constitute a genuinely new methodology of science or even a new scientific paradigm, which, moreover also raise a new host of philosophical issues (Humphreys 2004, 2009, Rohrlich 1991, Winsberg 2001 and 2003, and various contributions to Sismondo and Gissis 1999). Hence the claim is that simulations call into question our philosophical understanding of many aspects of science. However, this enthusiasm is not shared universally and some argue that simulations, far from demanding a new philosophy of science, raise few if any new philosophical problems (Frigg and Reiss 2009).
Whether or not one sees computer simulations as raising fundamentally new philosophical issues, there is no doubt about their practical significance. When standard methods fail, computer simulations are often the only way to learn something about a dynamical model; they help us to ‘extend ourselves’ (Humphreys 2004), as it were. An important question that arises in this context is the justification of simulation results: why should we trust the outputs of a computer simulation? An influential strand of attempts to come to grips with this question exploits similarities between traditional experiments and computer simulations, which raises vexing questions about the relation between computer simulations and experiments (Barberousse, Franceschelli and Imbert 2009, Morgan 2003, Morrison 2009, Parker 2008, 2009, Winsberg 2003).
This question of trustworthiness can be broken down into sub-questions: (a) do the equations of the model represent the target system accurately enough for the purpose at hand and (b) does the computer provide accurate enough solutions of these equations. Practicioners refer to these, respectively, as the problem of validation and the problem of verification. In practice we often face a version of the Duhem problem because one can only evaluate the "net outcome" of a simulation and it is not possible to address these two issues one by one. This had led scientists to develop various methods to test whether the outcome of simulation is on target; for a discussion of these see Winsberg (2009, 2010).
Computer simulations are also heuristically important. They may suggest new theories, models and hypotheses, for example based on a systematic exploration of a model's parameter space (Hartmann 1996). But computer simulations also bear methodological perils. They may provide misleading results because due to the discrete nature of the calculations carried out on a digital computer they only allow for the exploration of a part of the full parameter space; and this subspace may not reveal certain important features of the model. The severity of this problem is somehow mitigated by the increasing power of modern computers. But the availability of more computational power also may have adverse effects. It may encourage scientists to swiftly come up with increasingly complex but conceptually premature models, involving poorly understood assumptions or mechanisms and too many additional adjustable parameters (for a discussion of a related problem in the context of individual actor models in the social sciences see Schnell 1990). This may lead to an increase in empirical adequacy—which may be welcome when it comes, for example, to forecasting the weather—but not necessarily to a better understanding of the underlying mechanisms. As a result, the use of computer simulations may change the weight we assign to the various goals of science. Finally, the availability of computer power may seduce scientists into making calculations that do not have the degree of trustworthiness that one would expect them to have. This happens, for instance, when computers are used to propagate probability distributions forward in time, which are then taken to be decision relevant probabilities even though they turn out not to be upon closer examination (see Frigg et al. 2012). So it is important not to be carried away with the means that new powerful computers offer and to thereby place out of sight the actual goals of research.
Once we have knowledge about the model, this knowledge has to be ‘translated’ into knowledge about the target system. It is at this point that the representational function of models becomes important again. Models can instruct us about the nature of reality only if we assume that (at least some of) the model's aspects have counterparts in the world. But if learning is tied to representation and if there are different kinds of representation (analogies, idealizations, etc.), then there are also different kinds of learning. If, for instance, we have a model we take to be a realistic depiction, the transfer of knowledge from the model to the target is accomplished in a different manner than when we deal with an analogue, or a model that involves idealizing assumptions.
What are these different ways of learning? Although numerous case studies have been made of how certain specific models work, there do not seem to be any general accounts of how the transfer of knowledge from a model to its target is achieved (this with the possible exception of theories of analogical reasoning, see references above). This is a difficult question, but it is one that deserves more attention than it has gotten so far.
One of the most perplexing questions in connection with models is how they relate to theories. The separation between models and theory is a very hazy one and in the jargon of many scientists it is often difficult, if not impossible, to draw a line. So the question is: is there a distinction between models and theories and if so how do they relate to one another?
In common parlance, the terms ‘model’ and ‘theory’ are sometimes used to express someone's attitude towards a particular piece of science. The phrase ‘it's just a model’ indicates that the hypothesis at stake is asserted only tentatively or is even known to be false, while something is awarded the label ‘theory’ if it has acquired some degree of general acceptance. However, this way of drawing a line between models and theories is of no use to a systematic understanding of models.
The syntactic view of theories, which is an integral part of the logical positivist picture of science, construes a theory as a set of sentences in an axiomatized system of first order logic. Within this approach, the term model is used in a wider and in a narrower sense. In the wider sense, a model is just a system of semantic rules that interpret the abstract calculus and the study of a model amounts to scrutinizing the semantics of a scientific language. In the narrower sense, a model is an alternative interpretation of a certain calculus (Braithwaite 1953, Campbell 1920, Nagel 1961, Spector 1965). If, for instance, we take the mathematics used in the kinetic theory of gases and reinterpret the terms of this calculus in a way that makes them refer to billiard balls, the billiard balls are a model of the kinetic theory of gases. Proponents of the syntactic view believe such models to be irrelevant to science. Models, they hold, are superfluous additions that are at best of pedagogical, aesthetical or psychological value (Carnap 1938, Hempel 1965; see also Bailer-Jones 1999).
The semantic view of theories (see e.g. van Fraassen 1980, Giere 1988, Suppe 1989, and Suppes 2002) reverses this standpoint and declares that we should dispense with a formal calculus altogether and view a theory as a family of models. Although different versions of the semantic view assume a different notion of model (see above) they all agree that models are the central unit of scientific theorizing.
One of the most perspicuous criticisms of the semantic view is that it mislocates the place of models in the scientific edifice. Models are relatively independent from theory, rather than being constitutive of them; or to use Morrison's (1998) slogan, they are ‘autonomous agents’. This independence has two aspects: construction and functioning (Morgan and Morrison 1999).
A look at how models are constructed in actual science shows that they are neither derived entirely from data nor from theory. Theories do not provide us with algorithms for the construction a model; they are not ‘vending machines’ into which one can insert a problem and a models pops out (Cartwright 1999, Ch. 8). Model building is an art and not a mechanical procedure. The London model of superconductivity affords us with a good example of this relationship. The model's principal equation has no theoretical justification (in the sense that it could be derived from electromagnetic or any other fundamental theory) and is motivated solely on the basis of phenomenological considerations (Cartwright et al. 1995). Or, to put it another way, the model has been constructed ‘bottom up’ and not ‘top down’ and therefore enjoys a great deal of independence from theory.
The second aspect of the independence of models is that they perform functions which they could not perform if they were a part of, or strongly dependent on, theories.
Models as complements of theories. A theory may be incompletely specified in the sense that it imposes certain general constraints but remains silent about the details of concrete situations, which are provided by a model (Redhead 1980). A special case of this situation is when a qualitative theory is known and the model introduces quantitative measures (Apostel 1961). Redhead's example for a theory that is underdetermined in this way is axiomatic quantum field theory, which only imposes certain general constraints on quantum fields but does not provide an account of particular fields.
While Redhead and others seem to think of cases of this sort as somehow special, Cartwright (1983) has argued that they are the rule rather than the exception. On her view, fundamental theories such as classical mechanics and quantum mechanics do not represent anything at all as they do not describe any real world situation. Laws in such theories are schemata that need to be concretized and filled with the details of a specific situation, which is a task that is accomplished by a model.
Models stepping in when theories are too complex to handle. Theories may be too complicated to handle. In such a case a simplified model may be employed that allows for a solution (Apostel 1961, Redhead 1980). Quantum chromodynamics, for instance, cannot easily be used to study the hadron structure of a nucleus, although it is the fundamental theory for this problem. To get around this difficulty physicists construct tractable phenomenological models (e.g. the MIT bag model) that effectively describes the relevant degrees of freedom of the system under consideration (Hartmann 1999). The advantage of these models is that they yield results where theories remain silent. Their drawback is that it is often not clear how to understand the relationship between the theory and the model as the two are, strictly speaking, contradictory.
A more extreme case is the use of a model when there are no theories at all available. We encounter this situation in all domains, but it is particularly rampant in biology and economics where overarching theories are often not to be had. The models that scientists then construct to tackle the situation are sometimes referred to as ‘substitute models’ (Groenewold 1961).
Models as preliminary theories. The notion of models as substitutes for theories is closely related to the notion of a developmental model. This term has been coined by Leplin (1980), who pointed out how useful models were in the development of early quantum theory and is now used as an umbrella notion covering cases in which models are some sort of a preliminary exercises to theory.
A closely related notion is the one of probing models (also ‘study models’ or ‘toy models’). These are models which do not perform a representational function and which are not expected to instruct us about anything beyond the model itself. The purpose of these models is to test new theoretical tools that are used later on to build representational models. In field theory, for instance, the so-called φ4-model has been studied extensively not because it represents anything real (it is well-known that it doesn't) but because it serves several heuristic functions. The simplicity of the φ4-model allows physicist to ‘get a feeling’ for what quantum field theories are like and to extract some general features that this simple model shares with more complicated ones. One can try complicated techniques such as renormalization in a simple setting and it is possible to get acquainted with mechanisms—in this case symmetry breaking—that can be used later on (Hartmann 1995). This is true not only for physics. As Wimsatt (1987) points out, false models in genetics can perform many useful functions, among them the following: the false model can help to answer questions about more realistic models, provide an arena for answering questions about properties of more complex models, ‘factor out’ phenomena that would not otherwise be seen, serve as a limiting case of a more general model (or two false models may define the extreme of a continuum of cases in which the real case is supposed to lie), or it can lead to the identification of relevant variables and the estimation of their values.
The debate over scientific models has important repercussions for other debates in the philosophy of science. The reason for this is that traditionally the debates over scientific realism, reductionism, explanation, and laws of nature were couched in terms of theories, because only theories were acknowledged as carriers of scientific knowledge. So the question is whether, and if so how, discussions of these matters change when we shift the focus from theories to models. Up to now, no comprehensive model-based accounts of any of these issues have been developed; but models did leave some traces in the discussions of these topics.
It has been claimed that the practice of model building favors antirealism over realism. Antirealists point out that truth is not the main goal of scientific modeling. Cartwright (1983), for instance, presents several case studies illustrating that good models are often false and that supposedly true theories might not help much when it comes to understanding, say, the working of a laser.
Realists deny that the falsity of models renders a realist approach to science impossible by pointing out that a good model, thought not literally true, is usually at least approximately true. Laymon (1985) argues that the predictions of a model typically become better when we relax idealizations (i.e. de-idealize the model), which he takes to support realism (see also McMullin 1985, Nowak 1979 and Brzezinski and Nowak 1992).
Apart from the usual complaints about the elusiveness of the notion of approximate truth, antirealists have taken issue with this reply for two (related) reasons. First, as Cartwright (1989) points out, there is no reason to assume that one can always improve a model by adding de-idealizing corrections. Second, it seems that the outlined procedure is not in accordance with scientific practice. It is unusual that scientists invest work in repeatedly de-idealizing an existing model. Rather, they shift to a completely different modeling framework once the adjustments to be made get too involved (Hartmann 1998). The various models of the atomic nucleus are a case in point. Once it has been realized that shell effects are important to understand various phenomena, the (collective) liquid drop model has been put aside and the (single-particle) shell model has been developed to account for these findings. A further difficulty with de-idealization is that most idealizations are not ‘controlled’. It is, for example, not clear in what way one could de-idealize the MIT-Bag Model to eventually arrive at quantum chromodynamics, the supposedly correct underlying theory.
A further antirealist argument, the ‘incompatible models argument’, takes as its starting point the observation that scientists often successfully use several incompatible models of one and the same target system for predictive purposes (Morrison 2000). These models seemingly contradict each other as they ascribe different properties to the same target system. In nuclear physics, for instance, the liquid drop model explores the analogy of the atomic nucleus with a (charged) fluid drop, while the shell model describes nuclear properties in terms of the properties of protons and neutrons, the constituents of an atomic nucleus. This practice appears to cause a problem for scientific realism. Realists typically hold that there is a close connection between the predictive success of a theory and its being at least approximately true. But if several theories of the same system are predictively successful and if these theories are mutually inconsistent, they cannot all be true, not even approximately.
Realists can react to this argument in various ways. First, they can challenge the claim that the models in question are indeed predictively successful. If the models aren't good predictors, the argument is blocked. Second, they can defend a version of perspectival realism (Giere 1999, Rueger 2005) according to which each model reveals one aspect of the phenomenon in question, and when taken together a full (or fuller) account emerges. Third, realists can deny that there is a problem in the first place because scientific models, which are always idealized in one way or another and therefore strictly speaking false, are just the wrong vehicle to make a point about realism.
The multiple-models problem mentioned in the last section raises the question of how different models are related. Evidently, multiple models for the same target system do not generally stand in a deductive relation as they often contradict each other. Given that most of these models seem to be indispensable to the practice of science, a simple picture of the organization of science along the lines of Nagel's (1961) model of reduction or Oppenheim and Putnam's (1958) pyramid picture does not seem plausible.
Some have suggested (Cartwright 1999, Hacking 1983) a picture of science according to which there are no systematic relations that hold between different models. Some models are tied together because they represent the same target system, but this does not imply that they enter into any further relationships (deductive or otherwise). We are confronted with a patchwork of models, all of which hold ceteris paribus in their specific domains of applicability (see also the papers collected in Falkenburg and Muschik 1998).
Some argue that this picture is at least partially incorrect because there are various interesting relations that hold between different models or theories. These relations range from controlled approximations over singular limit relations (Batterman 2004) to structural relations (Gähde 1997) and rather loose relations called stories (Hartmann 1999; see also Bokulich 2003). These suggestions have been made on the basis of cases studies (for instance of so-called effective quantum field theories, see Hartmann 2001) and it remains to be seen whether a more general account of these relations can be given and whether a deeper justification for them can be provided, for instance, with a Bayesian framework (first steps towards a Bayesian understanding of reduction can be found in Dizadji-Bahmani et al. 2011).
It is widely held that science aims at discovering laws of nature. Philosophers, in turn, have been faced with the challenge of explicating what laws of nature are. According to the two currently dominant accounts, the best systems approach and the universals approach, laws of nature are understood to be universal in scope, meaning that they apply to everything that there is in the world. This take on laws does not seem to square with a view that sees models at the center of scientific theorizing. What role do general laws play in science if it is models that represent what is happening in the world and how are models and laws related?
One possible response is to argue that laws of nature govern entities and processes in a model rather than in the world. Fundamental laws, on this approach, do not state facts about the world but hold true of entities and processes in the model. Different variants of this view have been advocated by Cartwright (1983, 1999), Giere (1999), and van Fraassen (1989). Surprisingly, realists about laws do not seem to have responded to this challenge and so it remains an open question whether (and if so how) a realistic understanding of laws and a model-based approach to science can be made compatible.
Laws of nature play an important role in many accounts of explanation, most prominently in the deductive-nomological model and the unification approach. Unfortunately, these accounts inherit the problems that beset the relationship between models and laws. This leaves us with two options. Either one can argue that laws can be dispensed with in explanations, an idea which is employed in both van Fraassen's (1980) pragmatic theory of explanation and approaches to causal explanation such as Woodward's (2003). According to the latter, models are tools to find out about the causal relations that hold between certain facts or processes and it is these relations that do the explanatory job. Or one can shift the explanatory burden on models. A positive suggestion along these lines is Cartwright's so-called ‘simulacrum account of explanation’, which suggests that we explain a phenomenon by constructing a model that fits the phenomenon into the basic framework of a grand theory (1983, Ch. 8). On this account, the model itself is the explanation we seek. This squares well with basic scientific intuitions but leaves us with the question of what notion of explanation is at work (see also Elgin and Sober 2002). Bokulich (2008, 2009) pursues a similar line of reasoning and sees the explanatory power of models as being closely related to their fictional nature.
Models play an important role in science. But despite the fact that they have generated considerable interest among philosophers, there remain significant lacunas in our understanding of what models are and of how they work.
- Achinstein, Peter (1968), Concepts of Science. A Philosophical Analysis. Baltimore: Johns Hopkins Press.
- Ackerlof, George A (1970), “The Market for ‘Lemons’: Quality Uncertainty and the Market Mechanism”, Quarterly Journal of Economics 84: 488–500.
- Ankeny, Rachel (2009), “Model Organisms as Fictions”, In Mauricio Suárez (ed.): Fictions in Science, Philosophical Essays on Modelling and Idealisation, London: Routledge, 194–204.
- Apostel, Leo (1961), “Towards the Formal Study of Models in the Non-Formal Sciences”, in Freudenthal 1961, 1–37.
- Bailer-Jones, Daniela M. (1999), “Tracing the Development of Models in the Philosophy of Science”, in Magnani, Nersessian and Thagard 1999, 23–40.
- ––– (2003) “When Scientific Models Represent”, International Studies in the Philosophy of Science 17: 59–74.
- ––– and Bailer-Jones C. A. L. (2002), “Modelling Data: Analogies in Neural Networks, Simulated Annealing and Genetic Algorithms”, in Magnani and Nersessian 2002: 147–165.
- Barbrousse, Anouk and Pascal Ludwig (2009), “Fictions and Models”, In Mauricio Suárez (ed.): Fictions in Science, Philosophical Essays on Modelling and Idealisation, London: Routledge, 56–75.
- ––– Sara Franceschelli, Cyrille Imbert (2009), “Computer Simulations as Experiments”, Synthese, 169 (3): 557–574.
- Batterman, Robert (2004), “Intertheory Relations in Physics”, The Stanford Encyclopedia of Philosophy (Spring 2004 Edition), Edward N. Zalta (ed.), URL = http://plato.stanford.edu/archives/spr2004/entries/physics-interrelate/.
- Bell, John and Moshé Machover (1977), A Course in Mathematical Logic, Amsterdam: North-Holland.
- Black, Max (1962), Models and Metaphors. Studies in Language and Philosophy. Ithaca, New York: Cornell University Press.
- Bogen, James and James Woodward (1988), “Saving the Phenomena,” Philosophical Review 97: 303–352.
- Bokulich, Alisa (2003), “Horizontal Models: From Bakers to Cats”, Philosophy of Science 70: 609–627.
- ––– (2008), Reexamining the Quantum-Classical Relation: Beyond Reductionism and Pluralism, Cambridge: Cambridge University Press.
- ––– (2009), “Explanatory Fictions”, In: Mauricio Suárez (ed.): Fictions in Science. Philosophical Essays on Modelling and Idealisation London: Routledge, 91–109.
- Braithwaite, Richard (1953), Scientific Explanation. Cambridge: Cambridge University Press.
- Brewer, W. F. and C. A. Chinn (1994), “Scientists’ Responses to Anomalous Data: Evidence from Psychology, History, and Philosophy of Science,” in: Proceedings of the 1994 Biennial Meeting of the Philosophy of Science Association, Volume 1: Symposia and Invited Papers, 304–313.
- Brown, James (1991), The Laboratory of the Mind: Thought Experiments in the Natural Sciences. London: Routledge.
- Brzezinski, Jerzy and Leszek Nowak (eds.) (1992), Idealization III: Approximation and Truth. Amsterdam: Rodopi.
- Callender, Craig and Jonathan Cohen (2006), “There Is No Special Problem About Scientific Representation,” Theoria, forthcoming.
- Campbell, Norman (1920), Physics: The Elements. Cambridge: Cambridge University Press. Reprinted as Foundations of Science. New York: Dover, 1957.
- Carnap, Rudolf (1938), “Foundations of Logic and Mathematics”, in Otto Neurath, Charles Morris and Rudolf Carnap (eds.), International Encyclopaedia of Unified Science. Vol. 1. Chicago: University of Chicago Press, 139–213.
- Cartwright, Nancy (1983), How the Laws of Physics Lie. Oxford: Oxford University Press.
- ––– (1989), Nature's Capacities and their Measurement. Oxford: Oxford University Press.
- ––– (1999), The Dappled World. A Study of the Boundaries of Science. Cambridge: Cambridge University Press.
- ––– Towfic Shomar and Mauricio Suárez (1995), “The Tool-box of Science”, in Herfel 1995, 137–150.
- Contessa, Gabrielle (2007) “Scientific Representation, Interpretation and Surrogative Reasoning”, Philosophy of Science 74(1): 48–68.
- ––– (2010), “Scientific Models and Fictional Objects”, Synthese 172 (2), 215–229.
- Da Costa, Newton, and Steven French (2000) “Models, Theories, and Structures: Thirty Years On”, Philosophy of Science 67, Supplement, S116–127.
- ––– (2003), Science and Partial Truth: A Unitary Approach to Models and Scientific Reasoning. Oxford: Oxford University Press.
- Dizadji-Bahmani, Foad, Roman Frigg and Stephan Hartmann (2011) “Confirmation and Reduction: A Bayesian Account”, Synthese 179(2): 321–38.
- Downes, Stephen (1992), “The Importance of Models in Theorizing: A Deflationary Semantic View”. Proceedings of the Philosophy of Science Association, Vol.1, edited by David Hull et al., 142–153. East Lansing: Philosophy of Science Association.
- Elgin, Catherine (2010), “Telling Instances”, In: Roman Frigg and Matthew Hunter (eds.): Beyond Mimesis and Nominalism: Representation in Art and Science Berlin and Ney York: Springer, 1–17.
- Elgin, Mehmet and Elliott Sober (2002), “Cartwright on Explanation and Idealization”, Erkenntnis 57: 441–50.
- Falkenburg, Brigitte, and Wolfgang Muschik (eds.) (1998), Models, Theories and Disunity in Physics, Philosophia Naturalis 35.
- Fine, Arthur (1993), “Fictionalism”, Midwest Studies in Philosophy 18: 1–18.
- Forster, Malcolm, and Elliott Sober (1994), “How to Tell when Simple, More Unified, or Less Ad Hoc Theories will Provide More Accurate Predictions”, British Journal for the Philosophy of Science 45: 1–35.
- Freudenthal, Hans (ed.) (1961), The Concept and the Role of the Model in Mathematics and Natural and Social Sciences. Dordrecht: Reidel.
- Frigg, Roman (2006), “Scientific Representation and the Semantic View of Theories”, Theoria 55: 37–53.
- ––– and Julian Reiss (2009), “The Philosophy of Simulation: Hot New Issues or Same Old Stew?”, Synthese 169 (3): 593–613.
- ––– (2010a), “Fiction in Science”, In: John Woods (ed.): Fictions and Models: New Essays, Munich: Philosophia Verlag, 247–287
- ––– (2010b), “Models and Fiction”, Synthese, 172(2): 251–268
- ––– (2010c), “Fiction and Scientific Representation”, In: Roman Frigg and Metthew Hunter (eds.): Beyond Mimesis and Nominalism: Representation in Art and Science, Berlin and Ney York: Springer, 97–138.
- ––– Seamus Bradley, Reason L. Machete and Leonard A. Smith (2012), “Probabilistic Forecasting: Why Model Imperfection Is a Poison Pill”, forthcoming in Hanne Anderson, Dennis Dieks, Gregory Wheeler, Wenceslao Gonzalez and Thomas Uebel (eds): New Challenges to Philosophy of Science, Berlin and New York: Springer.
- Gähde, Ulrich (1997), “Anomalies and the Revision of Theory-Nets. Notes on the Advance of Mercury's Perihelion”, in Marisa Dalla Chiara et al. (eds.), Structures and Norms in Science. Dordrecht: Kluwer.
- Galison, Peter (1997) Image and Logic. A Material Culture of Microphysics. Chicago: Chicago: University of Chicago Press.
- Gendler, Tamar (2000) Thought Experiment: On the Powers and Limits of Imaginary Cases. New York and London: Garland.
- Gibbard, Allan and Hal Varian (1978), “Economic Models”, Journal of Philosophy 75: 664–677.
- Giere, Ronald (1988), Explaining Science: A Cognitive Approach. Chicago: University of Chicago Press.
- ––– (1999), Science Without Laws. Chicago: University of Chicago Press.
- ––– (2004), “How Models Are Used to Represent Reality”, Philosophy of Science 71, Supplement, S742–752.
- ––– (2009), “Why Scientific Models Should Not be Regarded as Works of Fiction”, In: Mauricio Suárez (ed.): Fictions in Science. Philosophical Essays on Modelling and Idealisation London: Routledge, 248–258.
- Godfrey-Smith, P. (2006), “The Strategy of Model-based Science”, Biology and Philosophy, 21: 725–740.
- ––– (2009), “Models and Fictions in Science”Philosophical Studies, 143: 101–116.
- Groenewold, H. J. (1961), “The Model in Physics” in Freudenthal 1961, 98–103.
- Hacking, Ian (1983), Representing and Intervening. Cambridge: Cambridge University Press.
- Harris, Todd (2003), “Data Models and the Acquisition and Manipulation of Data”, Philosophy of Science 70: 1508–1517.
- Hartmann, Stephan (1995), “Models as a Tool for Theory Construction: Some Strategies of Preliminary Physics”, in Herfel et al. 1995, 49–67.
- ––– (1996), “The World as a Process. Simulations in the Natural and Social Sciences”, in Hegselmann et al. 1996, 77–100.
- ––– (1998), “Idealization in Quantum Field Theory”, in Shanks 1998, 99–122.
- ––– (1999), “Models and Stories in Hadron Physics”, in Morgan and Morrison 1999, 326–346.
- ––– (2001), ‘Effective Field Theories, Reduction and Scientific Explanation’, Studies in History and Philosophy of Modern Physics 32, 267–304.
- Hegselmann, Rainer, Ulrich Müller and Klaus Troitzsch (eds.) (1996), Modelling and Simulation in the Social Sciences from the Philosophy of Science Point of View. Theory and Decision Library. Dordrecht: Kluwer.
- Hellman, D. H. (ed.) (1988), Analogical Reasoning. Kluwer: Dordrecht.
- Hempel, Carl G. (1965), Aspects of Scientific Explanation and other Essays in the Philosophy of Science. New York: Free Press.
- Herfel, William, Wladiyslaw Krajewski, Ilkka Niiniluoto and Ryszard Wojcicki (eds.) (1995), Theories and Models in Scientific Process. (Poznan Studies in the Philosophy of Science and the Humanities 44.) Amsterdam: Rodopi.
- Hesse, Mary (1963), Models and Analogies in Science. London: Sheed and Ward.
- ––– (1974), The Structure of Scientific Inference. London: Macmillan.
- Hodges, Wilfrid (1997), A Shorter Model Theory. Cambridge: Cambridge University Press.
- Holyoak, Keith and Paul Thagard (1995), Mental Leaps. Analogy in Creative Thought. Cambridge, Mass.: Bradford.
- Horowitz, Tamara and Gerald Massey (eds.) (1991), Thought Experiments in Science and Philosophy. Lanham: Rowman and Littlefield.
- Hughes, R. I. G. (1997), “Models and Representation”, Philosophy of Science 64: S325–336.
- Humphreys, Paul (2004), Extending Ourselves: Computational Science, Empiricism, and Scientific Method. Oxford: Oxford University Press.
- ––– (2009), “The Philosophical Novelty of Computer Simulation Methods”, Synthese 169: 615–626.
- Knuuttila, Taria (2009), “Representation, Idealisation and Fiction in Economics: From the Assumptions Issue to the Epistemology of Modelling”, In: Mauricio Suárez (ed.): Fictions in Science. Philosophical Essays on Modelling and Idealisation, London: Routledge, 205–233.
- Kroes, Peter (1989), “Structural Analogies between Physical Systems”, British Journal for the Philosophy of Science 40: 145–154.
- Laymon, Ronald (1982), “Scientific Realism and the Hierarchical Counterfactual Path from Data to Theory”, Proceedings of the Biennial Meeting of the Philosophy of Science Association, Volume 1, 107–121.
- ––– (1985), “Idealizations and the Testing of Theories by Experimentation”, in Peter Achinstein and Owen Hannaway (eds.), Observation Experiment and Hypothesis in Modern Physical Science. Cambridge, Mass.: M.I.T. Press, 147–173.
- ––– (1991), “Thought Experiments by Stevin, Mach and Gouy: Thought Experiments as Ideal Limits and Semantic Domains”, in Horowitz and Massey 1991, 167–91.
- Leng, Mary (2010), Mathematics and Reality, Oxford.
- Leonelli, Sabina (2010), “Packaging Data for Re-Use: Databases in Model Organism Biology”, In: Howlett P,Morgan MS (eds.): How Well Do Facts Travel? The Dissemination of Reliable Knowledge, Cambridge: Cambridge University Press.
- Leonelli, Sabina and R. Ankeny (2012), “Re-Thinking Organisms: The Epistemic Impact of Databases on Model Organism Biology”, Studies in the History and Philosophy of the Biological and Biomedical Sciences, 43, 29–36.
- Leplin, Jarrett (1980), “The Role of Models in Theory Construction”, in: T. Nickles (ed.), Scientific Discovery, Logic, and Rationality. Reidel: Dordrecht: 267–284.
- Lloyd, Elisabeth (1984), “A Semantic Approach to the Structure of Population Genetics”, Philosophy of Science 51: 242–264.
- ––– (1994), The Structure and Confirmation of Evolutionary Theory. Princeton: Princeton University Press.
- Magnani, Lorenzo, and Nancy Nersessian (eds.) (2002), Model-Based Reasoning: Science, Technology, Values. Dordrecht: Kluwer.
- ––– (2012), Scientific Models Are Not Fictions: Model-Based Science as Epistemic Warfar, Forthcoming in L. Magnani and P. Li (eds.): Philosophy and Cognitive Science: Western and Eastern Studies, Heidelberg/Berlin: Springer.
- ––– and Paul Thagard (eds.) (1999), Model-Based Reasoning In Scientific Discovery. Dordrecht: Kluwer.
- Mäki, Uskali (1994), “Isolation, Idealization and Truth in Economics”, in Bert Hamminga and Neil B. De Marchi (eds.), Idealization VI: Idealization in Economics. Poznan Studies in the Philosophy of the Sciences and the Humanities, Vol. 38: 147–168. Amsterdam: Rodopi.
- Mayo, Deborah (1996), Error and the Growth of Experimental Knowledge. Chicago: University of Chicago Press.
- McMullin, Ernan (1968), “What Do Physical Models Tell Us?”, in B. van Rootselaar and J. F. Staal (eds.), Logic, Methodology and Science III. Amsterdam: North Holland, 385–396.
- ––– (1985), “Galilean Idealization”, Studies in the History and Philosophy of Science 16: 247–73.
- Morgan, Mary (1999), “Learning from Models”, in Morgan and Morrison 1999, 347–88.
- ––– and Margaret Morrison (1999), Models as Mediators. Perspectives on Natural and Social Science. Cambridge: Cambridge University Press.
- ––– and Margaret Morrison (1999), “Models as Mediating Instruments”, In: Morgan and Morrison 1999, 10–37.
- ––– (2001) “Models, Stories and the Economic World”, Journal of Economic Methodology 8:3, 361–84. Reprinted in Fact and Fiction in Economics, edited by Uskali Mäki, 178–201. Cambridge: Cambridge University Press, 2002.
- ––– (2003) “Experiments without Material Intervention: Model Experiments, Virtual Experiments and Virtually Experiments”, In: H. Radder (ed.): The Philosophy of Scientific Experimentation, Pittsburgh: University of Pittsburgh Press, 217–235
- ––– and Boumans, Marcel J. (2004), “Secrets Hidden by Two-Dimensionality: The Economy as a Hydraulic Machine”, In: S. de Chadarevian and N. Hopwood (eds.): Model: The Third Dimension of Science, Stanford: Stanford University Press, 369–401.
- Morrison, Margaret (1998), “Modelling Nature: Between Physics and the Physical World”, Philosophia Naturalis 35: 65–85.
- ––– (1999) “Models as Autonomous Agents”, in Morgan and Morrison 1999, 38–65.
- ––– (2000), Unifying Scientific Theories. Cambridge: Cambridge University Press.
- ––– (2009), “Fictions, Representations and Reality”, In: Mauricio Suárez (ed.): Fictions in Science. Philosophical Essays on Modelling and Idealisation, London: Routledge, 110–135.
- ––– (2009), “Models, Measurement and Computer Simulation: The Changing Face of Experimentation”, Philosophical Studies, 143 (1): 33–57.
- Mundy, Brent (1986), “On the General Theory of Meaningful Representation”, Synthese 67: 391–437.
- Musgrave, Alan (1981), “‘Unreal Assumptions’ in Economic Theory: The F-Twist Untwisted”, Kyklos 34: 377–387.
- Nagel, Ernest (1961), The Structure of Science. Problems in the Logic of Scientific Explanation. New York: Harcourt, Brace and World.
- Norton, John (1991), “Thought Experiments in Einstein's Work”, in Horowitz and Massey 1991, 129–148.
- Nowak, Leszek (1979), The Structure of Idealization: Towards a Systematic Interpretation of the Marxian Idea of Science. Reidel: Dordrecht.
- Oppenheim, Paul, and Hilary Putnam (1958), “Unity of Science as a Working Hypothesis”, in Herbert Feigl, Grover Maxwell, and Michael Scriven (eds.), Minnesota Studies in the Philosophy of Science. Minneapolis: University of Minnesota Press, 3–36. Reprinted in The Philosophy of Science, edited by Richard Boyd et al., Ch. 22. Cambridge, Mass.: M.I.T. Press, 1991.
- Parker, W.S. (2008), “Franklin, Holmes and the Epistemology of Computer Simulation”, International Studies in the Philosophy of Science 22(2): 165–183.
- ––– (2009), “Does Matter Really Matter? Computer Simulations, Experiments and Materiality”, Synthese 169: 483–496
- Peirce, Charles Sanders (1931–1958), Collected Papers of Charles Sanders Peirce. Vol 3. Edited by Charles Hartshorne, Paul Weiss, and Arthur Burks. Harvard University Press, Cambridge, Mass.
- Pincock, Christopher (2012) Mathematics and Scientific Representation, Oxford, Ch.12.
- Psillos, Stathis (1995), “The Cognitive Interplay between Theories and Models: The Case of 19th Century Physics”, in Herfel et al. 1995, 105–133.
- Quine, Willard Van Orman (1953), “On What There Is”, in From a Logical Point of View. Cambridge, Mass.: Harvard University Press.
- Redhead, Michael (1980), “Models in Physics”, British Journal for the Philosophy of Science 31: 145–163.
- Reiss, Julian (2003), “Causal Inference in the Abstract or Seven Myths about Thought Experiments”, in Causality: Metaphysics and Methods Research Project. Technical Report 03/02. LSE.
- ––– (2006), “Beyond Capacities”, in Luc Bovens and Stephan Hartmann (eds.), Nancy Cartwright's Philosophy of Science. London: Routledge.
- Rohrlich, Fritz (1991) “Computer Simulations in the Physical Sciences”, in Proceedings of the Philosophy of Science Association, Vol. 2, edited by Arthur Fine et al., 507–518. East Lansing: The Philosophy of Science Association.
- Rueger, Alexander (2005), “Perspectival Models and Theory Unification”, British Journal for the Philosophy of Science 56.
- Schnell, Rainer (1990), “Computersimulation und Theoriebildung in den Sozialwissenschaften”, Kölner Zeitschrift für Soziologie und Sozialpsychologie 1, 109–128.
- Schaffner, Kenneth F. (1969, “The Watson-Crick Model and Reductionism”, The British Journal for the Philosophy of Science, 20 (4): 325–348.
- Shanks, Niall (ed.). (1998), Idealization in Contemporary Physics. Amsterdam: Rodopi.
- Sismondo, Sergio and Snait Gissis (eds.) (1999), Modeling and Simulation. Special Issue of Science in Context 12.
- Skyrms, Brian (1996), Evolution of the Social Contract. Cambridge: Cambridge University Press.
- Sorensen, Roy (1992), Thought Experiments. New York: Oxford University Press.
- Spector, Marshall (1965), “Models and Theories”, British Journal for the Philosophy of Science 16: 121–142.
- Staley, Kent W. (2004), The Evidence for the Top Quark: Objectivity and Bias in Collaborative Experimentation. Cambridge: Cambridge University Press.
- Suárez, Mauricio (2003), “Scientific Representation: Against Similarity and Isomorphism.” International Studies in the Philosophy of Science 17: 225–244.
- ––– (2004), “An Inferential Conception of Scientific Representation”, Philosophy of Science 71, Supplement, S767–779.
- ––– and Albert Solé (2006), “On the Analogy between Cognitive Representation and Truth”, Theoria 55, 27–36.
- ––– (2009), “Scientific Fictions as Rules of Inference” In: Mauricio Suárez (ed.): Fictions in Science. Philosophical Essays on Modelling and Idealisation, Routledge: London, 158–178.
- Suppe, Frederick. (1989), The Semantic View of Theories and Scientific Realism. Urbana and Chicago: University of Illinois Press.
- Suppes, Patrick. (1960), “A Comparison of the Meaning and Uses of Models in Mathematics and the Empirical Sciences”, Synthèse 12: 287–301. Reprinted in Freudenthal (1961), 163–177, and in Patrick Suppes: Studies in the Methodology and Foundations of Science. Selected Papers from 1951 to 1969. Dordrecht: Reidel 1969, 10–23.
- ––– (1962), “Models of Data”, in Ernest Nagel, Patrick Suppes and Alfred Tarski (eds.), Logic, Methodology and Philosophy of Science: Proceedings of the 1960 International Congress. Stanford: Stanford University Press, 252–261. Reprinted in Patrick Suppes: Studies in the Methodology and Foundations of Science. Selected Papers from 1951 to 1969. Dordrecht: Reidel 1969, 24–35.
- ––– (2002), Representation and Invariance of Scientific Structures. Stanford: CSLI Publications.
- Swoyer, Chris (1991), “Structural Representation and Surrogative Reasoning”, Synthese 87: 449–508.
- Teller, Paul (2001), “Twilight of the Perfect Model”, Erkenntnis 55, 393–415.
- ––– (2004), “How We Dapple the World”, Philosophy of Science 71: 425–447.
- ––– (2009), “Fictions, Fictionalization, and Truth in Science”, In: Mauricio Suárez (ed.): Fictions in Science. Philosophical Essays on Modelling and Idealisation, London: Routledge, 235–247.
- Thomson-Jones, Martin (2010), “Missing Systems and the Face Value Practice”, Synthese 172 (2): 283–299.
- Toon, A. (2010), “Models as Make-Believe”, In: Frigg, R and Hunter, M. (eds.): Beyond Mimesis and Convention: Representation in Art and Science, Boston Studies in the Philosophy of Science: Springer, 71–96.
- ––– (2010), “The Ontology of Theoretical Modelling: Models as Make-Believe”, Synthese 172: 301–315.
- ––– (2011), “Playing with Molecules”, Studies in History and Philosophy of Science 42: 580–589.
- ––– (2012), Models as Make-Believe: Imagination, Fiction and Scientific Representation, Palgrave Macmillan.
- Vaihinger, Hans (1911), The Philosophy of ‘As If’. German original. English translation: London: Kegan Paul 1924.
- van Fraassen, Bas C. (1980), The Scientific Image. Oxford: Oxford University Press.
- ––– (1989), Laws and Symmetry. Oxford: Oxford University Press.
- ––– (2004), “Science as Representation: Flouting the Criteria”, Philosophy of Science 71, Supplement, S794–804.
- ––– (2008), Scientific Representation: Paradoxes of Perspective, Oxford: Oxford University Press.
- Weisberg, M. (2012), Simulation and Similarity: Using Models to Understand the World, forthcoming from Oxford University Press, Ch. 4.
- Wimsatt, William. (1987), “False Models as Means to Truer Theories”, in N. Nitecki and A. Hoffman(eds.), Neutral Models in Biology. Oxford: Oxford University Press, 23–55.
- Winsberg, Eric (2001), “Simulations, Models and Theories: Complex Physical Systems and their Representations”, Philosophy of Science 68 (Proceedings): 442–454.
- ––– (2003), “Simulated Experiments: Methodology for a Virtual World”, Philosophy of Science 70: 105–125.
- ––– (2009), “A Function for Fictions: Expanding the Scope of Science”, In: Mauricio Suárez (ed.): Fictions in Science. Philosophical Essays on Modelling and Idealisation, London: Routledge, 197–191.
- ––– (2010, Science in the Age of Computer Simulation, Chicago: Chicago University Press.
- Woodward, James (2003), Making Things Happen: A Theory of Causal Explanation. New York: Oxford University Press.
How to cite this entry. Preview the PDF version of this entry at the Friends of the SEP Society. Look up this entry topic at the Indiana Philosophy Ontology Project (InPhO). Enhanced bibliography for this entry at PhilPapers, with links to its database.
[Please contact the author with suggestions.]
We would like to thank Nancy Cartwright, Paul Humphreys, Julian Reiss, Elliott Sober, Chris Swoyer, and Paul Teller for helpful comments and suggestions on a draft of this entry.