V. Parsimonious Considerations regarding the Existence of God
Occasionalism is a variation upon Cartesian metaphysics. The latter is the most notorious case of dualism (mind and body, for instance). The mind is a "mental substance". The body - a "material substance". What permits the complex interactions which happen between these two disparate "substances"? The "unextended mind" and the "extended body" surely cannot interact without a mediating agency, God. The appearance is that of direct interaction but this is an illusion maintained by Him. He moves the body when the mind is willing and places ideas in the mind when the body comes across other bodies.
Descartes postulated that the mind is an active, unextended, thought while the body is a passive, unthinking extension. The First Substance and the Second Substance combine to form the Third Substance, Man. God - the Fourth, uncreated Substance - facilitates the direct interaction among the two within the third.
Foucher raised the question: how can God - a mental substance - interact with a material substance, the body. The answer offered was that God created the body (probably so that He will be able to interact with it). Leibniz carried this further: his Monads, the units of reality, do not really react and interact. They just seem to be doing so because God created them with a pre-established harmony. The constant divine mediation was, thus, reduced to a one-time act of creation. This was considered to be both a logical result of occasionalism and its refutation by a reductio ad absurdum argument.
But, was the fourth substance necessary at all? Could not an explanation to all the known facts be provided without it? The ratio between the number of known facts (the outcomes of observations) and the number of theory elements and entities employed in order to explain them is the parsimony ratio. Every newly discovered fact either reinforces the existing worldview or forces the introduction of a new one, through a "crisis" or a "revolution" (a "paradigm shift" in Kuhn's abandoned phrase).
The new worldview need not necessarily be more parsimonious. It could be that a single new fact precipitates the introduction of a dozen new theoretical entities, axioms and functions (curves between data points). The very delineation of the field of study serves to limit the number of facts, which could exercise such an influence upon the existing worldview and still be considered pertinent. Parsimony is achieved, therefore, also by affixing the boundaries of the intellectual arena and / or by declaring quantitative or qualitative limits of relevance and negligibility. The world is thus simplified through idealization. Yet, if this is carried too far, the whole edifice collapses. It is a fine balance that should be maintained between the relevant and the irrelevant, what matters and what could be neglected, the comprehensiveness of the explanation and the partiality of the pre-defined limitations on the field of research.
This does not address the more basic issue of why do we prefer simplicity to complexity. This preference runs through history: Aristotle, William of Ockham, Newton, Pascal - all praised parsimony and embraced it as a guiding principle of work scientific. Biologically and spiritually, we are inclined to prefer things needed to things not needed. Moreover, we prefer things needed to admixtures of things needed and not needed. This is so, because things needed are needed, encourage survival and enhance its chances. Survival is also assisted by the construction of economic theories. We all engage in theory building as a mundane routine. A tiger beheld means danger - is one such theory. Theories which incorporated fewer assumptions were quicker to process and enhanced the chances of survival. In the aforementioned feline example, the virtue of the theory and its efficacy lie in its simplicity (one observation, one prediction). Had the theory been less parsimonious, it would have entailed a longer time to process and this would have rendered the prediction wholly unnecessary. The tiger would have prevailed.
Thus, humans are Parsimony Machines (Ockham Machines): they select the shortest (and, thereby, most efficient) path to the production of true theorems, given a set of facts (observations) and a set of theories. Another way to describe the activity of Ockham Machines: they produce the maximal number of true theorems in any given period of time, given a set of facts and a set of theories.
Poincare, the French mathematician and philosopher, thought that Nature itself, this metaphysical entity which encompasses all, is parsimonious. He believed that mathematical simplicity must be a sign of truth. A simple Nature would, indeed, appear this way (mathematically simple) despite the filters of theory and language. The "sufficient reason" (why the world exists rather than not exist) should then be transformed to read: "because it is the simplest of all possible worlds". That is to say: the world exists and THIS world exists (rather than another) because it is the most parsimonious - not the best, as Leibniz put it - of all possible worlds.
Parsimony is a necessary (though not sufficient) condition for a theory to be labeled "scientific". But a scientific theory is neither a necessary nor a sufficient condition to parsimony. In other words: parsimony is possible within and can be applied to a non-scientific framework and parsimony cannot be guaranteed by the fact that a theory is scientific (it could be scientific and not parsimonious). Parsimony is an extra-theoretical tool. Theories are under-determined by data. An infinite number of theories fits any finite number of data. This happens because of the gap between the infinite number of cases dealt with by the theory (the application set) and the finiteness of the data set, which is a subset of the application set. Parsimony is a rule of thumb. It allows us to concentrate our efforts on those theories most likely to succeed. Ultimately, it allows us to select THE theory that will constitute the prevailing worldview, until it is upset by new data.
Another question arises which was not hitherto addressed: how do we know that we are implementing some mode of parsimony? In other words, which are the FORMAL requirements of parsimony?
The following conditions must be satisfied by any law or method of selection before it can be labeled "parsimonious":
1.. Exploration of a higher level of causality: the law must lead to a level of causality, which will include the previous one and other, hitherto apparently unrelated phenomena. It must lead to a cause, a reason which will account for the set of data previously accounted for by another cause or reason AND for additional data. William of Ockham was, after all a Franciscan monk and constantly in search for a Prima Causa.
2.. The law should either lead to, or be part of, an integrative process. This means that as previous theories or models are rigorously and correctly combined, certain entities or theory elements should be made redundant. Only those, which we cannot dispense with, should be left incorporated in the new worldview.
3.. The outcomes of any law of parsimony should be successfully subjected to scientific tests. These results should correspond with observations and with predictions yielded by the worldviews fostered by the law of parsimony under scrutiny.
4.. Laws of parsimony should be semantically correct. Their continuous application should bring about an evolution (or a punctuated evolution) of the very language used to convey the worldview, or at least of important language elements. The phrasing of the questions to be answered by the worldview should be influenced, as well. In extreme cases, a whole new language has to emerge, elaborated and formulated in accordance with the law of parsimony. But, in most cases, there is just a replacement of a weaker language with a more powerful meta-language. Einstein's Special Theory of Relativity and Newtonian dynamics are a prime example of such an orderly lingual transition, which was the direct result of the courageous application of a law of parsimony.
5.. Laws of parsimony should be totally subjected (actually, subsumed) by the laws of Logic and by the laws of Nature. They must not lead to, or entail, a contradiction, for instance, or a tautology. In physics, they must adhere to laws of causality or correlation and refrain from teleology.
6.. Laws of parsimony must accommodate paradoxes. Paradox Accommodation means that theories, theory elements, the language, a whole worldview will have to be adapted to avoid paradoxes. The goals of a theory or its domain, for instance, could be minimized to avoid paradoxes. But the mechanism of adaptation is complemented by a mechanism of adoption. A law of parsimony could lead to the inevitable adoption of a paradox. Both the horns of a dilemma are, then, adopted. This, inevitably, leads to a crisis whose resolution is obtained through the introduction of a new worldview. New assumptions are parsimoniously adopted and the paradox disappears.
7.. Paradox accommodation is an important hallmark of a true law of parsimony in operation. Paradox Intolerance is another. Laws of parsimony give theories and worldviews a "licence" to ignore paradoxes, which lie outside the domain covered by the parsimonious set of data and rules. It is normal to have a conflict between the non-parsimonious sets and the parsimonious one. Paradoxes are the results of these conflicts and the most potent weapons of the non-parsimonious sets. But the law of parsimony, to deserve it name, should tell us clearly and unequivocally, when to adopt a paradox and when to exclude it. To be able to achieve this formidable task, every law of parsimony comes equipped with a metaphysical interpretation whose aim it is to plausibly keep nagging paradoxes and questions at a distance. The interpretation puts the results of the formalism in the context of a meaningful universe and provides a sense of direction, causality, order and even "intent". The Copenhagen interpretation of Quantum Mechanics is an important member of this species.
8.. The law of parsimony must apply both to the theory entities AND to observable results, both part of a coherent, internally and externally consistent, logical (in short: scientific) theory. It is divergent-convergent: it diverges from strict correspondence to reality while theorizing, only to converge with it when testing the predictions yielded by the theory. Quarks may or may not exist - but their effects do, and these effects are observable.
9.. A law of parsimony has to be invariant under all transformations and permutations of the theory entities. It is almost tempting to say that it should demand symmetry - had this not been merely an aesthetic requirement and often violated.
10.. The law of parsimony should aspire to a minimization of the number of postulates, axioms, curves between data points, theory entities, etc. This is the principle of the maximization of uncertainty. The more uncertainty introduced by NOT postulating explicitly - the more powerful and rigorous the theory / worldview. A theory with one assumption and one theoretical entity - renders a lot of the world an uncertain place. The uncertainty is expelled by using the theory and its rules and applying them to observational data or to other theoretical constructs and entities. The Grand Unified Theories of physics want to get rid of four disparate powers and to gain one instead.
11.. A sense of beauty, of aesthetic superiority, of acceptability and of simplicity should be the by-products of the application of a law of parsimony. These sensations have been often been cited, by practitioners of science, as influential factors in weighing in favor of a particular theory.
12.. Laws of parsimony entail the arbitrary selection of facts, observations and experimental results to be related to and included in the parsimonious set. This is the parsimonious selection process and it is closely tied with the concepts of negligibility and with the methodology of idealization and reduction. The process of parsimonious selection is very much like a strategy in a game in which both the number of players and the rules of the game are finite. The entry of a new player (an observation, the result of an experiment) sometimes transforms the game and, at other times, creates a whole new game. All the players are then moved into the new game, positioned there and subjected to its new rules. This, of course, can lead to an infinite regression. To effect a parsimonious selection, a theory must be available whose rules will dictate the selection. But such a theory must also be subordinated to a law of parsimony (which means that it has to parsimoniously select its own facts, etc.). a meta-theory must, therefore, exist, which will inform the lower-level theory how to implement its own parsimonious selection and so on and so forth, ad infinitum.
13.. A law of parsimony falsifies everything that does not adhere to its tenets. Superfluous entities are not only unnecessary - they are, in all likelihood, false. Theories, which were not subjected to the tests of parsimony are, probably, not only non-rigorous but also positively false.
14.. A law of parsimony must apply the principle of redundant identity. Two facets, two aspects, two dimensions of the same thing - must be construed as one and devoid of an autonomous standing, not as separate and independent.
15.. The laws of parsimony are "back determined" and, consequently, enforce "back determination" on all the theories and worldviews to which they apply. For any given data set and set of rules, a number of parsimony sets can be postulated. To decide between them, additional facts are needed. These will be discovered in the future and, thus, the future "back determines" the right parsimony set. Either there is a finite parsimony group from which all the temporary groups are derived - or no such group exists and an infinity of parsimony sets is possible, the results of an infinity of data sets. This, of course, is thinly veiled pluralism. In the former alternative, the number of facts / observations / experiments that are required in order to determine the right parsimony set is finite. But, there is a third possibility: that there is an eternal, single parsimony set and all our current parsimony sets are its asymptotic approximations. This is monism in disguise. Also, there seems to be an inherent (though solely intuitive) conflict between parsimony and infinity.
16.. A law of parsimony must seen to be at conflict with the principle of multiplicity of substitutes. This is the result of an empirical and pragmatic observation: The removal of one theory element or entity from a theory - precipitates its substitution by two or more theory elements or entities (if the preservation of the theory is sought). It is this principle that is the driving force behind scientific crises and revolutions. Entities do multiply and Ockham's Razor is rarely used until it is too late and the theory has to be replaced in its entirety. This is a psychological and social phenomenon, not an inevitable feature of scientific progress. Worldviews collapse under the mere weight of their substituting, multiplying elements. Ptolemy's cosmology fell prey to the Copernican model not because it was more efficient, but because it contained less theory elements, axioms, equations. A law of parsimony must warn against such behaviour and restrain it or, finally, provide the ailing theory with a coup de grace.
17.. A law of parsimony must allow for full convertibility of the phenomenal to the nuomenal and of the universal to the particular. Put more simply: no law of parsimony can allow a distinction between our data and the "real" world to be upheld. Nor can it tolerate the postulation of Platonic "Forms" and "Ideas" which are not entirely reflected in the particular.
18.. A law of parsimony implies necessity. To assume that the world is contingent is to postulate the existence of yet another entity upon which the world is dependent for its existence. It is to theorize on yet another principle of action. Contingency is the source of entity multiplication and goes against the grain of parsimony. Of course, causality should not be confused with contingency. The former is deterministic - the latter the result of some kind of free will.
19.. The explicit, stated, parsimony, the one formulated, formalized and analyzed, is connected to an implicit, less evident sort and to latent parsimony. Implicit parsimony is the set of rules and assumptions about the world that are known as formal logic. The latent parsimony is the set of rules that allows for a (relatively) smooth transition to be effected between theories and worldviews in times of crisis. Those are the rules of parsimony, which govern scientific revolutions. The rule stated in article (a) above is a latent one: that in order for the transition between old theories and new to be valid, it must also be a transition between a lower level of causality - and a higher one.
Efficient, workable, parsimony is either obstructed, or merely not achieved through the following venues of action:
1.. Association - the formation of networks of ideas, which are linked by way of verbal, intuitive, or structural association, does not lead to more parsimonious results. Naturally, a syntactic, grammatical, structural, or other theoretical rule can be made evident by the results of this technique. But to discern such a rule, the scientist must distance himself from the associative chains, to acquire a bird's eye view , or, on the contrary, to isolate, arbitrarily or not, a part of the chain for closer inspection. Association often leads to profusion and to embarrassment of riches. The same observations apply to other forms of chaining, flowing and networking.
2.. Incorporation without integration (that is, without elimination of redundancies) leads to the formation of hybrid theories. These cannot survive long. Incorporation is motivated by conflict between entities, postulates or theory elements. It is through incorporation that the protectors of the "old truth" hope to prevail. It is an interim stage between old and new. The conflict blows up in the perpetrators' face and a new theory is invented. Incorporation is the sworn enemy of parsimony because it is politically motivated. It keeps everyone happy by not giving up anything and accumulating entities. This entity hoarding is poisonous and undoes the whole hyper-structure.
3.. Contingency - see (r) above.
4.. Strict monism or pluralism - see (o) above.
5.. Comprehensiveness prevents parsimony. To obtain a description of the world, which complies with a law of parsimony, one has to ignore and neglect many elements, facts and observations. Gödel demonstrated the paradoxality inherent in a comprehensive formal logical system. To fully describe the world, however, one would need an infinite amount of assumptions, axioms, theoretical entities, elements, functions and variables. This is anathema to parsimony.
6.. The previous excludes the reconcilement of parsimony and monovalent correspondence. An isomorphic mapping of the world to the worldview, a realistic rendering of the universe using theoretical entities and other language elements would hardly be expected to be parsimonious. Sticking to facts (without the employ of theory elements) would generate a pluralistic multiplication of entities. Realism is like using a machine language to run a supercomputer. The path of convergence (with the world) - convergence (with predictions yielded by the theory) leads to a proliferation of categories, each one populated by sparse specimen. Species and genera abound. The worldview is marred by too many details, crowded by too many apparently unrelated observations.
7.. Finally, if the field of research is wrongly - too narrowly - defined, this could be detrimental to the positing of meaningful questions and to the expectation of receiving meaningful replies to them (experimental outcomes). This lands us where we started: the psychophysical problem is, perhaps, too narrowly defined. Dominated by Physics, questions are biased or excluded altogether. Perhaps a Fourth Substance IS the parsimonious answer, after all.
It would seem, therefore, that parsimony should rule out the existence of a Necessary and Supreme Being or Intelligence (God). But is Nature really parsimonious, as Poincare believed? Our World is so complex and includes so many redundancies that it seems to abhor parsimony. Doesn't this ubiquitous complexity indicate the existence of a Mind-in-Chief, a Designer-Creator?
VI. Complexity as Proof of Design
"Everything is simpler than you think and at the same time more complex than you imagine."
(Johann Wolfgang von Goethe)
Complexity rises spontaneously in nature through processes such as self-organization. Emergent phenomena are common as are emergent traits, not reducible to basic components, interactions, or properties.
Complexity does not, therefore, imply the existence of a designer or a design. Complexity does not imply the existence of intelligence and sentient beings. On the contrary, complexity usually points towards a natural source and a random origin. Complexity and artificiality are often incompatible.
Artificial designs and objects are found only in unexpected ("unnatural") contexts and environments. Natural objects are totally predictable and expected. Artificial creations are efficient and, therefore, simple and parsimonious. Natural objects and processes are not.
As Seth Shostak notes in his excellent essay, titled "SETI and Intelligent Design", evolution experiments with numerous dead ends before it yields a single adapted biological entity. DNA is far from optimized: it contains inordinate amounts of junk. Our bodies come replete with dysfunctional appendages and redundant organs. Lightning bolts emit energy all over the electromagnetic spectrum. Pulsars and interstellar gas clouds spew radiation over the entire radio spectrum. The energy of the Sun is ubiquitous over the entire optical and thermal range. No intelligent engineer - human or not - would be so wasteful.
Confusing artificiality with complexity is not the only terminological conundrum.
Complexity and simplicity are often, and intuitively, regarded as two extremes of the same continuum, or spectrum. Yet, this may be a simplistic view, indeed.
Simple procedures (codes, programs), in nature as well as in computing, often yield the most complex results. Where does the complexity reside, if not in the simple program that created it? A minimal number of primitive interactions occur in a primordial soup and, presto, life. Was life somehow embedded in the primordial soup all along? Or in the interactions? Or in the combination of substrate and interactions?
Complex processes yield simple products (think about products of thinking such as a newspaper article, or a poem, or manufactured goods such as a sewing thread). What happened to the complexity? Was it somehow reduced, "absorbed, digested, or assimilated"? Is it a general rule that, given sufficient time and resources, the simple can become complex and the complex reduced to the simple? Is it only a matter of computation?
We can resolve these apparent contradictions by closely examining the categories we use.
Perhaps simplicity and complexity are categorical illusions, the outcomes of limitations inherent in our system of symbols (in our language).
We label something "complex" when we use a great number of symbols to describe it. But, surely, the choices we make (regarding the number of symbols we use) teach us nothing about complexity, a real phenomenon!
A straight line can be described with three symbols (A, B, and the distance between them) - or with three billion symbols (a subset of the discrete points which make up the line and their inter-relatedness, their function). But whatever the number of symbols we choose to employ, however complex our level of description, it has nothing to do with the straight line or with its "real world" traits. The straight line is not rendered more (or less) complex or orderly by our choice of level of (meta) description and language elements.
The simple (and ordered) can be regarded as the tip of the complexity iceberg, or as part of a complex, interconnected whole, or hologramically, as encompassing the complex (the same way all particles are contained in all other particles). Still, these models merely reflect choices of descriptive language, with no bearing on reality.
Perhaps complexity and simplicity are not related at all, either quantitatively, or qualitatively. Perhaps complexity is not simply more simplicity. Perhaps there is no organizational principle tying them to one another. Complexity is often an emergent phenomenon, not reducible to simplicity.
The third possibility is that somehow, perhaps through human intervention, complexity yields simplicity and simplicity yields complexity (via pattern identification, the application of rules, classification, and other human pursuits). This dependence on human input would explain the convergence of the behaviors of all complex systems on to a tiny sliver of the state (or phase) space (sort of a mega attractor basin). According to this view, Man is the creator of simplicity and complexity alike but they do have a real and independent existence thereafter (the Copenhagen interpretation of a Quantum Mechanics).
Still, these twin notions of simplicity and complexity give rise to numerous theoretical and philosophical complications.
In human (artificial and intelligent) technology, every thing and every action has a function within a "scheme of things". Goals are set, plans made, designs help to implement the plans.
Not so with life. Living things seem to be prone to disorientated thoughts, or the absorption and processing of absolutely irrelevant and inconsequential data. Moreover, these laboriously accumulated databases vanish instantaneously with death. The organism is akin to a computer which processes data using elaborate software and then turns itself off after 15-80 years, erasing all its work.
Most of us believe that what appears to be meaningless and functionless supports the meaningful and functional and leads to them. The complex and the meaningless (or at least the incomprehensible) always seem to resolve to the simple and the meaningful. Thus, if the complex is meaningless and disordered then order must somehow be connected to meaning and to simplicity (through the principles of organization and interaction).
Moreover, complex systems are inseparable from their environment whose feedback induces their self-organization. Our discrete, observer-observed, approach to the Universe is, thus, deeply inadequate when applied to complex systems. These systems cannot be defined, described, or understood in isolation from their environment. They are one with their surroundings.
Many complex systems display emergent properties. These cannot be predicted even with perfect knowledge about said systems. We can say that the complex systems are creative and intuitive, even when not sentient, or intelligent. Must intuition and creativity be predicated on intelligence, consciousness, or sentience?
Thus, ultimately, complexity touches upon very essential questions of who we, what are we for, how we create, and how we evolve. It is not a simple matter, that...
The fact that the Universe is "fine-tuned" to allow for Life to emerge and evolve does not necessarily imply the existence of a Designer-Creator (although this cannot be ruled out conclusively). All forms and manner of Anthropic Principles are teleological and therefore non-scientific. This, though, does not ipso facto render them invalid or counterfactual.
Still, teleological explanations operate only within a context within which they acquire meaning. God cannot serve as His own context because he cannot be contained in anything and cannot be imperfect or incomplete. But, to have designed the Universe, He must have had a mind and must have used a language. His mind and His language combined can serve as the context within which he had labored to create the cosmos.
The rule of parsimony applies to theories about the World, but not to the World itself. Nature is not parsimonious. On the contrary: it is redundant. Parsimony, therefore, does not rule out the existence of an intelligent Designer-Creator (though it does rule out His incorporation as an element in a scientific theory of the world or in a Theory of Everything).
Finally, complexity is merely a semantic (language) element that does not denote anything in reality. It is therefore meaningless (or at the very least doubtful) to claim the complexity of the Universe implies (let alone proves) the existence of an intelligent (or even non-intelligent) Creator-Designer.