Machine Minds

download Machine Minds

of 79

Transcript of Machine Minds

  • 8/14/2019 Machine Minds

    1/79

    The Emerging Mind of the Machine

    John Cameron Glasgow II

    1

  • 8/14/2019 Machine Minds

    2/79

    ABSTRACT

    A philosophy of emerging systems based on general systems theory concepts is propounded and

    applied to the problem of creating intelligence in a machine. It is proposed that entropy, a

    principle that provides a framework for understanding the organization of physical systems

    throughout the universe, serves also as a basis for understanding the creation of an intelligence.

    That position is defended by the argument that the universe is a statistically described place in

    which the entropic processes of information bearing physical systems (e.g. RNA, DNA, neural

    networks, etc.) describe the same thing under interpretation as information or physical structures.

    As a consequence the observations concerning hierarchical stratification and rule creation

    described in general systems theory and so obvious in physical systems are assumed to apply to

    information systems. The mind and its intelligence is identified as such an information system

    that emerges into the human environment as a system of hierarchically structured mental objects

    and associated rules. It is inferred that those rules/objects that emerge at a level in a mind are not

    reducible to the functions and interactions of constituent parts at lower levels. All of this implies

    that any implementation of humanlike intelligence in a machine will have to include lower levels,

    and that the mechanism of that implementation will have to be emergence rather than

    construction. It is not concluded that this precludes the implementation of intelligence in a

    machine. On the contrary, these ideas are developed in the context of an attempt to implement anundeniably intelligent machine, that is, a machine that will be considered intelligent in the same

    manner that one person normally and implicitly considers another person intelligent.

    2

  • 8/14/2019 Machine Minds

    3/79

    INTRODUCTION

    Emergence is a controversial philosophical viewpoint that maintains that the universe exhibits an

    ongoing creativity. It claims that creativity is manifest in the orderly systems that emerge at a

    level from constituent parts at lower levels. More explicitly, such systems are seen to exhibit

    behavior that is coherent and deterministic but not predictable and not reducible to the laws that

    apply at the lower levels. That is, new levels exhibit new laws. Thus a person who accepts the

    concept of emergence might claim that, while the laws of physics and chemistry can explain the

    internal bonds and the physical attributes of any and all molecules, those laws cannot predict

    which molecules will be produced or the subsequent behavior of a particular molecule or a group

    of molecules, even though the nature of those activities can be codified. As a more concrete

    example, while the structure and chemical proclivities of a molecule of deoxyribonucleic acid(DNA) can be explained by the laws of chemistry and its behavior in the world as it encounters

    other objects can be explained by laws of physics, the spectacular behavior exhibited over time

    by groups of these molecules (as codified in synthetic theories of evolution) is not implicit in

    those laws. As an example at a higher level, while demographers might determine that the world

    human population will reach a particular level in fifty years, they cannot predict the nature of the

    social, governmental, or economic institutions that will arise any more than they can predict that

    some specific set of people will produce offspring who will in turn produce offspring who will

    produce a great leader who will materially affect those institutions (although it is reasonable to

    assert that such will be the case). The concept of a universe in a state of ongoing creative activity

    lies somewhere between the two most popular world models; the idea that the universe is in

    some manner controlled or directed from without and the idea that the universe consists of a set

    of unchanging rules and objects to which all observed phenomena can be reduced.

    The philosophical concept of reductionism maintains that the nature and activity of everything is

    implicit in some basic finite set of rules. According to reductionism, given those rules and an

    initial state of the universe all of the past and future states of the universe can be calculated (an

    assertion made by Pierre Simon Laplace in the late seventeenth century at a time when it seemed

    that the clockwork mechanism of the universe would soon be completely revealed). This

    3

  • 8/14/2019 Machine Minds

    4/79

    philosophy is easily accepted by scientists, for whom it can serve as both the justification of

    continued search for underlying laws, and the assurance that such laws exist. Philosophically,

    reductionism has come under attack from many quarters. The implied existence of a regress of

    explanation of phenomena in terms of their constituent phenomena requires (for any practical

    application of the concept) the hypothesis of some particular set of low level, basic and

    fundamental objects and rules whose validity is assumed. In a macroscopic sense the job of

    science can be taken to be the reduction of phenomena at high levels to some hierarchical

    combination of the basic set, thus revealing why things are as they are. But the selection of a

    basic set of rules and objects has often proven problematical. In practice, an assumed truth often

    turns out to be just a particular case in a larger scheme. In geometry, Euclids parallel lines axiom

    is famous as an assumed truth that is only true in flat or uncurved space. Newtons laws are

    true only in the context of a restricted portion of an Einsteinian universe, albeit the portion of

    which we happen to be most aware. At the quantum level of existence some of the laws of logic

    that seem so evident at the human level of existence, break down. Even some of Einstein's

    assumptions break down at the quantum level (e.g. the assumption that the speed of light is

    fundamental limit). A reductionist would take this to mean that the basis set of fundamental

    truths had not yet been discovered.

    In the early part of this century the school of philosophical thought that exemplified the

    reductionists attitude was logical positivism. The two famous proponents of logical positivismwere Alfred North Whitehead and Bertrand Russell. Their attempt to provide the philosophy

    with a sound analytical basis ( Principia Mathematica) was brought into question in 1931 by the

    famous incompleteness theorem of Kurt Goedel. Goedel proved that in any formal system based

    on a set of axioms and that is sufficiently complex to be able to do arithmetic, either there will be

    assertions that can be made within the framework of the system that are true but that cannot be

    proven true (i.e. the system is incomplete), or there will be contradictory assertions that can be

    proven true (i.e. the system is inconsistent). That is, there are inherent weaknesses in any such

    system that tend to invalidate the system as a whole. Still, scientists can work within a

    reductionist framework if their work is restricted to the phenomena at some particular level or

    some small closely associated group of levels. In the normal conduct of their research they do not

    come up against an infinite regress of explanations. Problems of incompleteness or inconsistency4

  • 8/14/2019 Machine Minds

    5/79

    are rarely encountered. When they are encountered, the necessary assumptions concerning the

    effects of other levels phenomena can usually be revised to remove the specific problem. Indeed,

    the very process of revision of assumptions can be considered an integral part of scientific

    process (see Kuhn 1962). Consequently scientists have not found it necessary to seriously alter

    the methodological framework within which they have worked since the development and

    adoption of that methodology in the eighteenth century. It is natural then, that scientifically

    oriented researchers, when faced with the problem of creating intelligence in a machine feel that

    such intelligence can be achieved by an appropriate basic set of rules and objects installed in the

    proper hierarchically structured control system. Assuming, that reductionism holds across the

    range of levels required for the implementation of a mind, and that humans have the perception,

    tools, and time to determine the content of those levels together with the skill to emulate and

    implement those contents in a machine, then certainly machines can be built to be intelligent.

    Reductionism is unpopular with philosophers defending concepts of ethics and morality as

    formulated in Judeo-Christian religions. Most of the doctrines of western religious institutions as

    they apply to ethical and moral considerations turn upon the human intuition of free choice. If

    the universe is predetermined as reductionism implies, then free choice is an illusion. In

    particular a choice to commit or not commit an immoral act, no matter how real the choice seems

    to the chooser, is a charade. Further, any supposed imposition of reward or punishment for such

    behavior becomes a compounded charade. The inclusion of a Deity in a reductionist framework results in a religious concept called Deism. It maintains that a Deity created and set the universe

    in motion after which the universe did not require (and does not require) any further

    intervention. Deism enjoyed a brief popularity among intellectuals about the time of the

    American revolution, a time at which the consequences of the Newtonian revolution dominated

    most of the thought of the times. But such a belief does not provide a sufficient base upon which

    to build the requisite ethical and moral dogma of a western style religion. Persons worried about a

    philosophical basis for ethical and moral behavior, or those who insist simply that the decisions

    they make are meaningful in that the future course of events is not fixed and does respond to

    those decisions, normally prefer some concept of a supervised, managed, or directed universe

    (well use the term spiritualism to refer to any such a concept of a directed universe).

    5

  • 8/14/2019 Machine Minds

    6/79

    Spiritualists will almost certainly disagree with the reductionist conclusion concerning machine

    intelligence. Inherent in spiritualism is the belief in the existence of a non-corporeal motivating

    and controlling force or forces that in some (inexplicable) manner interact with corporeal objects

    causing behavior that does not yield to reductionist explanation. This force or forces must be

    an eternal feature of the universe for if man is the result of a lengthy metamorphosis from simple

    molecular structures to his present state (a theory that has achieved the status of accepted fact

    for all but a tiny minority of people) then the motivating force must have been there all along,

    directing, maintaining, and imbuing each and every object or process with its unique non-

    reducible characteristics. To maintain otherwise requires either that evolution be rejected in favor

    of an instantaneous creation replete with all of the evidence that argues against it, or that it be

    accepted that unenlightened matter in some way can give rise to, or create, inexplicable forces

    that subsequently vitalize it. Thus an eternal and all pervasive life force or elan vital is

    hypothesized to account for the obvious but not easily described difference between living and

    non-living matter, and the human essence, mind, or soul, is considered an entity, separate from,

    but associated with the body, yet not subject to the bodys limitations. The believer is free to

    assert that the forces are responsible for any perceived phenomenon or state of nature. If the

    human mind is perceived as a unique, non-reducible phenomenon, then that is the intent of the

    forces of which we can have no knowledge. Certainly we cannot reproduce in a machine that of

    which we can have no knowledge.

    Phenomena for which no reductionist explanations exist serve as the only support for belief in

    spiritualism (e.g. inexplicable human mental activity such as intuition, a flash of insight, or

    creativity may be cited as evidence the non-reducible quality of the human mind). In fact the term

    belief applies only to assertions about such phenomena. At the other end of the spectrum the

    term fact applies only to those phenomena that are reducible (we use the term in its sense as an

    objective measure of reality rather than its use in referring to data or in making assertions of

    truth). The axioms to which facts may be reduced might be as informal as the (reproducible)

    evidence of the senses or as formal as an abstract mathematical system. While adherents of facts

    and beliefs both claim to have a corner on the market in truth , both are subject to challenge.

    Spiritualists are stuck with the problem of trying to support their beliefs by citing inexplicable

    phenomena that might at any moment be explained, while reductionists must fear (or they should6

  • 8/14/2019 Machine Minds

    7/79

    fear) that their systems might yield contradictory facts, or that facts will be discovered that are

    not explainable in their system. It might appear that there are no alternatives to these two

    opposed ways of conceptualizing the world. Such is not the case. The concept of emergence

    provides another 1 explanation of the way things are; one that allows for the existence of non-

    reducible phenomena , and yet one that is not spiritualistic . We stress that concepts

    associated with theories of emergence should not be confused with spiritualism. Certainly

    spiritual forces could be invoked to explain creative processes but that would simply be another

    version of a directed universe. Many real, non-spiritual mechanisms can be proposed to explain

    the process of ongoing creation (and some are detailed below). Nor are the concepts regarding

    emergence, philosophically compatible with reductionism. Theories concerned with emerging

    systems hypothesize a hierarchically structured universe in which the levels of the hierarchy

    exhibit a semi-permeability with respect to the effects of the rules and objects at one level onadjacent levels. The further removed two levels the less the phenomena in the one can be reduced

    to the phenomena in the other. Since the levels are assumed to extend throughout the universe,

    laws can be at once, both universal and restricted in scope. In effect the laws of the universe

    never die, they just fade away (across levels). This concept has many implications for what it

    means to exist and be intelligent at one level in such a universe. In the context of this paper it has

    implications for the implementation of machine intelligence. To sum up, reductionism easily

    admits of the possibility of building a machine intelligence while spiritualism argues against it.

    Emergence lies somewhere between these two extremes. As might be expected then, it implies

    that the creation of a machine intelligence lies somewhere between impossible and possible. First

    we develop support for the concept of emerging systems by considering the concept of a

    universe organized into hierarchical levels through the interaction of two seemingly opposed

    facts of nature, chaos and order. We follow that by stating the ideas more clearly as a hypothesis.

    Finally we make inferences from the hypothesis concerning the possibility of creating a machine

    intelligence (which we deem possible) and the conditions necessary to accomplish that feat in the

    context of an emergent universe.

    1 But not necessarily the only other one.

    7

  • 8/14/2019 Machine Minds

    8/79

    EMERGENT SYSTEMS

    Evolution

    The philosophicalreligious climate of Europe in the eighteenth and nineteenth centuries was

    favorable for investigations into the nature of biological organisms. This was due to two aspects

    of the JudeoChristian thought of that time: time was considered linear and events, measured

    from creation to eternity with (for the Christians) memorable deterministic events occurring at

    the advent and second coming of Christ. This was different from the cyclical nature of time linear

    and cyclical of the Greek philosophers (and of most other cultures).

    It was assumed that God had created the world in a methodical ordered manner. One perceptionof the order of things was the "scale of being," the "ladder of perfection" or the "ladder of life."

    Man was perceived to occupy the highest earthly rung of the ladder with all of the other various

    life forms occupying lower positions depending upon their perfection or complexity. The ladder

    did not stop at man but continued on with Angels, Saints and other heavenly creatures that

    occupied successively higher rungs until finally God was reached. Man thus had a dual nature; a

    spiritual nature that he shared with the Angels above him on the ladder and an animal nature that

    he shared with the creatures below him on the ladder.

    Scientific thought and religious thought often mixed during the renaissance and even into the

    nineteenth century. Investigations into biological complexity were encouraged so that the order of

    succession of creatures on the ladder of life could be more accurately ascertained. That, coupled

    with the perception that time moved inexorably from the past into the future set the stage for the

    discovery of evolutionary processes. All that was needed was the concept of vast periods of

    geologic 2 time, the idea that the ladder of life might have a dynamic character, and a non

    supernatural mechanism by which movement on it could occur. These ideas, or various forms of

    them, had already gained widespread acknowledgement when Charles Darwin wrote his Origin of

    Species (published in 1859). Origin united the idea of the transmutation of species over geologic

    2 The idea of geologic time was provided by James Hutton, the father of geology.

    8

  • 8/14/2019 Machine Minds

    9/79

    time with the mechanism of natural selection 3 to yield the theory of biological evolution. The

    theory required strong champions for it made formal and respectable the heretical ideas that had

    only been the object of speculation. Even in 1866 (seven years after publication of Origin of

    Species ) the ladder was still being defended (Eiseley , 1958). From a scientific point of view these

    were the last serious challenges to evolution. The evidence from biology, geology, and

    paleontology were too strong to support that particular form of theologically generated order. At

    about the same time (1866) an Austrian monk named Gregor Mendel published a report on his

    research into genes, the units of heredity. In it he outlined the rules that govern the passing of

    biological form and attribute from one generation to the next. The report was to lay

    unappreciated for the next thirty years until around the turn of the century when the rules were

    independently rediscovered. The Mendelian theory provided the Darwinian mechanism of natural

    selection with an explanation for the necessary diversification and mutation upon which it relied.

    The mendelian laws were incorporated into Darwinism which became known as NeoDarwinism.

    In 1953 J. D. Watson and F. H. C. Crick proposed the double helix molecular structure of the

    deoxyribonucleic acid (DNA)) from which genes are constructed and which contain the actual

    coded information needed to produce an organism. This and various loosely related theories such

    as population genetics, speciation, systematics, paleontology and developmental genetics

    complement neodarwinism and combine with it to form what is termed the synthetic theory of

    evolution. The synthetic theory of evolution represents the state of the theory today (Salthe

    1985).

    This new wave of scientific discovery about the earth and the particularly the biology of the

    earth worked a new revolution in philosophical thought in which man had to be perceived as

    playing a smaller role in the scheme of things if for no other reason than the true grandeur of the

    scheme of things was becoming apparent. The first philosopher to embrace the new biological

    theories was Herbert Spencer. He was a philosopher of the latter half of the nineteenth century

    who seized upon the theory of evolution as presented by Darwin and generalized it to a

    philosophical principle that applied to everything, as Will Durant (Durant 1961) demonstrates

    3 The result of the uneven reproduction of the genotypephenotype in a group of members of a species thatcan mate with each other. The mechanism of selection is that of the interaction of the phenotype with itsenvironment. The concept was popularized as 'the survival of the fittest' by Herbert Spencer.

    9

  • 8/14/2019 Machine Minds

    10/79

    by partial enumeration, to:

    "The growth of the planets out of the nebulae; the formation of oceans and mountains on the earth; the metabolismof elements by plants, and of animal tissues by men; the development of the heart in the embryo, and the fusion of the bones after birth; the unification of sensations and memories into knowledge and thought, and of knowledge

    into science and philosophy; the development of families into clans and gentes and cities and states and alliancesand the 'federation of the world'."

    the last, of course, being a prediction. In everything Spencer saw the differentiation of the

    homogeneous into parts and the integration of those parts into wholes. The process from

    simplicity to complexity or evolution is balanced by the opposite process from complexity to

    simplicity or dissolution. He attempted to show that this followed from mechanical principles by

    hypothesizing a basic instability of the homogeneous, the similar parts being driven by external

    forces to separated areas in which the different environments produce different results.

    Equilibration follows; the parts form alliances in a differentiated whole. But all things run down

    and equilibration turns into dissolution. Inevitably, the fate of everything is dissolution. This

    philosophy was rather gloomy but it was in accord with the second law of thermodynamics 4 or

    the law of entropy that states that all natural processes run down. Entropy was proposed by

    Rudolf Clausius and William Thompson in the 1850's and was derived from the evidence of

    experiments with mechanical systems, in particular heat engines (see the section on order and

    entropy below). That scientific fact tended to support a philosophy that the universe and

    everything in it was bound for inevitable dissolution. This was shocking and depressing to philosophers and to the educated populace at the turn of the twentieth century. However, the

    universal application of the Darwinian theory of evolution is not so simple as Spencer would

    have it.

    In response to the overwhelming wealth of detail, the biological sciences began to turn inward.

    The detail allowed for tremendous diversification without leaving the confines of the life sciences.

    The subject matter of biology was, and is, perceived as so different from the subject matter of

    4 The first law of thermodynamics deals with the conservation of energy, stating that the sum of the flow of heat and the rate of change of work done, are equal to the rate of change of energy in a system. There is azero'th law (that, as you may suspect, was proposed after the first and second laws, but that logically precedesthem). It states that two objects that are in thermal equilibrium with a third object, will be in thermalequilibrium with each other.

    10

  • 8/14/2019 Machine Minds

    11/79

    other areas of science that it is considered unique and to some extent a closed discipline. But the

    isolationist tendency is widespread in all of the sciences and, as has been shown on many

    occasions, uniqueness is an illusion. In recent years there has been a move on the part of some

    biologists to find a more general framework within which biology fits together with the other

    disciplines of science. Such a move toward integration should, of course, be of interest to all of

    the sciences. One such approach is espoused by Stanley N. Salthe 5 .

    Salthe notes (Salthe 1985) that there are at least seven levels of biological activity that must be

    considered for the investigation of what are considered common biological processes. These

    include, but are not limited to, the molecular, cellular, organismic, population, local ecosystem,

    biotic regional, and the surface of the earth levels. The systems at these levels are usually studied

    as autonomous systems largely because of the difficulty in attributing cause to other levels. There

    is, inherent in biological phenomena, a hierarchical structure differentiated mainly by spacial and

    temporal scaling. The hierarchy is not limited to biological systems, but extends upward to the

    cosmos and downward to quantum particles. Nor is it limited to any particular enumeration such

    as the seven levels mentioned above. Interactions between levels are largely constrained to those

    that occur between adjacent levels. At a particular level (focal level) and for a particular entity the

    surrounding environment (or next higher level) and the material substance of the entity (next

    lower level) are seen as providing initial conditions and constraints for the activities or evolution

    of the entity. Salthe proposes that a triadic system of three levels (the focal level and theimmediately adjacent levels) are the appropriate context in which to investigate and describe

    systemic, in particular, biologic phenomena. So the theory of evolution that depends upon

    natural selection applies at the population level and may or may not apply at other levels. Salthe

    distinguishes several factors that provide differentiation into levels. These include 1) scale in both

    size and time that prohibits any dynamic interaction between entities at the different levels, 2)

    polarity, that causes the phenomena at higher and lower levels to be radically different in a

    5 The specific approach presented by Salthe has its roots in general system theory. That theory's inception wasin Norbert Wiener's cybernetics. It was given its strong foundations and extended far beyond the originalcontent by such researchers as Ludwig Von Bertalanffy (1968), Ervin Laszlo (1972), Ilya Prigogine (1980),Erich Jantsch (1980), David Layzer (1990) and many others (the cited works are included in the bibliography,and are not intended to represent the principle work of the cited researcher). General systems science is now afirmly established discipline. It provides the context in which the discussion concerning 'levels' on theensuing pages should be considered.

    11

  • 8/14/2019 Machine Minds

    12/79

  • 8/14/2019 Machine Minds

    13/79

    homogeneous classes. There is then a 'secondary' type of order or regularity which arises only through the(usually cumulative) effect of individualities in inhomogeneous systems and classes. Note that such order neednot violate the laws of physics"

    Salthe and others who attempt to distinguish levels as more than mere handy abstractions to

    locate or specify objects provides a means by which mechanists and vitalists have some commonground. The processes and objects that occupy a level are not expected to be completely

    explainable in terms of any other level, in fact any complete explanation is precluded by the

    increasing impermeability that characterizes the correspondence between increasingly distant

    levels. Only the most immediate levels have any appreciable impact on the activities of the other.

    This saves the mechanists from the infinite regress of explanations but leaves the explanation of

    elan vital for further investigation. As Elsasser points out, these need not be mystic,

    spiritualistic, or nonscientific. One such possible explanation for the observed propensity of

    systems to selforganize is given below in the section on entropy and order. It is an unexpected

    explanation, advocating that organization results not because of any positive motivating force

    but rather because of the inability of some open systems to pursue their entropic imperative at a

    rate commensurate with the expanding horizons of the local state space. It will be argued that the

    nature of the system at a level in an environment together with existence of chaotic tendencies,

    dictates the form that the resulting organization takes. These ideas have applicability to virtually

    any system at any level, not just biological systems at the levels identified above. We shall

    comment and expand upon these ideas in later sections, but first we will review the developmentof knowledge about the nature of the physical world beyond (and below) the lowest levels

    mentioned above. We skip over the molecular level, whose rules and objects are detailed in

    chemistry, the atomic level described by atomic physics, and the nuclear level described by

    nuclear physics, to the quantum level whose rules and objects are expressed in particle physics

    by quantum mechanics. We do this because the quantum level vividly illustrates the increasing

    inability to relate the actions and objects at different levels far removed from the human level of

    existence.

    Quantum Theory

    At the end of the nineteenth century physicists were perplexed by the fact that objects heated to

    13

  • 8/14/2019 Machine Minds

    14/79

    the glowing point gave off light of various colors (the spectra). Calculations based on the then

    current model of the atom indicated that blue light should always be emitted from intensely hot

    objects. To solve the puzzle, the German scientist, Max Planck proposed that light was emitted

    from atoms in discrete quanta or packets of energy according to the formula, E = hf, where f was

    the frequency of the observed light in hertz and h was a constant unit of action (of dimensions

    energy/frequency). Planck calculated h = 6.63 x 10 27 ergs/hertz and the value became known as

    Planck's constant. In 1905 Albert Einstein suggested that radiation might have a corpuscular

    nature. Noting that mc 2 = E = hf, or more generally mv 2 = hf (where v stands for any velocity)

    provides a relation between mass and frequency he suggested that waves should have particle

    characteristics. In particular since the wavelength l, of any wave, is related to the frequency by f

    = v/l, then l = h/mv, or setting the momentum mv = p, then p = h/l. That is the momentum of a

    wave is Planck's constant divided by the wavelength. Years later (in 1923), Prince Louis

    DeBroglie, a French scientist, noted that it should be true that the inverse relation also exists and

    that l = h/p. That is, particles have the characteristics of waves. This hypothesis was quickly

    verified by observing the interference patterns formed when an electron beam from an electron

    gun was projected through a diffraction grating onto a photosensitive screen. An interference

    pattern forms from which the wavelength of the electrons can be calculated and the individual

    electron impact points can be observed. This, in itself, is not proof of the wave nature of

    particles since waves of particles, (for instance sound waves or water waves) when projected

    through a grating will produce the same effect. However, even when the electrons are propagated

    at widely separated intervals (say one a day), the same pattern is observed. This can only occur

    if the particles themselves and not merely their collective interactions have the characteristics of

    waves. In other words, all of the objects of physics (hence all of the objects in the universe), have

    a waveparticle dual nature. This idea presented something of a problem to the physicists of the

    early part of this century. It can be seen that the wave nature of objects on the human scale of

    existence can be ignored (the wavelength of such objects will generally be less than 10 27 meters

    so, for instance, if you send a stream of billiard balls through a (appropriately large) grating the

    interference pattern will be too small to measure, or at least, small enough to safely ignore).

    Classical physics, for most practical purposes, remained valid, but a new physics was necessary

    14

  • 8/14/2019 Machine Minds

    15/79

    for investigating phenomena on a small scale. That physics became known as the particle 6

    physics.

    The Austrian physicist, Erwin Schr_dinger applied the wave mechanics developed by James

    Clerk Maxwell to develop an appropriate wave equation (y) for calculating the probability of the

    occurrence of the attributes of quantum particles. For example, solving y for a specified position

    of a particle (say the position of an electron near an atom) yields a value which, when squared,

    gives the probability of finding the particle at that position. Then, to investigate the trajectory of

    an electron in relation to its atom, the equation can be solved for a multitude of positions. The

    locus of points of highest probability can be thought of as a most likely trajectory for the

    electron. There are many attributes that a quantum particle might have, among which are mass,

    position, charge, momentum, spin, and polarization. The probability of the occurrence of

    particular values for all of these attributes can be calculated from Shr_dinger's wave function.

    Some of these attributes (for example mass and charge), are considered static and may be thought

    of as fixed and always belonging to the particle. They provide, in effect, immutable evidence for

    the existence of the particle. But other attributes of a particle are complementary in that the

    determination of one affects the ability to measure the other. These other attributes are termed

    dynamic attributes and come in pairs. Foremost among the dynamic complementary attributes

    are position, q, and momentum, p. In 1927 Werner Heisenberg, introduced his famous

    uncertainty principle in which he asserted that the product of the uncertainty with which the position is measured, Dq, and the uncertainty with which the momentum is measured, D p must

    always be greater than Planck's constant. That is, Dq D p h. In other words, if you measure one

    attribute to great precision (e.g. Dq is very small) then the complementary attribute must

    necessarily have a very large uncertainty (e.g. D p h/Dq). A similar relation is true of any pair of

    6 All things at the quantum level (even forces) are represented by particles. Some of these particles arefamiliar to most people (e.g. the electron and photon), others are less familiar (muons, gluons, neutrinos,etc.). Basically, quantum particles consist of Fermions and Bosons. The most important Fermions are quarksand leptons. Quarks are the components of protons and neutrons, while leptons include electrons. TheFermions then, contain all the particles necessary to make atoms. The Bosons carry the forces between theFermions (and their constructs). For example, the electromagnetic force is intermediated by photons, and thestrong and weak nuclear forces are intermediated by mesons, and intermediate vector bosons respectively.Quantum physicists have found it convenient to speak of all of the phenomena at the quantum level in termsof particles. This helps avoid confusion that might arise because of the wave/particle duality of quantum

    phenomena. It gives nonphysicists a comfortable but false image of a quantum level populated by tiny billiard balls.

    15

  • 8/14/2019 Machine Minds

    16/79

    dynamic attributes. In the early days of quantum mechanics it was believed that this fact could be

    attributed to the disturbance of the attribute not being measured by the measuring process. This

    view was necessitated by the desire to view quantum particles as objects that conform to the

    normal human concept of objects with a real existence and tangible attributes (call such a concept

    normal reality ). Obviously, if Heisenberg's principle was true and the quantum world possessed

    a normal reality, the determination of a particle's position must have altered the particle's

    momentum in a direct manner. Unfortunately, as will be discussed below, the assumption of

    normal reality at quantum levels leads to the necessity of assuming faster than light

    communication among the parts of the measured particle and the measuring device. The idea that

    quantum objects conform to a normal reality is now out of favor with physicists (though not

    discarded). In any case, the principle was not intended as recognition of the clumsiness of

    measuring devices. Heisenberg's uncertainty principle arose from considerations concerning

    the wave nature of the Schr_dinger equations and result directly from them. Richard P. Feynman,

    in one of his famous lectures on quantum electrodynamics (QED) tried to put the principle in its

    historical perspective (Feynman, 1985):

    When the revolutionary ideas of quantum physics were first coming out, people still tried to understand themin terms of old-fashioned ideas (such as, light goes in straight lines). But at a certain point the old-fashionedideas would begin to fail, so a warning was devised that said, in effect, Your old-fashioned ideas are nodamned good when... If you get rid of the old-fashioned ideas and instead use the ideas Im explaining inthese lectures [ ] there is no need for an uncertainty principle.

    The old-fashioned ideas were of course, the normal physics of normal reality, and as Feynman

    readily admitted, the new ideas appeared non-sensical in that reality.

    But the predictions made by quantum mechanics (which now consist of four different systems

    equivalent to Schr_dinger equations) 7 of the results of quantum particle experiments have proven

    accurate for fifty years. They consistently and precisely predict the observations made by

    physicists of the effects produced by the interactions of single particles (from perhaps cosmic

    sources) and beams of particles produced in laboratories. The tools for the prediction of 7 Three other systems equivalent to the Schr_dinger equations are Heisenberg's matrix mechanics, Dirac'stransformation theory and Feynman's sum over histories system.

    16

  • 8/14/2019 Machine Minds

    17/79

    experimental results are perfected and used daily by physicists. So, under the test that a

    successful theory can be measured by its ability to predict, quantum theory is very successful. In

    spite of all of their accuracy and predictive ability the systems do not provide a picture of, or

    feeling for the nature of quantum reality any more than the equation F = ma provides a picture of

    normal reality. But the quantum mechanics does provide enough information to indicate that

    quantum reality cannot be like normal reality. Unfortunately (or fortunately if you are a

    philosopher) there are many different versions of quantum reality that are consistent with the

    facts; all of which are difficult to imagine. For instance, the concept of attribute, so strong and

    unequivocal in normal reality, is qualified at the quantum level as indicated above. Worse, if the

    quantum mechanics equations are to be interpreted as a model of the quantum level of existence

    then the inhabitants of quantum reality consist of waves of probability, not objects with

    attributes (strange and quirky though they may be). As Werner Heisenberg said "Atoms are not

    things." The best one can do is to conjure up some version of an Alice in Wonderland place in

    which exist entities identifiable by their static attributes but in which all values of dynamic

    attributes are possible but in which none actually exist. Occasionally the quantum peace is

    disturbed by an act that assigns an attribute and forces a value assignment (e.g. a measurement or

    observation occurs). Nick Herbert in his book Quantum Reality (Herbert, 1989) has identified

    eight versions of quantum reality held by various physicists (the versions are not all mutually

    exclusive):

    1) The Copenhagen interpretation (and the reigning favorite) originated by Niels Bohr, Heisenberg, and Max Bornholds that there is no underlying reality. Quantum entities possess no dynamic attributes. Attributes arise as a joint

    product of the probability wave and the normal reality measuring device.

    2) The Copenhagen interpretation part II maintains that the world is created by an act of observation made in thenormal world. This has the effect of choosing an attribute and forcing the assignment of values.

    3) The world is a connected whole arising from the history of phase entanglements of the quantum particles thatmake up the world. That is, when any two possibility waves (representative of some quantum particle) interact, their

    phases become entangled. Forever afterward, no matter their separation in space, whatever actuality may manifestitself in the particles, the particles share a common part. Originally the phase entanglement was thought of as just amathematical fiction required by the form of the equations. Recent developments (Bell's theorem, see below) lend

    credence to the actuality of the entanglements in the sense that those entanglements can explain, and in fact areneeded to explain, experimental data. Phase entanglement was originally proposed by Erwin Schr_dinger.

    4) There is no normal reality hypothesis (normal reality is a function of the mind). John Von Neumann felt that the problem with explaining the reality behind the experimental observation arose because the measurement instrumentswere treated as normally real while the particles being measured were considered as waveforms in a quantum reality.He undertook to treat the measurement devices as quantum waveforms too. The problem then became one of

    17

  • 8/14/2019 Machine Minds

    18/79

    determining when a quantum particle as represented by a wave of all possible attributes and values, collapsed (took a quantum leap) to a particular set of attributes and values. He examined all of the possible steps along the path of the process of measurement and determined that there was no distinguished point on that path and that thewaveform could collapse anywhere without violating the observed data. Since there was no compelling place toassume the measurement took place, Von Neumann placed it in the one place along the path that remainedsomewhat mysterious, the human consciousness.

    5) There are an infinity of worlds hypothesis. In 1957, Hugh Everett, then a Princeton University graduate studentmade the startling proposal that the wave function does not collapse to one possibility but that it collapses to all

    possibilities. That is, upon a moment requiring the assignment of a value, every conceivable value is assigned, onefor each of a multitude of universes that split off at that point. We observe a wave function collapse only because weare stuck in one branch of the split. Strange as it may seem this hypothesis explains the experimental observations.

    6) Logic at the quantum level is different than in normal reality. In particular the distributive laws in logic do notapply to quantum level entities, that is, A or (B and C) is not the same as (A or B) and (A or C). For example, if club Swell admits people who are rich or it admits people who come from a good family and have goodconnections, while club Upper Crust admits people who are rich or come from a good family and who are rich or have good connections, then in normal reality we would find that after one thousand people apply for membershipto both clubs the same group of people will have been accepted at both clubs. In the quantum world, not only willthe memberships be different but in club Upper Crust there will be members who are not rich and do not come froma good family or are not well connected.

    7) Neo-realism (the quantum world is populated with normal objects). Albert Einstein said that he did not believethat God played dice with the universe. This response was prompted by his distaste for the idea that there was noquantum reality expressible in terms of the normal reality. He didn't want to except probability waves as in somesense real. He and DeBroglie argued that atoms are indeed things and that the probabilistic nature of the quantummechanics is just the characteristic of ensembles of states of systems as commonly presented in statisticalmechanics. John Von Neumann proved a theorem that stated that objects that displayed reasonable characteristicscould not possibly explain the quantum facts as revealed by experimental data. This effectively destroyed the neo-realist argument until it was rescued by David Bohm who developed a quantum reality populated by normal objectsconsistent with the facts of quantum mechanics. The loophole that David Bohm found that allowed him to develophis model was the assumption by Von Neumann of reasonable behavior by quantum entities. To Von Neumann,reasonableness did not include faster than light phenomena. In order to explain the experimental data in a quantumreality populated by normal objects, Bohm found it necessary to hypothesize a pilot wave associated with each

    particle that was connected to distant objects and that was able to receive superluminal communications. The pilotwave was then able to guide the particle to the correct disposition to explain the experimental data. Unfortunatelyfor the neoreality argument faster than light communications puts physicists off. As discussed below, theacceptance of a variation of that concept, at least for objects in the quantum world, may be required by recentfindings.

    8) Werner Heisenberg champions a combination of 1 and 2 above. He sees the quantum world as populated by potentia or "tendencies for being." Measurement is the promotion of potentia to real status. The universe is observer created (which is not the same as Von Neumann's consciousness created universe).

    Albert Einstein did not like the concept of a particle consisting of a wave of probabilities. In

    1935, in collaboration with Boris Podowski and Nathan Rosen he proposed a thought experiment

    that would show that, even if quantum mechanics could not be proven wrong, it was anincomplete theory. The idea was to create a hypothetical situation in which it would have to be

    concluded that there existed quantum attributes/features that were not predictable by quantum

    mechanics. The thought experiment is known as the EPR experiment.

    18

  • 8/14/2019 Machine Minds

    19/79

    The EPR experiment consists of the emission in opposite directions from some source of two

    momentum correlated quantum particles. Correlated particles (correlated in all attributes, not just

    by the momentum attribute) may be produced, for example, by the simultaneous emission of two

    particles from a given energy level of an atom. Call the paths that the first and second particles

    take the right hand and left hand paths respectively. At some point along the right hand path

    there is a measuring device that can measure the momentum of the first particle. On the left hand

    path there is an identical measuring device at a point twice the distance from the source than the

    first device. When the particles are simultaneously emitted along their respective paths, according

    to quantum theory, they both consist of a wave of probability that will not be converted into an

    actuality until they are measured. At the point of being measured their probability wave is

    collapsed, or the momentum attribute is forced to take a value, or the consciousness of the

    observer assigns a momentum value (e.g. a human is monitoring the measuring device), or the

    universe splits, or some other such event takes place to fix the momentum. Now consider the

    events of the experiment. The particles are emitted at the same time in opposite directions. Soon

    the first particle (on the right hand path) is measured and its momentum is fixed to a particular

    value. What is the status of the second particle at this point in time? According to quantum

    mechanics it is still a wave of probability that won't become actualized until it encounters the

    measuring device (still off in the distance). And yet it is known that the left hand measuring

    device will produce the same momentum that the right hand device has already produced andwhen the second particle finally gets to that measuring device it does show the expected

    momentum. Two possibilities present themselves, either the results of the measurement of the

    first particle is somehow communicated to the second particle in time for it to assign the correct

    value to its momentum attribute, or the second particle already 'knows' in some sense, which

    actual momentum to exhibit when it gets to the measuring device. The particles are moving at or

    near the speed of light so the former possibility requires superluminal communication and must

    be rejected. But then the quantum particle must contain knowledge that is not accounted for by

    quantum theory. In fact it must contain a whole list of values to assign to attributes upon

    measurement because it can't possibly know which attribute will be measured by the measuring

    device it encounters. So, Einstein concludes, quantum theory is incomplete since it says nothing

    about such a list.19

  • 8/14/2019 Machine Minds

    20/79

    Einsteins rejection of superluminal communication is equivalent to an assumption of locality.

    That is, any communication that takes place must happen through a chain of mediated

    connections and may not exceed the speed of light. Electric, magnetic and gravitational fields may

    provide such mediation but the communication must occur through distortions of those fields

    that can only proceed at subluminal speeds. In 1964 John Stewart Bell, an Irish Physicist

    attached to CERN, devised a means of using a real version of the EPR thought experiment to test

    the assumption of locality. He substituted two beams of polarity correlated (but randomly

    polarized) photons for the momentum correlated particles of the EPR experiment.

    By 1972 John Clauser at University of California at Berkeley, using a mercury source for the

    beams obtained results indicating that the quantum mechanics predictions were correct. In 1982

    Alain Aspect of the University of Paris performed a more rigorous version of the experiment.

    The validity of quantum theory was upheld and the assumption of locality in quantum events

    shown to be wrong. That quantum reality must be nonlocal is known as Bell's theorem. It is a

    startling theorem that says that unmediated and superluminal or instantaneous communication

    occurs between quantum entities no matter their separation in space. It is so startling that many

    physicists do not accept it. It does little to change the eight pictures of quantum reality given

    above except to extend the scope of what might be considered a measurement situation from the

    normal reality description of components to include virtually everything in the universe. Theconcept of phase entanglement is promoted from mathematical device to active feature of

    quantum reality.

    Of course, we see none of this in our comfortable, normal reality. Objects are still constrained to

    travel below the speed of light. Normal objects communicate at subluminal rates. Attributes are

    infinitely measurable and their permanence is assured even though they may go for long periods

    of time unobserved and uncontemplated. We can count on our logic for solving our problems, it

    will always give the same correct answer, completely in keeping with the context of our normal

    reality. But we are constructed from quantum stuff. How can we not share the attributes of that

    level of existence? And we are one of the constituents of the Milky Way, (although we never

    think of ourselves as such); what role do we play in the nature of its existence? Does such a20

  • 8/14/2019 Machine Minds

    21/79

    question even make sense? It might help to look more closely at the nature of levels of existence.

    Levels

    Biologists, physicists, astronomers, and other scientists, when describing their work, often

    qualify their statements by specifying the scale at which their work occurs. Thus it's common to

    hear the 'quantum level', the 'atomic level', the 'molecular level', the 'cellular level', the 'planetary

    level' or the 'galactic level' mentioned in their work. And they might further qualify things by

    mentioning the time frame in which the events of their discipline occur, e.g. picoseconds,

    microseconds, seconds, days, years, millions of years, or billions of years.

    The fact that such specifications must be made is not often noted; it's just taken as the nature of

    things that, for example, objects and rules at the quantum level don't have much to do with the

    objects and rules at the cellular level, and the objects and rules at the cellular level don't have

    much to do with objects and rules at the galactic level and so on. A bit of reflection however,

    reveals that it's really quite an extraordinary fact! All of these people are studying exactly the

    same universe, and yet the objects and rules being studied are so different at the different levels,

    that an ignorant observer would guess that they belonged to different universes. Further, when

    the level at which humankind exists is placed on a scale representing the levels that are easily

    identified, it is very nearly in the center (see figure 1).

    Three reasons why that might be, come immediately to mind: 1) we're in the center by pure

    chance, 2) the universe was created around us (and therefore, probably for us), or 3) it only looks

    like we're in the center because we can 'see' about the same distance in both directions. Is our

    perception of these levels, embedded in the scale of being, just an artifact of our failure to see the

    underlying scheme (i.e. the universe can be explained by a few immutable object types together

    with their rules of interaction, so that the different objects and rules that we perceive at different

    levels are simply a complex combination of those primitives), or were these levels set up, in

    some way for our support, or do these levels (and their objects and rules) have a real and, in

    some sense, independent existence? The answer we give is part of the hypothesis that will be

    made below: levels exist and contain rules not derived from surrounding levels of existence, and 21

  • 8/14/2019 Machine Minds

    22/79

    the further removed any two levels the less the one can be interpreted or explained in terms of the

    other .

    If accepted, this answers the question of why we seem to be at the center of the scale of being;

    the further removed a level from the human level of existence, the less that level has in common

    with the human level and the less humans are able to interpret what occurs at that level in familiar

    terms. For sufficiently remote levels, no interpretation can be made. We cannot 'see' beyond what

    our minds can comprehend. So, for example, we don't understand quantum reality because it is on

    the very edge of that to which we as humans can relate. For the same reason, at the other end of

    the scale, we can't conceive of all that is as the universe, and at the same time, the universe as all

    that is, nor do our minds comfortably deal with infinity and eternity. If we were a quantum

    entity, we

    22

  • 8/14/2019 Machine Minds

    23/79

    radii of proton or neutron

    radii of electron shells of atomsmoleculesmacro-molecules(DNA etc.)virusesbacteriablood corpusclesvegetable and animal cells

    atomic Nuclei

    manmousewhale

    snowflakes

    small town

    radius of the Earth

    radius of Solar System

    large city

    radius of galaxy (Milky Way)

    radius of the Sun

    radius of largest structures (clusters of galaxies)

    range of strong nuclear force

    range of weak nuclear force

    distance to nearest star

    10-25

    25

    -20

    -15

    -10

    - 5

    0

    5

    10

    15

    20

    10

    Meters

    Meters

    Figure 1: New scale of being

    would undoubtedly express the same wonder at the impossible nature of the strange subluminal

    world populated by unconnected, virtually uncommunicative, objects with frozen attributes, that

    exist at the very edge of the ability to comprehend. But we shall pull back from such tempting

    speculations. Our purpose is not to investigate the concept of levels in general. We will leave

    many questions unanswered. We are mainly interested in the fact of the existence of levels and

    the effect of that on what we call intelligence. We take up that problem in subsequent sections of

    this paper. It would be good, however, to have a firmer grasp of how and why levels occur.

    23

  • 8/14/2019 Machine Minds

    24/79

    Why should the universe be hierarchically structured and what process or processes leads to the

    semi-permeability of objects and rules across levels?

    Chaos

    Mathematics was invented as an aid to reasoning about objects and processes in the world. Its

    main feature is abstraction and its main implementation mechanism is deduction from a set of

    axioms. In the beginning the axioms were restricted to self evident, non-reducible truths about the

    world. Tools of proof were restricted to real world artifacts such as the compass and divider.

    Eventually it was discovered that the restriction to self-evident truths was not fundamental to the

    process of deduction, and finite tools were not sufficient for the proof of all theorems that could

    otherwise be shown to be true. It was found that different axiomatic systems could give rise to

    diverse and interesting mathematical structures, some of which could be applied to real world

    problems. Throughout all of its history, the axiomatic method remained the unquestioned

    approach of mathematics. The cracks in that foundation evidenced by competing yet seemingly

    valid axiomatic systems spurred the attempt to find an encompassing set of axioms upon which

    to base all of mathematics. That search came to an abrupt end with the publishing of Goedel's

    Theorem which stated that no such system could exist. Mathematics had moved from an aid to

    reckoning, to acceptance as a main revealer of truth about the world ; it was considered the

    handmaiden or even the queen of the sciences. But like a skyrocket that rises from a bottle to thesky and then explodes into a thousand beautiful whirligigs, mathematics, while a beautiful sight to

    behold, no longer had, nor does it have, a discernible direction. The importance of Goedel's

    theorem should not be overlooked, it does not just apply to abstract mathematics. It applies to

    any codified (formal) system and so denies the validity of highly mathematical world models and

    loosely argued cosmologies alike.

    When a system is determined to be incomplete or inconsistent for some specific reason, then

    those inconsistencies or the incompleteness can often be corrected by enlarging the system to

    include appropriate axioms. This is reminiscent of the way in which one scientific theory

    supplants another. Unfortunately, any such measure simply creates another system to which

    Goedel's theorem applies as strongly as the first. To avoid an infinite regress it becomes24

  • 8/14/2019 Machine Minds

    25/79

    necessary at some point to accept a world system and the fact that, in it, there will be

    unexplainable (non-reducible) or contradictory phenomena. We will argue below that this is not

    simply a mathematical conundrum but a feature inherent in a creative hierarchically structured

    universe. In preparation for that argument we first discuss an obvious manifestation of the

    inexplicable in the world system that most humans find comfortable, that is the phenomena

    subject to description (not explanation) in the mathematics of chaos theory.

    Geometry, algebra, calculus, logic, computability theory, probability theory, and the various

    other areas and subdivisions of mathematics that have been developed over the years serve to

    describe and aid in the development of the various hard and soft sciences. As the sciences have

    come to treat more complex systems (e.g. global weather systems, the oceans, economic systems

    and biological systems to name just a few) the traditional mathematical tools have proved, in

    practice, to be inadequate. With the advent of computers a new tactic emerged. The systems

    could be broken down into manageable portions. A prediction would be calculated for a small

    step ahead in time for each piece, taking into account the effects of the nearby pieces. In this

    manner the whole system could be stepped ahead as far as desired. A shortcoming of the

    approach is the propagation of error inherent in the immense amount of calculation, in the

    approximation represented by the granularity of the pieces, and in the necessary approximation

    of nonlocal effects. To circumvent these problems, elaborate precautions consisting of

    redundancy and calculations of great accuracy have been employed. To some extent such tacticsare successful, but they always break down at some point in the projection. The hope has been

    that with faster more accurate computers the projections could be extended as far as desired.

    In 1980 Mitchell J. Feigenbaum published a paper (Feigenbaum 1980) in which he presented

    simple functionally described systems whose stepwise iterations (i.e. the output of one iteration

    serves as the input to the next iteration in a manner similar to the mechanism of the models

    mentioned above) produced erratic, nonpredictable trajectories. Beyond a certain point in the

    iteration, the points take on a verifiably random nature. That is the nature of the points

    25

  • 8/14/2019 Machine Minds

    26/79

    produced satisfy the condition for a truly random sequence of numbers 8. The resulting

    unpredictability of such systems arises not as an artifact of the accumulation of computational

    inaccuracies but rather as a fundamental sensitivity of the process to the initial conditions of the

    system. 9 Typically these systems display behavior that depends upon the value of parameters in

    that system. For example the equation x i+1 = ax i(1 x i) where 0 < x 0 < 1, has a single solution

    for each a < 3, (that is the iteration settles down so that for above some value of i, x i remains the

    same). At a = 3 there is an onset of doubling of solutions (i.e. beyond a certain i, x i alternates

    between a finite set of values; the number of values in that set doubles at incremental values as a

    increases). Above a = 3.5699... the system quits doubling and begins to produce random values.

    If you denote by l n the value for which the function doubles for the nth time then ( l n+1

    l n)/( l n l n1 ) = 4.6692. The value 4.6692 and another value 2.5029 are universal constants

    that have been shown to be associated with all such doubling systems. The points at which

    doubling occurs are called points of bifurcation. They are of considerable interest because for real

    systems that behave according to such an iterative procedure 10, there is an implication that the

    system might move arbitrarily to one of two or more solutions. The above example has one

    dimension but the same activity has been observed for many dimensional systems and

    apparently occurs for systems of any dimension. When the points generated by a two

    dimensional system are plotted on a coordinate system (for various initial conditions) they form

    constrained but unpredictable, and often beautiful trajectories. The space to which they areconstrained often exhibits coordinates to which some trajectories are attracted, sometimes

    8 For example a random sequence on an interval is random if the probability that the process will generate avalue in any given subinterval is the same as the probability of it generating a number in any other subintervalof the same length.

    9 Consider the following simple system: x 0 = pi 3, x i+1 = 10*x i trunc(10*x i), where trunc means todrop the decimal part of the argument. The sequence of numbers generated by the indicated iteration arenumbers between 0 and 1 where the nth number is the decimal portion of pi with the first n digits deleted.This is a random sequence of numbers arising from a deterministic equation. Changing the value of x 0,however slightly, inevitably leads to a different sequence of numbers.

    10 For example the equation x i+1 = ax i(1 x i) is called the logistics function and can serve as a model for the growth of populations.

    26

  • 8/14/2019 Machine Minds

    27/79

    providing a simple stable point to which the points gravitate or around which the points orbit. At

    other times trajectories orbit around multiple coordinates in varying patterns.

    More spectacularly, some trajectories gravitate toward what have been called strange attractors .

    Strange attractors are objects with fractional dimensionality. Such objects have been termed

    fractals by Benoit B. Mandelbrot. who developed a geometry based on the concept of fractional

    dimensionality (Mandelbrot, 1977). Fractals of dimension 0 < d < 1, consists of sets of points

    that might best be described as a set of points trying to be a line. An example of such a set of

    points is the Cantor set which is constructed by taking a line segment, removing the middle one

    third and repeating that operation recursively on all the line segments generated (stipulating that

    the remaining sections retain specifiable endpoints). In the limit an infinite set of points remains

    that exhibits a fractal dimensionality of log 2 / log 3 = .6309. Objects of varying dimensionality in

    the range [0,1] may be constructed by varying the size of the chunk removed from the middle of

    each line. An object with dimensionality 1 < d < 3 can be described as a line trying to become a

    plane. It consists of a line, so contorted, that it can have infinite length yet still be contained to a

    finite area of a plane. An example of such an object is the Koch snowflake as a fractal. The Koch

    snowflake is generated by taking an equilateral triangle and deforming the middle one third of each

    side outward into two sides of another equilateral triangle. After the first deformation a star of

    David is formed. Each of the twelve sides of the star of David is deformed in the same manner,

    and the sides of the resulting figure deformed, and so on. In the limit a snowflake shaped figurewith an infinitely long perimeter and a fractal dimensionality of log 4 / log 3 = 1.2618 results.

    Beyond a few iterations the snowflake does not change much in appearance.

    Many processes in the real world display the kind of abrupt change from stability to chaos that

    Feigenbaum's systems display. Turbulence, weather, population growth and decline, material

    phase changes (e.g. ice to water), dispersal of chemicals in solution and so forth. Many objects of

    the real world display fractal like dimensionality. Coastlines consists of convolutions, in

    convolutions, in convolution, and so on. Mountains consist of randomly shaped lumps, on

    randomly shaped lumps, on randomly shaped lumps, and so on. Cartographers have long noted

    that coastlines and mountains appear much the same no matter the scale. This happens because

    at large scales detail is lost and context is gained while at small scale detail is gained but context is27

  • 8/14/2019 Machine Minds

    28/79

    lost. In other words, shorelines look convoluted and mountains bumpy. It might be expected,

    given that the chaotic processes described above give rise to points, lines, orbits, and fractal

    patterns that real world systems might be described in terms of their chaotic tendencies. And

    indeed that is the case. The new area of research in mathematics and physics arising from these

    discoveries has come to be known as chaos theory . Chaos theory holds both promise and peril

    for model builders. Mathematical models whose iterations display the erratic behavior that

    natural systems display, capture to some extent the nature of the bounds, constraints and

    attractors to which the systems are subject. On the other hand, the fact that any change, however

    slight (quite literally), in the initial conditions from which the system starts, results in totally

    different trajectories implies that such models are of little use as predictors of a future state of the

    system. Further, it would appear that it may be very difficult, perhaps even impossible, to

    extract the equations of chaos (no matter that they may be very simple) from observation of the

    patterns to which those equations give rise. Whatever the case, the nature of the activity of these

    systems has profound implications for the sciences.

    If PierreSimon Laplace turns out to have been correct in assuming that the nature of the

    universe is deterministic (which the tenets of chaos theory do not preclude, but which they do

    argue against 11), he was wrong when he conjectured that he only needed enough calculating

    ability and an initial state to determine all the past and future history of the universe. A major

    error was in assuming that initial conditions could ever be known to the degree necessary to provide for such calculations. As an example, imagine a frictionless game of billiards played on a

    perfect billiard table. The player, who has absolutely perfect control over his shot and can

    11 We reject it mainly because of the butterfly effect which states that the flapping of the wings of a butterflyin Tokyo might affect the weather at a later date in New York City. This is because the weather systems at thelatitude of New York City and Tokyo travel West to East so that the weather in New York can be thought of as arising from the initial conditions of the atmosphere over Tokyo. As another example, consider the oldsaying, "for want of a nail the shoe was lost, for want of a shoe the horse was lost, for want of a horse the

    battle was lost, and for want of the battle the kingdom was lost." Consider the nail in the horse's shoe. For some coefficient of friction between the nail and the hoof, the shoe would have been retained. The exact valueof that coefficient might have depended on the state of a few molecules at the nail/hoof interface. Supposethose molecules were behaving in a manner that had brought them to a point of bifurcation that would resultin either an increase or decrease in the coefficient of friction. In the one case the shoe would have been lost andin the other it would have been retained.The loss of the nail would then have been a true random event thatwould have propagated its way to the level of events in human civilization. If the battle in question had beenthe battle of Hastings, the strong French influence in the language you are reading might well have beenmissing.

    28

  • 8/14/2019 Machine Minds

    29/79

    calculate all of the resulting collisions, sends all of the balls careening around the table. If he

    neglects to take into account the effects of one electron on the edge of the galaxy, his calculations

    will begin to diverge from reality within one minute (Crutchfield et al, 1986). Knowing the

    impossibility of prediction even in a theoretically deterministic universe, decouples the concepts

    of determinism and predictability, and of describability and randomness . Further, chaotic systems

    have built into them an arrow of time. The observer of such a system at some given point in time

    can not simply extrapolate backward to the initial conditions nor can he (so far as we know)

    divine the underlying equations by the character and relationship of the events it generates. The

    laws that underlie those events are obscured and can be determined only by simulated trial and

    error, the judicious use of analogy, description in terms of probability, or blind luck. There is a

    strong sense that something of Goedel's theorem is revealing itself here. Just as in any

    mathematical system of sufficient complexity there are truths that cannot be proved, there are

    real deterministic systems for which observations can be made that cannot be predicted. Perhaps

    the most important result of the discoveries in chaotic processes, is the growing realization

    among scientists that the linear, closed system, mathematics that have served so well for

    describing the universe during the last two hundred and fifty years, and which gave rise to the

    metaphor of a clockwork universe, only describe exceptionally well behaved and relatively rare

    (but very visible) phenomena of that universe. It is becoming apparent that it may not be

    possible to derive a neat linear, closed system model for everything that exhibits a pattern. It is

    likely that new nonlinear, open system approaches that rely heavily on computers will berequired to proceed beyond the current state of knowledge, and that in the explanation of

    phenomenon in terms of laws, those laws may have to be constrained to the levels of existence of

    the systems they describe and not be reduced to more fundamental laws at more fundamental

    levels.

    It is not surprising that the investigation of chaotic activity waited on the development of fast

    computers and high resolution graphics display devices. The amount of calculation required to get

    a sense of chaos is staggering and the structures revealed can only be appreciated graphically with

    the use of computers. But chaos theory has implications for computing as well. Computers have

    been around for a long time. The first machine that can truly be said to have been a computer was

    the Analytic Engine, designed in England in the mid nineteenth century by Charles Babbage. The29

  • 8/14/2019 Machine Minds

    30/79

    machine was never finished, but Ada Augusta Byron, daughter of Lord Byron, and who held the

    title, Countess of Lovelace, wrote programs for it, making her the first computer programmer.

    The Analytic Engine, though incomplete, was a showpiece that impressed and intimidated

    people. The countess is said to have stated (McCorduck, 1979):

    "The Analytical Engine has no pretensions whatever to originate anything. It can do whatever we know how toorder it to perform."

    Since then, those sentiments have often been repeated to reassure novice programmers and others

    whose lack of understanding causes them to see the computer as an intimidating, intelligent, and

    perhaps malevolent entity. More experienced programmers, however, find that the machine is

    truly an ignorant brute, incapable of doing anything on its own. And so the myth is propagated

    across generations. In chaos theory the myth is refuted. In pursuing chaotic processes themachine can truly create, though, in a mechanical manner that seems to lack purpose. We will

    argue that the seeming lack of purpose is a human prejudice based on our concept of what

    constitutes purpose, and that chaotic processes may very well be the mark of a creative universe.

    The structure revealed in chaotic processes is statistically accessible. While the sequence and

    position of the points generated within the constraints of the system are not predictable, if,

    taking as an example a two dimensional space, the points are treated as forming a histogram on a

    grid imposed upon the shape/form generated by the system then the system can be treated

    probabilistically. In terms of levels, the points may be seen as occupying one level and the form

    defined by the points as occupying the next higher level. The theory then has potential for the

    probabilistic prediction of system activities. This appeal to statistical description is not

    frivolous, but in fact is essential to any description of the universe as David Layzer indicates in

    the statement of his proposed cosmological principle.

    According to Layzer (Layzer 1990) (and as indicated above) the world/universe at its

    fundamental smallest scale (quantum level) consists of permanent , and indistinguishable objects

    that occupy discrete (but, in principle, unpredictable ) states as opposed to the continuous,

    theoretically predictable, distinguishable objects that populate the macro world. The

    30

  • 8/14/2019 Machine Minds

    31/79

    unpredictability permits truly random behavior as opposed to the deterministic behavior that is

    assumed to exist in a (classical) continuous macro world system. In combination with the

    discreteness this allows (in fact demands) a statistical description of the universe since the

    microstates of finite systems (consisting of particular configurations of quantum level objects) are

    finite in their description but statistical in their configuration. This has important ramifications in

    an infinite universe populated by such systems. An example that illuminates the connection

    between the discreteness of quantum level objects, the finiteness of systems composed of such

    objects, and the possibility of an adequate and complete statistical description of the universe is

    as follows:

    Consider a sequence of ones and zeros each of which has been chosen by the flip of a coin. If the

    sequence has a beginning or end then the sequence is unique and any other sequence (with a

    beginning or end) can be compared to it on an element to element basis (i.e. just compare the

    elements starting from the respective beginnings or ends). But if a sequence has no beginning or

    end then the only criteria for comparison is to stipulate that any sub-sequence, no matter how

    long, in the one sequence must have an identical counterpart in the second sequence (an

    impossible task to actually carry out). But, since in any such infinite sequence every finite

    subsequence must occur somewhere, then any two such infinite sequences must be identical. To

    produce a sequence of the first kind (one with a beginning or end) requires an infinite amount of

    information (the specific infinite sequence in question). But to produce an infinite sequence withno beginning or end only requires that the probability of a one or a zero at any cell in the infinite

    sequence be specified (in this example P(1) = P(0) = .5). A Laplacian, deterministic, universe

    corresponds to a specification of the first kind, and contains an infinite amount of specific

    information (the initial conditions). On the contrary, a statistically described universe contains

    very little information, and even that is statistical in nature. Yet obviously it describes a universe

    every bit as complex and varied as the Laplacian version.

    That is, a complete statistical description of the universe in terms of probabilities and averages

    can exist if the universe is

    1. infinite31

  • 8/14/2019 Machine Minds

    32/79

    2. statistically uniform (average properties are the same everywhere)3. finitely non-uniform (i.e. subsystems are finite)4. discrete (in the microstates of finite subsystems)

    All of this all gives rise to Layzers Cosmological Principle:

    A complete description of [an] initial state [of the universe] contains no information at all that would serve to define a preferred position in space or a preferred direction at any point.

    [That is, a statistical description that applies equally everywhere in the universe is preferred to anelaboration of the specific conditions that maintain at each place in the universe. Indeed, the

    statistical description is required in an infinite discrete universe.]

    The importance of a valid statistical description of the universe is very important for the next

    section in which order and entropy are explored. Given that we could explain (if not predict) the

    processes of nature using some combination of statistics and chaos theory, there still seems to be

    a missing element that spiritualists would happily call animus or moving force. What drives such

    systems in the first place? The traditional answer, given by thermodynamics, is that there is

    potential for work in thermodynamics in a system (force through a distance) anytime there is a

    heat source and a heat sink. Any system embodying such a source and sink is considered to be far

    from thermodynamic equilibrium until the temperatures of the source and sink have been

    equalized. A measure of the difficulty with which that heat can be converted to work is termedentropy. One of the most famous laws of science states that in a closed system in which work or

    heat interactions occur the entropy must increase. In the last century, An Austrian, Ludwig

    Boltzmann, recognized that this behavior is intimately connected with the concept of the

    possible states that a physical system might occupy. He was able to state the law simply as S =

    k log eP in which S is the entropy, P is the probability of the thermodynamic state occurring, and

    k is the ratio of ideal gas constant to Avogadro's number and is called Boltzmann's constant. The

    equation is engraved on Boltzmann's tombstone in Vienna. The description of systems by the use

    of statistics and probability theory as reflected in entropic processes is taken up next.

    32

  • 8/14/2019 Machine Minds

    33/79

    Order, entropy, and information (the emergence of levels)

    The word entropy is commonly used to indicate the process by which ordered systems become

    disordered 12. It is also the name for the measure of disorder. The next few paragraphs deal with

    the measure of entropy but the importance of entropy to this paper is as a process.

    There are two distinct types of systems to which the mathematical equation that describes

    entropy is commonly applied; the physical systems that are the object of study in physics, and

    the communication systems discussed in information theory. Entropy, as it applies to physical

    systems, is known as the second law of thermodynamics and is associated with timeirreversible

    characteristics of large systems whose sheer number of components and inaccessible initial

    conditions preclude deterministic analysis. Its form arises in the statistical consideration of the

    possible configurations of the components of a system and is often described as a measure of the

    randomness in motion and distribution of those components. The law states that, in any closed

    system 1 3 , the entropy, or disorder, or inability to convert heat or energy to work can only

    increase or remain the same over time. Among the laws of physics this irreversibility with

    respect to time is peculiar to the second law of thermodynamics (with the possible exception that

    might result from investigations in chaos theory). This view of entropy is intimately connected

    to the micro unpredictable nature of large real physical systems with many components and

    consequently does not often appear in classical physics where the behavior of small deterministickinetic systems is investigated.

    This form of the law is expressed in a simple form that first appeared in formulas attributable to

    DeMoivre and that he used to investigate games of chance (Wicken, 1988). When applied to

    thermodynamic systems (perhaps a container full of a gas or a liquid), it involves the12 A system is a structure of components that function together. This distinguishes them from a structurewhich may be an arrangement of non-dynamic items or an aggregate which may be an arrangement of nonrelated items. 'Function together' may be interpreted liberally to mean constrained to interact. We may,for instance, consider a gas in a container a system. The terms system, structure and aggregate depend to someextent upon context. The phrase "aggregate of stones" is meant to imply that we are speaking of a collectionof stones whose size, hardness and other stony characteristics are not related. We could speak of the samecollection as a structure if we had reason to refer to the spacial distribution of those stones.

    13 A closed system is one that cannot exchange energy, space or matter with its surroundings.

    33

  • 8/14/2019 Machine Minds

    34/79

    probabilities with which the states 14 of the system manifest themselves. The entropy of a

    system is given by the natural logarithm of the probability 'P i' of each state 'i' , weighted by itself

    and summed over all of the states, this yields the formula:

    S = k * S Pi *log ePi (i = 1 to W), (1)

    in which 'S' is the entropy, 'W' is the number of possible or attainable states of the system, and

    'k' is Boltzmann's constant 15. When the probability of occurrence of any state is the same as the

    probability of the occurrence of any other state, i.e. when all state probabilities are equal to 1/W,

    the maximum disorder occurs, (1) collapses to

    max(S) = k * log e W. (2)

    Given this measure of the maximum entropy possible in a system and the actual entropy (1), and

    allowing that the percentage disorder displayed by a system can be represented as the ratio of

    actual entropy to the maximum entropy, then we can define

    disorder S/max(S), (3)

    so that it is natural to define

    order 1 disorder. (4)

    14 For a physical system 'State' implies one specific configuration of components (probably but notnecessarily at the molecular level). A state is thought of as occupying a point in phase space. Phase space issimply a multidimensional space whose dimensions represent the spacial coordinates and momenta of thecomponents of the system so that one point in pha