Principles of Complex Systems: How to think like nature Part 1 Russ Abbott Sr.

Download Report

Transcript Principles of Complex Systems: How to think like nature Part 1 Russ Abbott Sr.

Principles of Complex Systems:
How to think like nature
Part 1
Russ Abbott
Sr. Engr. Spec.
310-336-1398
Does nature really think?
[email protected]
 1998-2007. The Aerospace Corporation. All Rights Reserved.
1
Complex systems course overview
9:00–9:10.
9:10–9:25.
9:25–9:45.
9:45–9:55.
9:55–10:05.
10:05–10:15.
10:15–10:30.
10:30–10:45.
10:45–10:55.
10:55–11:00.
Introduction and motivation.
Overview – unintended consequences, mechanism,
function, and purpose; levels of abstraction, emergence,
introduction to NetLogo.
Emergence, levels of abstraction, and the reductionist blind
spot.
Modeling; thought externalization; how engineers and
computer scientists think.
Lots of echoes and
Break.
repeated themes from
Evolution and evolutionary computing. one section to another.
Innovation – exploration and exploitation.
Platforms – distributed control and systems of systems.
Groups – how nature builds systems; the wisdom of crowds.
Summary/conclusions – remember this if nothing else.
2
Principles of Complex Systems:
How to think like nature
What “complex systems” means,
and why you should care.
Russ Abbott
Sr. Engr. Spec.
310-336-1398
[email protected]
 1998-2007. The Aerospace Corporation. All Rights Reserved.
3
What we will be talking about.
“Complex systems” refers to an
anti-reductionist way of thinking
that developed in the 1980s in
Biology, Computer Science,
Economics, Physics, and other
fields. (The term complexity is
also used this way.)
– It is not intended to refer to a
particular category of systems, which
are presumably distinguished from
other systems that aren’t “complex.”
Isn’t that true
of all systems?
But if I had to define what a
“complex system” is …
– A collection of autonomous
elements that interact both with
each other and with their
environment and that exhibits
aggregate, ensemble, macro
behaviors that none of the
elements exhibit individually.
System: a construct or collection of
different elements that together
produce results not obtainable by the
elements alone. — Eberhardt Rechtin
We are in the business of
Systems Architecting of Organizations:
building “complex systems.” Why Eagles Can't Swim, CRC, 1999.
4
A satellite in a geostationary orbit:
one of the simplest possible “complex systems”
Fixed with respect to the earth as a reference frame.
An “emergent” property
But nothing is tying it down.
No cable is holding it in place.
What is the environment?
period of the orbit = period of the earth’s rotation
Typical of complex system mechanisms.
Multiple independent or quasi-independent processes
— which are not directly connected causally (agents) —
interact within an environment to produce a result.
5
Complex systems terms
• Emergence. A level of abstraction that can be
described independently of its implementation.
• Multi-scalar. Applicable to systems that are
understood on multiple levels simultaneously,
especially when a lower level implements the
emergence of some functionality at a higher level.
6
See next
few slides
Why should you care?
• Because our corporate leadership, our customers, and their contractors
think it’s important.
– Rumsfeld’s inspiration for transformation in the military grew out of this way
of thinking. This field is not new. It’s at least 2 decades old.
– The Command and Control Research Program (CCRP) in the Pentagon
(Dave Alberts) is successfully promoting this style of thinking within the DoD.
– Complex systems thinking is a generalization of and the foundation for netcentric thinking—and the way the world has changed as a result of the web.
• You should understand what they are talking about
– So that you can explain it to them.
• Because it offers a powerful new way to think about how systems work.
• Because large systems—and especially systems of systems (another
important buzz-word)—tend to be complex in the ways we will discuss.
• Because the ideas are interesting, important, and good for you.
7
General Hamel and Dr. Austin think it’s important
• General Michael Hamel (moderator)
– “Where Commercial, Civil, and Military Space
Intersect: Cooperation, Conflict, and Execution of
the Mission”
• Plenary Session, AIAA Space 2007.
• Dr. Wanda Austin,
– “Space System of Systems Engineering”
• USC Center for Systems and Software Engineering
Convocation, October 2006.
8
What is System of Systems Engineering?*
The process of planning, analyzing, organizing, and integrating
the capability of a mix of existing and new systems into a
system-of-system capability that is greater than the sum of
the capabilities of the constituent parts.
Emergence
The process emphasizes the process of discovering,
developing, and implementing standards that promote
interoperability among systems developed via different
sponsorship, management, and primary acquisition processes.
Many people know that something is missing
from the way we look at systems traditionally. But
most are groping to express just what it is. This
course is intended to help sharpen the picture.
Platforms
* USAF SAB Report: System of Systems Engineering for Air Force
Capability Development, July 2005
9
What is a System of Systems?
Functional decomposition
Small stovepipes
to large stovepipes – NO
Level of
abstraction
Loosely coupled and tightly
integrated – YES
Platforms
10
10
Planning Complex Endeavors (April 2007)
by David S. Alberts and Richard E. Hayes
Alberts’ term for what a
complex system does.
The Command and Control Research Program (CCRP)
has the mission of improving DoD’s understanding of
the national security implications of the Information Age.
• John G. Grimes, Assistant Secretary of Defense (NII) & Chief
Information Officer
• Dr. Linton Wells, II, Principal Deputy Assistant Secretary of
Defense (NII)
• Dr. David S. Alberts, Special Assistant to the ASD(NII) & Director
of Research
11
From the forward by John G. Grimes
As this latest book from the CCRP explains, we can no longer be
content with building an “enterprise-wide” network that stops at the
edges of our forces, nor with a set of information sources and
channels that are purely military in nature. We need to be able to work
with a large and diverse set of entities and information sources. We
also need to develop new approaches to planning that are better
suited for these coalition operations.
The implications are significant for a CIO as it greatly expands the
who, the what, and the how of information sharing and collaboration.
What is this “new
way of thinking?”
It also requires a new way of thinking about
effectiveness, increasing the emphasis we place on agility, which, as
is explained in this book, is the necessary response to uncertainty and
complexity.
From Chapter 1. Introduction
Platforms
The economics of communications and information technologies has created enormous
opportunities to leverage the power of information and collaboration cost effectively by adopting
Power to the Edge principles and network-centric concepts.
Exploration and
exploitation
Fine to use these
terms, but what do we
really mean by them?
12
From Chapter 2. Key Concepts
Complicated Systems
Systems that have many moving parts or actors and are highly
dynamic, that is, the elements of these systems constantly interact
with and impact upon one another. However, the cause and effect
relationships within a complicated situation are generally well
understood, which allows planners to predict the consequences of
specific actions with some confidence.
I think this misses the point. Most systems and interactions are (eventually) “well
understood.” Complicated systems are often fully entrained with one locus of control.
13
From Chapter 2. Key Concepts
Both chaos and phase transitions
Complex Endeavors (Systems)
Complex endeavors involve changes and behaviors that
I disagree—although
sometimes the only
way to predict it is to
run (a model of) it.
cannot be
predicted in detail, although those behaviors and changes can be
expected to form recognizable patterns. Complex endeavors are also characterized by
small differences in initial conditions or
relatively small perturbations (seemingly tactical actions) are associated with very
large changes in the resulting patterns of
behavior and/or strategic outcomes.
circumstances in which relatively
Some complex situations develop into complex adaptive systems (CAS), which tend to
be robust—to persist over time and across a variety of circumstances. These are often
observed in nature in the form of
biological or ecological
systems. However, while these systems are thought of as robust, they can be
pushed out of balance even to the point of collapse through cascades of negatively
reinforcing conditions and behaviors. Such perturbations are what ecologists fear when
a habitat is reduced to an isolated geographic area or when invasive, nonnative
species are introduced.
Note biological reference
14
Like many networks, this complex system
involves many independently operating
elements (the trains) that together enable one to
get from any one station to any other without a
massive number of point-to-point connections.
Simon Patterson is fascinated by the information which orders our lives. He humorously dislocates and subverts sources of information
such as maps, diagrams and constellation charts; one of his best known works is The Great Bear, in which he replaced the names of stations
on the London Underground map with names of philosophers, film stars, explorers, saints and other celebrities. By transforming
authoritative data with his own associations he challenges existing rationales.
15
The world in a grain of sand
To see the world in a grain of sand,
and heaven in a wild flower,
to hold infinity in the palm of your hands,
and eternity in an hour. –William Blake
Do you understand what that means? It’s
profound, but it uses poetry to make its point.
A primary objective of this class is to say in as plain
a way as possible what many people have been
groping for when talking about complex systems.
Many of the ideas may seem like common sense.
It’s just that we’ll be looking at them more closely.
16
Introduction to Complex Systems:
How to think like nature
Unintended consequences;
mechanism, function, and purpose
Russ Abbott
Sr. Engr. Spec.
310-336-1398
[email protected]
This segment introduces
some basic concepts.
 1998-2007. The Aerospace Corporation. All Rights Reserved.
17
A fable
• Once upon a time, a state in India had too many snakes.
• To solve this problem the government instituted an incentivebased program to encourage its citizens to kill snakes.
• It created the No Snake Left Alive program.
– Anyone who brings a dead snake into a field office of the
Dead Snake Control Authority (DSCA) will be paid a
generous Dead Snake Bounty (DSB).
• A year later the DSB budget was exhausted. DSCA had paid
for a significant number of dead snakes.
• But there was no noticeable reduction in the number of
snakes plaguing the good citizens of the state.
• What went wrong?
18
The DSCA mechanism
Receive dead snake
certificate. Submit
certificate to DSCA.
DSCA
What would you do if
this mechanism were
available in your world?
Receive
money.
Start a
snake
farm.
Catch, kill, and submit
a dead snake.
Dead snake
verifier
19
Moral: unintended consequences
• A mechanism is installed in an environment.
• The mechanism is used/exploited in unanticipated
ways.
• Once a mechanism is installed in the environment, it
will be used for whatever purposes “users” can think
to make of it …
– which may not be that for which it was originally
intended.
The first lesson of complex systems thinking is
that one must always be aware of the relationship
between systems and their environments.
20
Parasites that control their hosts
• Dicrocoelium dendriticum causes host ants
to climb grass blades where they are eaten
by grazing animals, where D. dendriticum
lives out its adult life.
• Toxoplasma gondii causes host mice not to
fear cats, where T. gondii reproduces.
• Spinochordodes tellinii causes host insects
to jump into the water and drown, where S.
tellinii grows to adulthood.
It’s amazing how far exploitation
of environmental mechanisms
can go. (See platforms, later.)
21
Locomotion in E. coli
• E. coli movements consist of
short straight runs, each lasting
a second or less, punctuated by
briefer episodes of random
tumbling.
• Each tumble reorients the cell and sets it off in a
new direction.
Exploration
Exploitation
• Cells that are moving up the gradient of an attractant
tumble less frequently than cells wandering in a
homogeneous medium or moving away from the source.
• In consequence, cells take longer runs toward the source
and shorter ones away.
Gain benefit
Harold, Franklyn M. (2001) The Way of the Cell: Molecules,
Organisms, and the Order of Life, Oxford University Press.
22
Mechanism, function, and purpose
• Mechanism: The physical processes within
an entity.
– The chemical reactions built into E.coli that result in its
flagella movements.
– The DSCA mechanism.
• Function: The effect of a mechanism on the
environment and on the relationship between
an entity and its environment.
– E. coli moves about. In particular, it moves up nutrient
gradients.
– Snakes are killed and delivered; money is exchanged.
Wikipedia Commons
• Purpose: The (presumably positive)
consequence for the entity of the change in
its environment or its relationship with its
environment. (But Nature is not teleological.)
– E. coli is better able to feed, which is necessary for its
survival.
– Snake farming is encouraged?
Socrates
Compare to Measures of Performance, Effectiveness, and Utility
23
Teleology: building “purpose”
Nature
Designed
E.g., E. coli locomotion to food
E.g., Reduce snake population
• Evolve a new mechanism
• Envision a purpose
• Experience the resulting
functionality
• Imagine how a function can
achieve that purpose
• If the functionality enhances
survival, keep the mechanism
• Design and develop a
mechanism to perform that
function
• “Purpose” has been created
implicitly as part of a new
level of abstraction
Most of the design steps require
significant conceptualization abilities.
• Deploy the mechanism and
hope the purpose is achieved
In both cases, the world will
be changed by the addition
of the new functionality. The
purpose is more likely to be
achieved in nature.
24
NetLogo: let’s try it
File > Models Library > Biology > Ants
Click Open
25
Two levels of emergence
• No individual chemical reaction inside the
ants is responsible for making them follow
the rules that describe their behavior.
• That the internal chemical reactions
together do is an example of emergence.
• No individual rule and no individual ant is
responsible for the ant colony gathering
food.
• That the ants together bring about that
result is a second level of emergence.
Colony results
Ant behaviors
Ant chemistry
Each layer is a
level of abstraction
Notice the similarity to layered
communication protocols
26
Two levels of emergence
Applications, e.g., email, IM, Wikipedia
• No individual chemical reaction inside the
WWW
(HTML) — for
browsers
+ servers
ants
is responsible
making
them follow
the rules that describe
their behavior.
Presentation
• That the internal Session
chemical reactions
together do is an
example of emergence.
Transport
• No individual rule
and no individual ant is
Network
responsible for the ant colony gathering
Physical
food.
• That the ants together bring about that
result is a second level of emergence.
Colony results
Ant behaviors
Ant chemistry
Each layer is a
level of abstraction
Notice the similarity to layered
communication protocols
27
Principles of Complex Systems:
How to think like nature
Emergence: what’s right and what’s
wrong with reductionism
Russ Abbott
Sr. Engr. Spec.
310-336-1398
[email protected]
Philosophical, but
with a practical
corollary at the end.
Presumptuous?
 1998-2007. The Aerospace Corporation. All Rights Reserved.
28
Emergence: the holy grail of complex systems
How macroscopic behavior arises from microscopic behavior.
Emergent entities (properties
or substances) ‘arise’ out of
more fundamental entities and
yet are ‘novel’ or ‘irreducible’
with respect to them.
Stanford Encyclopedia of Philosophy
http://plato.stanford.edu/entries/properties-emergent/
Plato
The ‘scare’ quotes identify
problematic areas.
Emergence: Contemporary Readings in Philosophy and Science
Mark A. Bedau and Paul Humphreys (Eds.), MIT Press, April 2008.
29
Cosma Shalizi
http://cscs.umich.edu/~crshalizi/reviews/holland-on-emergence/
Someplace … where quantum field theory
meets general relativity and atoms and void
merge into one another, we may take “the
rules of the game” to be given.
Call this
emergence
if you like.
It’s a fine-sounding
word, and brings to
mind southwestern
creation myths in
an oddly apt way.
But the rest of the observable, exploitable order
in the universe
benzene molecules, PV = nRT, snowflakes, cyclonic
storms, kittens, cats, young love, middle-aged
remorse, financial euphoria accompanied with acute
gullibility, prevaricating candidates for public office,
tapeworms, jet-lag, and unfolding cherry blossoms
Where do all these regularities come from?
30
Erwin Schrödinger
“[L]iving matter, while not eluding the ‘laws of
physics’ … is likely to involve ‘other laws,’ [which]
will form just as integral a part of [its] science.”
Erwin Schrödinger, What is Life?, 1944.
Jerry Fodor
Steven Weinberg
The ultimate reductionist.
Why is
there
anything
except
physics?
Philip Anderson
John
Holland
The ability to reduce everything to simple
fundamental laws [does not imply] the ability to
start from those laws and reconstruct the
universe. … [We] must all start with reductionism,
which I fully accept. “More is Different” (Science, 1972)
31
The fundamental dilemma of science
Emergence
Are there autonomous
higher level laws of nature?
The functionalist claim
The reductionist position
How can that be if everything can be reduced
to the fundamental laws of physics?
My answer
It can all be explained in terms
of levels of abstraction.
32
The Game of Life
File > Models Library > Computer Science > Cellular Automata > Life
Click Open
33
Gliders
• Gliders are causally powerless.
– A glider does not change how the rules operate or which cells will be
switched on and off. A glider doesn’t “go to an cell and turn it on.”
– A Game of Life run will proceed in exactly the same way whether one
notices the gliders or not. A very reductionist stance.
• But …
– One can write down equations that characterize glider motion and
predict whether—and if so when—a glider will “turn on” a particular cell.
– What is the status of those equations? Are they higher level laws?
Like shadows, they don’t “do” anything.
The rules are the only “forces!”
34
Game of Life as a Programming Platform
• Amazing as they are, gliders are also trivial.
– Once we know how to produce a glider, it’s
simple to make them.
• Can build a library of Game of Life patterns and
their interaction APIs.
By suitably arranging these patterns,
one can simulate a Turing Machine.
Paul Rendell. http://rendell.server.org.uk/gol/tmdetails.htm
A second level of emergence.
Emergence is not particularly mysterious.
35
Downward causation entailment
• The unsolvability of the TM halting problem entails the
unsolvability of the GoL halting problem.
– How strange! We can conclude something about the GoL because
we know something about Turing Machines.
– Yet the theory of computation is not derivable from GoL rules.
• One can use glider “velocity” laws to draw conclusions (make
predictions) about which cells will be turned on and when that will
happen. (Also downward entailment.)
GoL gliders and Turing Machines are
causally reducible but ontologically real.
– You can reduce them away without
changing how a GoL run will proceed.
– Yet they obey higher level laws, not
derivable from the GoL rules.
36
Level of abstraction: the reductionist blind spot
A concept computer science has contributed to the world.
A collection of concepts and relationships that can be described
independently of its implementation.
Every computer application creates one.
A level of abstraction is causally reducible to its implementation.
– You can look at the implementation to see how it works.
Its independent specification—its properties and way of being in
the world—makes it ontologically real.
– How it interacts with the world is based on its specification
and is independent of its implementation.
– It can’t be reduced away without losing something
37
Practical corollary: feasibility ranges
• Physical levels of abstraction
are implemented only within
feasibility ranges.
• When the feasibility range is
exceeded a phase transition
generally occurs.
Require contractors to identify the feasibility range
within which the implementation will succeed and
describe the steps taken to ensure that those
feasibility ranges are honored—and what happens
if they are not. (Think O-rings.)
38
Principles of Complex Systems:
How to think like nature
Modeling, the externalization of
thought, and how engineers and
computer scientists think
Russ Abbott
Sr. Engr. Spec.
310-336-1398
[email protected]
 1998-2007. The Aerospace Corporation. All Rights Reserved.
39
Modeling problems:
the difficulty of looking downward
Models of computer security or
terrorism will always be incomplete.
Can only model
unimaginative
enemies.
• It is not possible to find a non-arbitrary base level
for models.
– What are we leaving out that might matter?
• Use Morse code to transmit messages on encrypted lines.
• No good models of biological arms races.
– Insects vs. plants: bark, bark boring, toxin, anti-toxin, … .
• Geckos use the Van der Waals “force” to climb.
Nature is not segmented into
a strictly layered hierarchy.
Epiphenomenal
40
Modeling problems:
the difficulty of looking upward
• Don’t know how to build models that can notice emergent
phenomena and characterize their interactions. We don’t
know what we aren’t noticing.
– We/they can use our commercial airline system to deliver
mail/bombs.
Exploit an
• Model gravity as an agent-based system. existing process
– Ask system to find equation of earth’s orbit.
– Once told what to look for, system can find ellipse. (GP)
– But it won’t notice the yearly cycle of the seasons — even
though it is similarly emergent.
Models of computer security or
terrorism will always be incomplete.
Can only model
unimaginative
enemies.
41
Turning dreams into reality
• Computer Scientists and Engineers both turn
dreams (ideas) into reality—systems that operate
in the world.
• But we do it in very different ways.
42
Intellectual leverage in Computer Science:
executable externalized thought
• Computer languages enable executable externalized thought—
different from (nearly) all other forms of externalized thought
throughout history!
– Software is both intentional—has meaning—and executable.
– All other forms of externalized thought (except music) require a
human being to interpret them.
• The bit provides a floor that is both symbolic and real.
– Bits are: symbolic, physically real, and atomic.
– Bits don’t have error bars.
– Can build (ontologically real) levels of abstraction above them.
• But the bit limits realistic modeling.
– E.g., no good models of evolutionary arms races and many
other multi-scale (biological) phenomena. No justifiable floor.
– Challenge: build a computer modeling framework that
supports dynamically varying floors.
43
Intellectual leverage in Engineering:
mathematical modeling
• Engineering gains intellectual leverage through
mathematical modeling and functional decomposition.
– Models approximate an underlying reality (physics).
– Models don’t create ontologically independent entities.
• Engineering is both cursed and blessed by its
attachment to physicality.
– There is no reliable floor in the material world.
• Engineering systems often fail because of
unanticipated interactions among well designed
components, e.g. acoustic coupling that could not be
identified in isolation from the operation of the full
systems. National Academy of Engineering, Design in the New Millennium, 2000.
– But, if a problem appears, engineers (like scientists) can
dig down to a lower level to solve it.
44
Engineers and computer scientists are different —
almost as different as Venus and Mars
• Engineers are grounded in physics.
– Ultimately there is nothing besides physics.
– Even though engineers build things that have very different
(emergent) properties from their components, engineers tend
to think at the level of physics.
– When designing systems, engineers start with an idea and
build down to the physics—using functional decomposition
and successive approximation.
• Engineering is (proudly) applied physics.
• Computer scientists live in a world of abstractions.
– Physics has very little to do with computer science worlds.
– For computer scientists, there is more than physics, but we
may have a hard time saying what it is—emergence.
– When designing systems, Computer scientists start with the
bit and build up to the idea—using levels of abstraction.
• Computer science is (cautiously) applied philosophy.
45
Complex systems course overview
9:00–9:10.
9:10–9:25.
9:25–9:45.
9:45–9:55.
9:55–10:05.
10:05–10:15.
10:15–10:30.
10:30–10:45.
10:45–10:55.
10:55–11:00.
Introduction and motivation.
Unintended consequences – mechanism, function, and
purpose; introduction to NetLogo.
Emergence – the reductionist blind spot and levels of
abstraction.
Modeling; thought externalization; how engineers and
computer scientists think.
Break.
Evolution and evolutionary computing.
Innovation – exploration and exploitation.
Platforms – distributed control and systems of systems.
Groups – how nature builds systems; the wisdom of crowds.
Summary/conclusions – remember this if nothing else.
46
Backups
47
The reductionist blind spot
• Darwin and Wallace’s theory of evolution by natural selection is
expressed in terms of
– entities
– their properties
– how suitable the properties of the entities are for the environment
– populations
– reproduction
– etc.
• These concepts are a level of abstraction.
– The theory of evolution is about entities at that level of abstraction.
• Let’s assume that it’s (theoretically) possible to trace how any state
of the world—including the biological organisms in it—came about
by tracking elementary particles
• Even so, it is not possible to express the theory of evolution in terms
of elementary particles.
• Reducing everything to the level of physics, i.e., naïve reductionism,
results in a blind spot regarding higher level entities and the laws
that govern them.
48
I’m showing this slide to invite anyone who is interested to work on this with me.
How are levels of abstraction built?
• By adding persistent constraints to what exists.
– Constraints “break symmetry” by ruling out possible future states.
– Should be able to relate this to symmetry breaking more generally.
• Easy in software.
– Software constrains a computer to operate in a certain way.
– Software (or a pattern set on a Game of Life grid) “breaks the
symmetry” of possible sequences of future states.
• How does nature build levels of abstraction? Two ways.
Isn’t this just common sense?
– Energy wells produce static entities.
Ice cubes act differently from
• Atoms, molecules, solar systems, …
water and water molecules.
– Activity patterns use imported energy to produce dynamic entities.
• The constraint is imposed by the processes that the dynamic
entity employs to maintain its structure.
• Biological entities, social entities, hurricanes.
• A constrained system operates differently (has additional laws—
the constraints) from one that isn’t constrained.
49