Emergence

From SEBoK
Revision as of 16:04, 12 April 2022 by Gsmith (talk | contribs)

Jump to navigation Jump to search

Lead Author: Rick Adcock, Contributing Authors: Scott Jackson, Dick Fairley, Janet Singer, Duane Hybertson


This topic forms part of the Systems Science knowledge area (KA). It gives the background to some of the ways in which emergenceemergence has been described, as well as an indication of current thinking on what it is and how it influences systems engineeringsystems engineering (SE) practice. It will discuss how these ideas relate to the general definitions of systemssystems given in What is a System?; in particular, how they relate to different engineered systemengineered system contexts. This topic is closely related to the complexity topic that precedes it.

Emergence is a consequence of the fundamental system conceptsconcepts of holismholism and interaction (Hitchins 2007, 27). System wholes have behaviorsbehaviors and properties arising from the organization of their elementselements and their relationships, which only become apparent when the system is placed in different environmentsenvironments.

Questions that arise from this definition include: What kinds of systems exhibit different kinds of emergence and under what conditions? Can emergence be predicted, and is it beneficial or detrimental to a system? How do we deal with emergence in the development and use of engineered systems? Can it be planned for? How?

There are many varied and occasionally conflicting views on emergence. This topic presents the prevailing views and provides references for others.

Overview of Emergence

As defined by Checkland, emergenceemergence is “the principle that entities exhibit properties which are meaningful only when attributed to the whole, not to its parts.” (Checkland 1999, 314). Emergent system behaviorbehavior can be viewed as a consequence of the interactions and relationships between system elementssystem elements rather than the behavior of individual elements. It emerges from a combination of the behavior and properties of the system elements and the systems structure or allowable interactions between the elements, and may be triggered or influenced by a stimulus from the systems environment.

Emergence is common in nature. The pungent gas ammonia results from the chemical combination of two odorless gases, hydrogen and nitrogen. As individual parts, feathers, beaks, wings, and gullets do not have the ability to overcome gravity; however, when properly connected in a bird, they create the emergent behavior of flight. What we refer to as “self-awareness” results from the combined effect of the interconnected and interacting neurons that make up the brain (Hitchins 2007, 7).

Hitchins also notes that technological systems exhibit emergence. We can observe a number of levels of outcome which arise from interaction between elements in an engineered systemengineered system context. At a simple level, some system outcomes or attributesattributes have a fairly simple and well defined mapping to their elementselements; for example, center of gravity or top speed of a vehicle result from a combination of element properties and how they are combined. Other behaviors can be associated with these simple outcomes, but their value emerges in complexcomplex and less predictable ways across a system. The single lap performance of a vehicle around a track is related to center of gravity and speed; however, it is also affected by driver skill, external conditions, component ware, etc. Getting the 'best' performance from a vehicle can only be achieved by a combination of good design and feedback from real laps under race conditions.

There are also outcomes which are less tangible and which come as a surprise to both system developers and usersusers. How does lap time translate into a winning motor racing team? Why is a sports car more desirable to many than other vehicles with performances that are as good or better?

Emergence can always be observed at the highest level of system. However, Hitchins (2007, 7) also points out that to the extent that the systems elements themselves can be considered as systems, they also exhibit emergence. Page (2009) refers to emergence as a “macro-level property.” Ryan (2007) contends that emergence is coupled to scopescope rather than system hierarchical levels. In Ryan’s terms, scope has to do with spatial dimensions (how system elements are related to each other) rather than hierarchical levels.

Abbott (2006) does not disagree with the general definition of emergence as discussed above. However, he takes issue with the notion that emergence operates outside the bounds of classical physics. He says that “such higher-level entities…can always be reduced to primitive physical forces.”

Bedau and Humphreys (2008) and Francois (2004) provide comprehensive descriptions of the philosophical and scientific background of emergence.

Types of Emergence

A variety of definitions of types of emergence exists. See Emmeche et al. (1997), Chroust (2003) and O’Connor and Wong (2006) for specific details of some of the variants. Page (2009) describes three types of emergence: "simple", "weak", and "strong".

According to Page, simple emergence is generated by the combination of element properties and relationships and occurs in non-complex or “ordered” systems (see Complexity) (2009). To achieve the emergent property of “controlled flight” we cannot consider only the wings, or the control system, or the propulsion system. All three must be considered, as well as the way these three are interconnected-with each other, as well as with all the other parts of the aircraft. Page suggests that simple emergence is the only type of emergence that can be predicted. This view of emergence is also referred to as synergysynergy (Hitchins 2009).

Page describes weak emergence as expected emergence which is desired (or at least allowed for) in the system structurestructure (2009). However, since weak emergence is a product of a complex system, the actual level of emergence cannot be predicted just from knowledge of the characteristics of the individual system componentscomponents.

The term strong emergence is used to describe unexpected emergence; that is, emergence not observed until the system is simulated or tested or, more alarmingly, until the system encounters in operation a situation that was not anticipated during design and development.

Strong emergence may be evident in failures or shutdowns. For example, the US-Canada Blackout of 2003 as described by the US-Canada Power System Outage Task Force (US-Canada Power Task Force 2004) was a case of cascading shutdown that resulted from the designdesign of the system. Even though there was no equipment failure, the shutdown was systemic. As Hitchins points out, this example shows that emergent properties are not always beneficial (Hitchins 2007, 15).

Other authors make a different distinction between the ideas of strong, or unexpected, emergence and unpredictable emergence:

  • Firstly, there are the unexpected properties that could have been predicted but were not considered in a systems development: "Properties which are unexpected by the observer because of his incomplete data set, with regard to the phenomenon at hand" (Francois, C. 2004, 737). According to Jackson et al. (2010), a desired level of emergence is usually achieved by iteration. This may occur as a result of evolutionary processesprocesses, in which element properties and combinations are "selected for", depending on how well they contribute to a system’s effectiveness against environmentalenvironmental pressures or by iteration of design parameters through simulationsimulation or build/test cycles. Taking this view, the specific values of weak emergence can be refined, and examples of strong emergence can be considered in subsequent iterations so long as they are amenable to analysis.
  • Secondly, there are unexpected properties which cannot be predicted from the properties of the system’s components: "Properties which are, in and of themselves, not derivable a priori from the behavior of the parts of the system" (Francois, C. 2004, 737). This view of emergence is a familiar one in social or natural sciences, but more controversial in engineeringengineering. We should distinguish between a theoretical and a practical unpredictability (Chroust 2002). The weather forecast is theoretically predictable, but beyond certain limited accuracy practically impossible due to its chaoticchaotic nature. The emergence of consciousness in human beings cannot be deduced from the physiological properties of the brain. For many, this genuinely unpredictable type of complexity has limited value for engineering. (See Practical Considerations below.)

A type of system particularly subject to strong emergence is the system of systems (sos)system of systems (sos). The reason for this is that the SoS, by definition, is composed of different systems that were designed to operate independently. When these systems are operated together, the interaction among the parts of the system is likely to result in unexpected emergence. Chaotic or truly unpredictable emergence is likely for this class of systems.

Emergent Properties

Emergent propertiesEmergent properties can be defined as follows: “A property of a complex system is said to be ‘emergent’ [in the case when], although it arises out of the properties and relations characterizing its simpler constituents, it is neither predictable from, nor reducible to, these lower-level characteristics” (Honderich 1995, 224).

All systems can have emergent properties which may or may not be predictable or amenable to modelingmodeling, as discussed above. Much of the literature on complexitycomplexity includes emergence as a defining characteristic of complex systems. For example, Boccara (2004) states that “The appearance of emergent properties is the single most distinguishing feature of complex systems.” In general, the more ordered a system is, the easier its emergent properties are to predict. The more complex a system is, the more difficult predicting its emergent properties becomes.

Some practitioners use the term “emergence” only when referring to “strong emergence”. These practitioners refer to the other two forms of emergent behavior as synergy or “system level behavior” (Chroust 2002). Taking this view, we would reserve the term "Emergent Property" for unexpected properties, which can be modeled or refined through iterations of the systems development.

Unforeseen emergence causes nasty shocks. Many believe that the main job of the systems approachsystems approach is to prevent undesired emergence in order to minimize the riskrisk of unexpected and potentially undesirable outcomes. This review of emergent properties is often specifically associated with identifying and avoiding system failures (Hitchins 2007).

Good SE isn't just focused on avoiding system failure, however. It also involves maximizing opportunityopportunity by understanding and exploiting emergence in engineered systemsengineered systems to create the required system level characteristics from synergistic interactions between the componentscomponents, not just from the components themselves (Sillitto 2010).

One important group of emergent properties includes properties such as agilityagility and resilienceresilience. These are critical system properties that are not meaningful except at the whole system level.

Practical Considerations

As mentioned above, one way to manage emergent properties is through iteration. The requirements to iterate the design of an engineered system to achieve desired emergence results in a designdesign processprocess are lengthier than those needed to design an ordered system. Creating an engineered system capable of such iteration may also require a more configurable or modular solution. The result is that complex systems may be more costly and time-consuming to develop than ordered ones, and the cost and time to develop is inherently less predictable.

Sillitto (2010) observes that “engineering design domains that exploit emergence have good mathematical models of the domain, and rigorously control variability of components and subsystems, and of process, in both design and operation.” The iterations discussed above can be accelerated by using simulation and modeling, so that not all the iterations need to involve building real systems and operating them in the real environment.

The idea of domain models is explored further by Hybertson in the context of general models or patternspatterns learned over time and captured in a model space (Hybertson 2009). Hybertson states that knowing what emergence will appear from a given design, including side effects, requires hindsight. For a new type of problem that has not been solved, or a new type of system that has not been built, it is virtually impossible to predict emergent behavior of the solution or system. Some hindsight, or at least some insight, can be obtained by modeling and iterating a specific system design; however, iterating the design within the development of one system yields only limited hindsight and often does not give a full sense of emergence and side effects.

True hindsight and understanding comes from building multiple systems of the same type and deploying them, then observing their emergent behavior in operation and the side effects of placing them in their environments. If those observations are done systematically, and the emergence and side effects are distilled and captured in relation to the design of the systems — including the variations in those designs — and made available to the community, then we are in a position to predict and exploit the emergence.

Two factors are discovered in this type of testing environment: what works (that is, what emergent behavior and side effects are desirable); and what does not work (that is, what emergent behavior and side effects are undesirable). What works affirms the design. What does not work calls for corrections in the design. This is why multiple systems, especially complex systems, must be built and deployed over time and in different environments - to learn and understand the relations among the design, emergent behavior, side effects, and environment.

These two types of captured learning correspond respectively to patterns and “antipatternsantipatterns,” or patterns of failure, both of which are discussed in a broader context in the Principles of Systems Thinking and Patterns of Systems Thinking topics.

The use of iterations to refine the values of emergent properties, either across the life of a single system or through the development of patterns encapsulating knowledge gained from multiple developments, applies most easily to the discussion of strong emergence above. In this sense, those properties which can be observed but cannot be related to design choices are not relevant to a systems approach. However, they can have value when dealing with a combination of engineering and managed problems which occur for system of systems contexts (Sillitto 2010). (See Systems Approach Applied to Engineered Systems.)

References

Works Cited

Abbott, R. 2006. "Emergence explained: Getting epiphenomena to do real work". Complexity, vol. 12, no. 1 (September-October), pp. 13-26.

Bedau, M.A. and P. Humphreys, P. (eds.). 2008. "Emergence" In Contemporary Readings in Philosophy and Science. Cambridge, MA, USA: The MIT Press.

Boccara, N. 2004. Modeling Complex Systems. New York, NY, USA: Springer-Verlag.

Checkland, P. 1999. Systems Thinking, Systems Practice. New York, NY, USA: John Wiley & Sons.

Chroust. G. 2002. "Emergent properties in software systems." 10th Interdisciplinary Information Management Talks; Hofer, C. and Chroust, G. (eds.). Verlag Trauner Linz, pp. 277-289.

Chroust, G., C. Hofer, C. Hoyer (eds.). 2005. The concept of emergence in systems engineering." The 12th Fuschl Conversation, April 18-23, 2004, Institute for Systems Engineering and Automation, Johannes Kepler University Linz. pp. 49-60.

Emmeche, C., S. Koppe, and F. Stjernfelt. 1997. "Explaining emergence: Towards an ontology of levels." Journal for General Philosophy of Science, vol. 28, no. 1, pp. 83-119. Accessed 3 December 2014. Available at:http://www.nbi.dk/~emmeche/coPubl/97e.EKS/emerg.html.

Francois, C. 2004. International Encyclopedia of Systems and Cybernetics, 2nd edition, 2 volumes. Munich, Germany: K.G.Saur Verlag.

Hitchins, D. 2007. Systems Engineering: A 21st Century Systems Methodology. Hoboken, NJ, USA: John Wiley & Sons.

Honderich. T. 1995. The Oxford Companion to Philosophy. New York, NY, USA: Oxford University Press.

Hybertson, D. 2009. Model-Oriented Systems Engineering Science: A Unifying Framework for Traditional and Complex Systems. Boca Raton, FL, USA: Auerbach/CRC Press.

Jackson, S., D. Hitchins, and H. Eisner. 2010. "What is the Systems Approach?" INCOSE Insight. 13(1) (April 2010): 41-43.

O’Connor, T. and H. Wong. 2006. "Emergent Properties," in Stanford Encyclopedia of Philosophy. Accessed December 3 2014: Available at: http://plato.stanford.edu/entries/properties-emergent/.

Page, S.E. 2009. Understanding Complexity. The Great Courses. Chantilly, VA, USA: The Teaching Company.

Ryan, A. 2007. "Emergence is coupled to scope, not level." Complexity, vol. 13, no. 2, November-December.

Sillitto, H.G. 2010. "Design principles for ultra-large-scale systems". Proceedings of the 20th Annual International Council on Systems Engineering (INCOSE) International Symposium, July 2010, Chicago, IL, USA, reprinted in “The Singapore Engineer,” April 2011.

US-Canada Power System Outage Task Force. 2004. Final Report on the August 14, 2003 Blackout in the United States and Canada: Causes and Recommendations. Washington-Ottawa. Accessed 3 December 3, 2014. Available: http://energy.gov/oe/downloads/blackout-2003-final-report-august-14-2003-blackout-united-states-and-canada-causes-and

Primary References

Emmeche, C., S. Koppe, and F. Stjernfelt. 1997. "Explaining emergence: Towards an ontology of levels." Journal for General Philosophy of Science, vol. 28, no. 1, pp. 83-119. Available: http://www.nbi.dk/~emmeche/coPubl/97e.EKS/emerg.html.

Hitchins, D. 2007. Systems Engineering: A 21st Century Systems Methodology. Hoboken, NJ, USA: John Wiley & Sons.

Page, S. E. 2009. Understanding Complexity. The Great Courses. Chantilly, VA, USA: The Teaching Company.

Additional References

Sheard, S.A. and A. Mostashari. 2008. "Principles of complex systems for systems engineering." Systems Engineering, vol. 12, no. 4, pp. 295-311.


< Previous Article | Parent Article | Next Article >
SEBoK v. 2.6, released 13 May 2022