Difference between revisions of "Principles of Systems Thinking"

From SEBoK
Jump to navigation Jump to search
m (Text replacement - "{{DISQUS}}" to "SEBoK v. 1.9.1 released 5 October 2018")
Line 223: Line 223:
 
----
 
----
 
<center> [[Concepts of Systems Thinking|< Previous Article]]  |  [[Systems Thinking|Parent Article]]  |  [[Patterns of Systems Thinking|Next Article >]] </center>
 
<center> [[Concepts of Systems Thinking|< Previous Article]]  |  [[Systems Thinking|Parent Article]]  |  [[Patterns of Systems Thinking|Next Article >]] </center>
 +
 +
<center>'''SEBoK v. 1.9.1, released 16 October 2018'''</center>
  
 
[[Category:Part 2]][[Category:Topic]][[Category:Systems Thinking]]
 
[[Category:Part 2]][[Category:Topic]][[Category:Systems Thinking]]
SEBoK v. 1.9.1 released 5 October 2018
 

Revision as of 03:59, 16 October 2018

This topic forms part of the Systems Thinking knowledge area (KA). It identifies systems principles as part of the basic ideas of systems thinking.

Some additional concepts more directly associated with engineered systems are described, and a summary of system principles associated with the concepts already defined is provided. A number of additional “laws” and heuristics are also discussed.

Systems Principles, Laws, and Heuristics

A principle is a general rule of conduct or behavior (Lawson and Martin 2008). It can also be defined as a basic generalization that is accepted as true and that can be used as a basis for reasoning or conduct (WordWeb 2012c). Thus, systems principles can be used as a basis for reasoning about systems thinking or associated conduct (systems approaches).

Separation of Concerns

A systems approach is focused on a systems-of-interest (SoI) of an open system. This SoI consists of open, interacting subsystems that as a whole interact with and adapt to other systems in an environment. The systems approach also considers the SoI in its environment to be part of a larger, wider, or containing system (Hitchins 2009).

In the What is Systems Thinking? topic, a “systems thinking paradox” is discussed. How is it possible to take a holistic system view while still being able to focus on changing or creating systems?

Separation of concerns describes a balance between considering parts of a system problem or solution while not losing sight of the whole (Greer 2008). Abstraction is the process of taking away characteristics from something in order to reduce it to a set of base characteristics (SearchCIO 2012). In attempting to understand complex situations it is easier to focus on bounded problems, whose solutions still remain agnostic to the greater problem (Erl 2012). This process sounds reductionist, but it can be applied effectively to systems. The key to the success of this approach is ensuring that one of the selected problems is the concerns of the system as a whole. Finding balance between using abstraction to focus on specific concerns while ensuring we continue to consider the whole is at the center of systems approaches.

A view is a subset of information observed of one or more entities, such as systems. The physical or conceptual point from which a view is observed is the viewpoint, which can be motivated by one or more observer concerns. Different views of the same target must be both separate, to reflect separation of concerns, and integrated such that all views of a given target are consistent and form a coherent whole (Hybertson 2009). Some sample views of a system are internal (What does it consist of?); external (What are its properties and behavior as a whole?); static (What are its parts or structures?); and dynamic (interactions).

encapsulation , which encloses system elements and their interactions from the external environment, is discussed in Concepts of Systems Thinking. Encapsulation is associated with modularity, the degree to which a system's components may be separated and recombined (Griswold 1995). Modularity applies to systems in natural, social, and engineered domains. In engineering, encapsulation is the isolation of a system function within a module and provides precise specifications for the module (IEEE Std. 610.12-1990).

Dualism is a characteristic of systems in which they exhibit seemingly contradictory characteristics that are important for the system (Hybertson 2009). The yin yang concept in Chinese philosophy emphasizes the interaction between dual elements and their harmonization, ensuring a constant dynamic balance through a cyclic dominance of one element and then the other, such as day and night (IEP 2006).

From a systems perspective the interaction, harmonization, and balance between system properties is important. Hybertson (Hybertson 2009) defines leverage as the duality between

  • Power, the extent to which a system solves a specific problem, and
  • Generality, the extent to which a system solves a whole class of problems.

While some systems or elements may be optimized for one extreme of such dualities, a dynamic balance is needed to be effective in solving complex problems.

Summary of Systems Principles

A set of systems principles is given in Table 1 below. The "Names" segment points to concepts underlying the principle. (See Concepts of Systems Thinking). Following the table, two additional sets of items related to systems principles are noted and briefly discussed: prerequisite laws for design science, and heuristics and pragmatic principles.

Table 1. A Set of Systems Principles. (SEBoK Original)
Name Statement of Principle
Abstraction A focus on essential characteristics is important in problem solving because it allows problem solvers to ignore the nonessential, thus simplifying the problem. (Sci-Tech Encyclopedia 2009; SearchCIO 2012; Pearce 2012)
Boundary A boundary or membrane separates the system from the external world. It serves to concentrate interactions inside the system while allowing exchange with external systems. (Hoagland, Dodson, and Mauck 2001)
Change Change is necessary for growth and adaptation, and should be accepted and planned for as part of the natural order of things rather than something to be ignored, avoided, or prohibited (Bertalanffy 1968; Hybertson 2009).
Dualism Recognize dualities and consider how they are, or can be, harmonized in the context of a larger whole (Hybertson 2009)
Encapsulation Hide internal parts and their interactions from the external environment. (Klerer 1993; IEEE 1990)
Equifinality In open systems, the same final state may be reached from different initial conditions and in different ways (Bertalanffy 1968). This principle can be exploited, especially in systems of purposeful agents.
Holism A system should be considered as a single entity, a whole, not just as a set of parts. (Ackoff 1979; Klir 2001)
Interaction The properties, capabilities, and behavior of a system are derived from its parts, from interactions between those parts, and from interactions with other systems. (Hitchins 2009 p. 60)
Layer Hierarchy The evolution of complex systems is facilitated by their hierarchical structure (including stable intermediate forms) and the understanding of complex systems is facilitated by their hierarchical description. (Pattee 1973; Bertalanffy 1968; Simon 1996)
Leverage Achieve maximum leverage (Hybertson 2009). Because of the power versus generality tradeoff, leverage can be achieved by a complete solution (power) for a narrow class of problems, or by a partial solution for a broad class of problems (generality).
Modularity Unrelated parts of the system should be separated, and related parts of the system should be grouped together. (Griswold 1995; Wikipedia 2012a)
Network The network is a fundamental topology for systems that forms the basis of togetherness, connection, and dynamic interaction of parts that yield the behavior of complex systems (Lawson 2010; Martin et al. 2004; Sillitto 2010)
Parsimony One should choose the simplest explanation of a phenomenon, the one that requires the fewest assumptions (Cybernetics 2012). This applies not only to choosing a design, but also to operations and requirements.
Regularity Systems science should find and capture regularities in systems, because those regularities promote systems understanding and facilitate systems practice. (Bertalanffy 1968)
Relations A system is characterized by its relations: the interconnections between the elements. Feedback is a type of relation. The set of relations defines the network of the system. (Odum 1994)
Separation of Concerns A larger problem is more effectively solved when decomposed into a set of smaller problems or concerns. (Erl 2012; Greer 2008)
Similarity/ Difference Both the similarities and differences in systems should be recognized and accepted for what they are. (Bertalanffy 1975 p. 75; Hybertson 2009). Avoid forcing one size fits all, and avoid treating everything as entirely unique.
Stability/ Change Things change at different rates, and entities or concepts at the stable end of the spectrum can and should be used to provide a guiding context for rapidly changing entities at the volatile end of the spectrum (Hybertson 2009). The study of complex adaptive systems can give guidance to system behavior and design in changing environments (Holland 1992).
Synthesis Systems can be created by choosing (conceiving, designing, selecting) the right parts, bringing them together to interact in the right way, and in orchestrating those interactions to create requisite properties of the whole, such that it performs with optimum effectiveness in its operational environment, so solving the problem that prompted its creation” (Hitchins 2008: 120).
View Multiple views, each based on a system aspect or concern, are essential to understand a complex system or problem situation. One critical view is how concern relates to properties of the whole. (Edson 2008; Hybertson 2009)

The principles are not independent. They have synergies and tradeoffs. Lipson (Lipson 2007), for example, argued that “scalability of open-ended evolutionary processes depends on their ability to exploit functional modularity, structural regularity and hierarchy.” He proposed a formal model for examining the properties, dependencies, and tradeoffs among these principles. Edson (Edson 2008) related many of the above principles in a structure called the conceptagon, which he modified from the work of Boardman and Sauser (Boardman and Sauser 2008). Edson also provided guidance on how to apply these principles. Not all principles apply to every system or engineering decision. Judgment, experience, and heuristics (see below) provide understanding into which principles apply in a given situation.

Several principles illustrate the relation of view with the dualism and yin yang principle; for example, holism and separation of concerns. These principles appear to be contradictory but are in fact dual ways of dealing with complexity. Holism deals with complexity by focusing on the whole system, while separation of concerns divides a problem or system into smaller, more manageable elements that focus on particular concerns. They are reconciled by the fact that both views are needed to understand systems and to engineer systems; focusing on only one or the other does not give sufficient understanding or a good overall solution. This dualism is closely related to the systems thinking paradox described in What is Systems Thinking?.

Rosen (Rosen 1979) discussed “false dualisms” of systems paradigms that are considered incompatible but are in fact different aspects or views of reality. In the present context, they are thus reconcilable through yin yang harmonization. Edson (Edson 2008) emphasized viewpoints as an essential principle of systems thinking; specifically, as a way to understand opposing concepts.

Derick Hitchins (Hitchins 2003) produced a systems life cycle theory described by a set of seven principles forming an integrated set. This theory describes the creation, manipulation and demise of engineered systems. These principles consider the factors which contribute to the stability and survival of man made systems in an environment. Stability is associated with the principle of connected variety, in which stability is increased by variety plus the cohesion and adaptability of that variety. Stability is limited by allowable relations, resistance to change, and patterns of interaction. Hitchins describes how interconnected systems tend toward a cyclic progression, in which variety is generated, dominance emerges to suppress variety, dominant modes decay and collapse and survivors emerge to generate new variety.

Guidance on how to apply many of these principles to engineered systems is given in the topic Synthesizing Possible Solutions, as well as in System Definition and other knowledge areas in Part 3 of the SEBoK.

Prerequisite Laws of Design Science

John Warfield (Warfield 1994) identified a set of laws of generic design science that are related to systems principles. Three of these laws are stated here:

  1. ‘’Law of Requisite Variety’’: A design situation embodies a variety that must be matched by the specifications. The variety includes the diversity of stakeholders. This law is an application of the design science of the Ashby (1956) Law of Requisite Variety, which was defined in the context of cybernetics and states that to successfully regulate a system, the variety of the regulator must be at least as large as the variety of the regulated system.
  2. ‘’Law of Requisite Parsimony’’: Information must be organized and presented in a way that prevents human information overload. This law derives from Miller’s findings on the limits of human information processing capacity (Miller 1956). Warfield’s structured dialog method is one possible way to help achieve the requisite parsimony.
  3. ‘’Law of Gradation’’: Any conceptual body of knowledge can be graded in stages or varying degrees of complexity and scale, ranging from simplest to most comprehensive, and the degree of knowledge applied to any design situation should match the complexity and scale of the situation. A corollary, called the Law of Diminishing Returns, states that a body of knowledge should be applied to a design situation to the stage at which the point of diminishing returns is reached.

Heuristics and Pragmatic Principles

A heuristic is a common sense rule intended to increase the probability of solving some problem (WordWeb 2012b). In the present context it may be regarded as an informal or pragmatic principle. Maier and Rechtin (Maier and Rechtin 2000) identified an extensive set of heuristics that are related to systems principles. A few of these heuristics are stated here.

  • Relationships among the elements are what give systems their added value. This is related to the ‘’Interaction’’ principle.
  • Efficiency is inversely proportional to universality. This is related to the ‘’Leverage’’ principle.
  • The first line of defense against complexity is simplicity of design. This is related to the ‘’Parsimony’’ principle.
  • In order to understand anything, you must not try to understand everything (attributed to Aristotle). This is related to the ‘’Abstraction’’ principle.

An International Council on Systems Engineering (INCOSE) working group (INCOSE 1993) defined a set of “pragmatic principles” for systems engineering (SE). They are essentially best practice heuristics for engineering a system. For example:

  • Know the problem, the customer, and the consumer
  • Identify and assess alternatives so as to converge on a solution
  • Maintain the integrity of the system

Hitchins defines a set of SE principles which include principles of holism and synthesis as discussed above, as well as principles describing how systems problems should be resolved that are of particular relevance to a Systems Approach Applied to Engineered Systems (Hitchins 2009).

References

Works Cited

Ackoff, R. 1979. "The Future of Operational Research is Past." Journal of the Operational Research Society. 30(2): 93–104, Pergamon Press.

Ashby, W.R. 1956. "Requisite variety and its implications for the control of complex systems." Cybernetica. 1(2):1–17.

Bertalanffy, L. von. 1968. General System Theory: Foundations, Development, Applications. Revised ed. New York, NY: Braziller.

Bertalanffy, L. von. 1975. Perspectives on General System Theory. E. Taschdjian, ed. New York: George Braziller.

Boardman, J. and B. Sauser. 2008. Systems Thinking: Coping with 21st Century Problems. Boca Raton, FL: Taylor & Francis.

Cybernetics (Web Dictionary of Cybernetics and Systems). 2012. "Principle of Parsimony or Principle of Simplicity." Accessed December 3 2014 at Web Dictionary of Cybernetics and Systems http://pespmc1.vub.ac.be/ASC/PRINCI_SIMPL.html

Edson, R. 2008. Systems Thinking. Applied. A Primer. Arlington, VA, USA: Applied Systems Thinking (ASysT) Institute, Analytic Services Inc.

Erl, T. 2012. "SOA Principles: An Introduction to the Service Orientation Paradigm." Accessed December 3 2014 at Arcitura http://www.soaprinciples.com/p3.php

Greer, D. 2008. "The Art of Separation of Concerns." Accessed December 3 2014 at Aspiring Craftsman http://aspiringcraftsman.com/tag/separation-of-concerns/

Griswold, W. 1995. "Modularity Principle." Accessed December 3 2014 at William Griswold http://cseweb.ucsd.edu/users/wgg/CSE131B/Design/node1.html

Hitchins D. K. 2003. Advanced systems thinking engineering and management. Boston, MA, USA: Artech House.

Hitchins, D. 2009. "What are the General Principles Applicable to Systems?" INCOSE Insight. 12(4): 59-63.

Hoagland, M., B. Dodson, and J. Mauck. 2001. Exploring the Way Life Works. Burlington, MA, USA: Jones and Bartlett Publishers, Inc.

Holland, J. 1992. Adaptation in Natural and Artificial Systems: An Introductory Analysis with Applications to Biology, Control, and Artificial Intelligence. Cambridge, MA: MIT Press.

Hybertson, D. 2009. Model-Oriented Systems Engineering Science: A Unifying Framework for Traditional and Complex Systems. Boca Raton, FL, USA: Auerbach/CRC Press.

IEEE. 1990. IEEE Standard Glossary of Software Engineering Terminology. Geneva, Switzerland: Institute of Electrical and Electronics Engineers. IEEE Std 610.12-1990.

IEP (Internet Encyclopedia of Philosophy). 2006. "Yinyang (Yin-yang)." Accessed December 3 2014 at Internet Encyclopedia of Philosophy http://www.iep.utm.edu/yinyang/

INCOSE 1993. An Identification of Pragmatic Principles - Final Report. SE Principles Working Group, January 21, 1993.

Klerer, S. “System Management Information Modeling.” IEEE Communications. 31(5)May 1993: 38-44.

Klir, G. 2001. Facets of Systems Science, 2nd ed. New York, NY: Kluwer Academic/Plenum Publishers.

Lawson, H. 2010. A Journey Through the Systems Landscape. London, UK: College Publications, Kings College, UK.

Lawson, H. and J. Martin. 2008. "On the Use of Concepts and Principles for Improving Systems Engineering Practice." INCOSE International Symposium 2008, 15-19 June 2008, The Netherlands.

Lipson, H. 2007. "Principles of modularity, regularity, and hierarchy for scalable systems." Journal of Biological Physics and Chemistry. 7: 125–128.

Maier, M. and E. Rechtin. 2000. The Art of Systems Architecting, 2nd ed. Boca Raton, FL: CRC Press.

Miller, G. 1956. "The magical number seven, plus or minus two: some limits on our capacity for processing information." The Psychological Review. 63: 81–97.

Odum, H. 1994. Ecological and General Systems: An Introduction to Systems Ecology (Revised Edition). Boulder, CO: University Press of Colorado.

Pattee, H. (ed.) 1973. Hierarchy Theory: The Challenge of Complex Systems. New York, NY: George Braziller.

Pearce, J. 2012. "The Abstraction Principle." Posting date unknown. Accessed December 3 2014 at Jon Pearce, San Jose State University http://www.cs.sjsu.edu/~pearce/modules/lectures/ood/principles/Abstraction.htm

Rosen, R. 1979. "Old trends and new trends in general systems research." International Journal of General Systems. 5(3): 173-184.

Sci-Tech Encyclopedia. 2009. "Abstract Data Type." McGraw-Hill Concise Encyclopedia of Science and Technology, Sixth Edition, The McGraw-Hill Companies, Inc.

SearchCIO. 2012. "Abstraction." Accessed December 3 2014 at SearchCIO http://searchcio-midmarket.techtarget.com/definition/abstraction

Sillitto, H. 2010. "Design principles for Ultra-Large-Scale (ULS) Systems." 'Proceedings of INCOSE International Symposium 2010', 12-15 July 2010, Chicago, IL.

Simon, H. 1996. The Sciences of the Artificial, 3rd ed. Cambridge, MA: MIT Press.

Warfield, J.N. 1994. A Science of Generic Design. Ames, IA: Iowa State University Press.

Wikipedia. 2012a. "Modularity." Accessed December 3 2014 at Wikipedia http://en.wikipedia.org/wiki/Modularity

WordWeb. 2012b. "Dualism." Accessed December 3 2014 at WordWeb http://www.wordwebonline.com/en/DUALISM.

WordWeb. 2012c. "Heuristic." Accessed December 3 2014 at WordWeb http://www.wordwebonline.com/en/HEURISTIC.

WordWeb. 2012d. "Principle." Accessed December 3 2014 at WordWeb http://www.wordwebonline.com/en/PRINCIPLE.

Primary References

Bertalanffy, L. von. 1968. General System Theory: Foundations, Development, Applications. Revised ed. New York, NY: Braziller.

Hybertson, D. 2009. Model-Oriented Systems Engineering Science: A Unifying Framework for Traditional and Complex Systems. Auerbach/CRC Press, Boca Raton, FL.

Klir, G. 2001. Facets of Systems Science, 2nd ed. New York: Kluwer Academic/Plenum Publishers.

Additional References

Francois, F. (ed.). 2004. International Encyclopedia of Systems and Cybernetics, 2nd ed. Munich, Germany: K. G. Saur Verlag.

Meyers, R. (ed.). 2009. Encyclopedia of Complexity and Systems Science. New York, NY: Springer.

Midgley, G. (ed.). 2003. Systems Thinking. Thousand Oaks, CA: Sage Publications Ltd.

Volk, T., and J.W. Bloom. (2007). "The use of metapatterns for research into complex systems of teaching, learning, and schooling. Part I: Metapatterns in nature and culture." Complicity: An International Journal of Complexity and Education. 4(1): 25—43.


< Previous Article | Parent Article | Next Article >
SEBoK v. 1.9.1, released 16 October 2018