Difference between revisions of "System Analysis"

From SEBoK
Jump to navigation Jump to search
(No difference)

Revision as of 20:58, 9 August 2011

Introduction, Definition and Purpose of System Analysis

System Analysis allows the developers of systems to carry out (in the most objective way possible) assessments of engineering data in order to select the most efficient system architecture , and also to provide a set of consistent and accurate engineering data including System Requirements.

During engineering, assessments should be performed every time technical choices or decisions have to be made or justified, not only to compare different design architectures, but taking into account the System Requirements. System Analysis provides a rigorous approach to technical decision-making. It is used to perform trade-off studies, including a set of analysis such as cost analysis, technical risks analysis, and effectiveness analysis. “Assess and select” is one of the major tasks of the system engineer.



Principles about System Analysis

One of the major development issues is to ensure that the global set of engineering data of a System of Interest (mainly Stakeholder Requirements, System Requirements, Design Properties) the engineers have created is consistent and that the values of data are relevant. Trade-off studies are at the centre of System Analysis, focusing on these aspects and providing means and techniques:

  • to define assessment criteria based on System Requirements;
  • to assess design properties of each candidate solution in comparison to these criteria;
  • to score globally the candidate solutions, and to justify the scores.


The number and importance of Assessment Criteria to be assessed depend on the considered type of system and on its context of operational use. The various assessments use theoretical models, representation models, mock-ups, simulations, analytical models, etc.


Trade-off studies

In the context of the definition of a system, a trade-off study consists of comparing the characteristics of each candidate solution to determine the solution that best globally addresses the Assessment Criteria. The various characteristics analyzed are gathered in Costs Analysis, Technical Risks Analysis, and Effectiveness Analysis (NASA. 2007) page 1-360. Each class of analysis is the subject of the following topics:

  1. assessment criteria are used to classify the various candidate solutions between themselves. They are absolute or relative. Example: maximum cost per unit produced is cc$; cost reduction shall be x%; effectiveness improvement is y%, risk mitigation is z%.
  2. Boundaries identify and limit the characteristics or criteria to be taken into account in the analysis. Examples: kind of costs to be taken into account; acceptable technical risks; type and level of effectiveness.
  3. Scales are used to quantify the characteristics or properties or criteria and to make comparisons. Their definition requires knowing the highest and lowest limits as well as the type of evolution of the characteristic (linear, logarithmic, etc.).
  4. An assessment score is assigned to a characteristic or criterion for each candidate solution. The goal of the trade-off study is to succeed in quantifying the three variables (and their decomposition in sub variables) of cost, risk, and effectiveness for each candidate solution. This operation is generally complex and requires the use of models.
  5. The optimization of the characteristics or properties improves the scoring of interesting solutions.


A decision-making process is not an accurate science and trade-off studies have limits. The following concerns shall be taken into account:

  • Subjective variables: for example, the component has to be beautiful. What is a beautiful component?
  • Uncertain data: for example, inflation has to be taken into account to estimate the cost of maintenance during the complete life cycle. What will be inflation for the next 5 years?
  • Sensitivity analysis: a global assessment score associated to every candidate solution is not absolute; it is recommended to get a robust selection by performing sensitivity analysis that consists to stimulate the decision model with small variations of assessment criteria values (weights). The selection is robust if the variations do not change the order of scores.


A relevant trade-off study specifies the assumptions, variables, and confidence intervals of the results. Knowledge management and a former experience that result in the building of significant databases and models that are able to demonstrate relevance and efficiency are a considerable advantage.

Costs Analysis

A cost analysis considers the full life cycle costs. The cost baseline can be adapted according to the project and the system. The global life cycle cost may includes example labor and non-labor cost items such as indicated in the Table 1.


Table 1 - Types of costs NEW TABLE


Methods for determining cost are described in the Planning topic.

Technical Risks Analysis

Every risk analysis concerning every domain is based on three things:

  • Analysis of potential threats or undesired events and their probability of occurrence.
  • Analysis of the consequences of these threats or undesired events and their classification on a scale of gravity.
  • The installation of protections or preventions to reduce the probabilities of threats and/or the levels of harmful effect to acceptable values.


The technical risks appear when the system cannot satisfy the System Requirements any longer. The causes reside in the solution itself and/or in the requirements. They are expressed in the form of insufficient effectiveness and can have multiple causes; for example, incorrect assessment of the technological capabilities, failure of parts, breakdowns, breakage, obsolescence of equipment, parts, or software, weakness from the supplier (non-compliant parts, delay for supply, etc.), human factors (insufficient training, wrong tunings, error handling, unsuited procedures, malice), etc. Because some risks come from context modifications, the project or the company cannot react: natural events (storm, flood, etc.), decisions or political events, regulation and standardization evolutions, change of policy or strategy.

Note: Technical risks have not to be confused with project risks even if the method to manage them is the same. Technical risks address the system itself, not the project for its development. Of course technical risks may react onto project risks. Technical risk can be managed on the same way that project risks all along the development till being satisfactorily mitigated.

See Risk Management for more details.


Effectiveness Analysis

Effectiveness studies use the different types of requirements as a starting point. The effectiveness of the system concerns several essential characteristics that are generally gathered in the following list of analyses, including but not limited to: performance, usability, dependability, manufacturing, maintenance or support, environment, etc. These analyses highlight candidate solutions under various aspects. It is essential to establish a classification in order to limit the number of analysis to the really significant aspects. The main difficulties of the effectiveness analysis are to sort and select the right set of effectiveness aspects; for example, if the product is made for a single use, the maintainability and the capability of evolution will not be relevant criteria.



Process Approach - System Analysis

Purpose and principles of the approach

The system analysis process is used to: (1) provide a rigorous basis for technical decision making, resolution of requirement conflicts, and assessment of alternative physical solutions; (2) determine progress in satisfying System Requirements and derived requirements; (3) support risk management; and (4) ensure that decisions are made only after evaluating the cost, schedule, performance, and risk effects on the engineering or reengineering of the system. (ANSI/EIA. 1998)

This process is named "Decision Analysis Process" in (NASA. 2007) page 1-360: The Decision Analysis Process is used to help evaluate technical issues, alternatives, and their uncertainties to support decision-making. See Decision Management for more details.


The system analysis supports other system definition processes:

  • Stakeholder Requirements definition and System Requirements definition processes use system analysis to solve issues relating to conflicts among the set of requirements, in particular those related to costs, technical risks, and effectiveness (performances, operational conditions, and constraints). System Requirements subject to high risks or which would require different architectures are discussed.
  • The Architectural Design process uses it to assess characteristics or Design Properties of candidate Functional and Physical Architectures, providing arguments for selecting the most efficient one in terms of costs, technical risks, and effectiveness (example: performances, dependability, human factors, etc.).

Like any system definition process, the System Analysis Process is iterative. Each operation is carried out several times; each step improves the precision of analysis.


Activities of the Process

Major activities and tasks performed during this process include:

  1. Planning the trade-off studies.
    1. Determine the number of candidate solutions to analyze, the methods and procedures to be used, the expected results (objects to be selected: Functional Architecture/Scenario, Physical Architecture, System Element, etc.), the justification items.
    2. Schedule the analyses according to the availability of models, engineering data (System Requirements, Design Properties), skilled personnel, and procedures.
  2. Define the selection criteria model.
    1. Select the Assessment Criteria from non-functional requirements (performances, operational conditions, constraints, etc.).
    2. Sort and order the Assessment Criteria.
    3. Establish a scale of comparison for each Assessment Criterion and weigh every Assessment Criterion according to its level of relative importance with the others.
  3. Identify candidate solutions and related models, and data.
  4. Assess candidate solutions using previously defined methods or procedures.
    1. Carry out costs analysis, technical risks analysis, and effectiveness analysis placing every candidate solution on every Assessment Criterion comparison scale.
    2. Score every candidate solution as an Assessment Score.
  5. Provide results to the calling process: Assessment Criteria, comparison scales, solutions’ scores, Assessment Selection (glossary), and eventually recommendations and related arguments.


Artifacts and Ontology Elements

This process may create several artifacts such as:

  1. Selection criteria model (list, scales, weighing)
  2. Costs, risks, effectiveness analysis reports
  3. Justification reports


This process handles the ontology elements of Table 2 within System Analysis.

Main Ontology Elements as Handled within System Analysis

Table 2. Main ontology elements as handled within System Analysis


The main relationships between ontology elements of System Analysis are presented in Figure 1.

System Analysis Elements Relationships with Other Engineering Elements

Figure 1. System Analysis elements relationships with other engineering elements (Faisandier, 2011)

Checking and correctness of System Analysis

TO BE WRITTEN

ADD ACTION RELATED TO COMMENT 408


Methods and Modeling Techniques

  • General usage of models: Various types of models can be used in the context of System Analysis:
    • Physical models are scale models allowing to simulate physical phenomena; they are specific to each discipline; associated tools are for example mocks-up, vibration tables, test benches, prototypes, decompression chamber, wind tunnels, etc.
    • Representation models are mainly used to simulate the behavior of a system; for example Enhanced Functional Flow Block Diagrams (EFFBD), Statecharts, State machine diagram (SysML), etc.
    • Analytical models are mainly used to establish values of estimates, and we can consider the deterministic models and probabilistic models (also known as stochastic models). Analytical models use equations or diagrams to approach the real operation of the system. They can be from simplest (addition) to most complicated (probabilistic distribution with several variables).


  • Use right models depending of the project progress:
    • At the beginning of the project, first studies use simple tools, allowing rough approximations which have the advantage of not requiring too much time and effort; these approximations are often sufficient to eliminate unrealistic or outgoing candidate solutions.
    • Progressively with the progress of the project it is necessary to improve precision of data to compare the candidate solutions still competing. The work is more complicated if the level of innovation is high.
    • A system engineer alone cannot model a complex system: he has to be supported by skilled people from different disciplines involved.


  • Specialist expertise: When the values of Assessment Criteria cannot be given in an objective or precise way, or because the subjective aspect is dominating, we can ask to specialists for expertise. The estimates proceeds in four steps:
    • Select interviewees to collect the opinion of qualified people for the considered field.
    • Draft a questionnaire; a precise questionnaire allows an easy analysis, but a questionnaire too closed risks to neglect significant points.
    • Interview a limited number of specialists with the questionnaire and having an in-depth discussion to get precise opinions.
    • Analyze the data with several different people comparing their impressions until agreement on a classification of Assessment Criteria and/or candidate solutions.


Often used analytical models in the context of System Analysis are summarized in Table 3.


Table 3 - Often used analytical models in the context of System Analysis NEW TABLE



Application to Product systems, Service systems, Enterprise systems

TO BE WRITTEN



Practical Considerations about System Analysis

Major pitfalls encountered with System Analysis are presented in Table 4.

Pitfalls with System Analysis

Table 4 – Pitfalls with System Analysis


Proven practices with System Analysis are presented in Table 5.

Proven practices with System Analysis

Table 5 – Proven practices with System Analysis



References

Citations

NASA. 2007. Systems engineering handbook. Washington, D.C.: National Aeronautics and Space Administration (NASA), NASA/SP-2007-6105.


ANSI/EIA. 1998. Processes for engineering a system. Philadelphia, PA, USA: American National Standards Institute (ANSI)/Electronic Industries Association (EIA), ANSI/EIA-632-1998.

Primary References

ANSI/EIA. 1998. Processes for engineering a system. Philadelphia, PA, USA: American National Standards Institute (ANSI)/Electronic Industries Association (EIA), ANSI/EIA-632-1998.


NASA. 2007. Systems engineering handbook. Washington, D.C.: National Aeronautics and Space Administration (NASA), NASA/SP-2007-6105.

Additional References

Faisandier, A. 2011. Engineering and architecting multidisciplinary systems. (expected--not yet published).


Article Discussion

[Go to discussion page]

<- Previous Article | Parent Article | Next Article ->>/center>

Signatures