Assessment and Control
The purpose of systems engineering assessment and control (SEAC) is to provide adequate visibility into the project’s actual technical progress and risks with respect to the technical plans (i.e., systems engineering management plan (SEMP) or Systems Engineering Plan (SEP), and subordinate plans). The visibility allows the project team to take timely preventive action when disruptive trends are recognized or corrective action when performance deviates beyond established thresholds or expected values. SEAC includes preparing for and conducting reviews and audits to monitor performance. The results of the reviews and measurement analyses are used to identify and record findings/discrepancies and may lead to causal analysis and corrective/preventive action plans. Action plans are implemented, tracked, and monitored to closure. (NASA 2007, Section 6.7; SEG-ITS, 2009, Section 3.9.3, 3.9.10; INCOSE, 2010, Clause 6.2; SEI, 2007)
Systems Engineering Assessment and Control Process Overview
The SEAC process includes determining and initiating appropriate handling strategies and actions for findings and/or discrepancies that are uncovered in the enterprise, infrastructure, or life cycle activities associated with the project. Analysis of the causes of the findings/discrepancies aids in the determination of appropriate handling strategies. Implementation of approved preventive, corrective, or improvement actions ensures satisfactory completion of the project within planned technical, schedule, and cost objectives. Potential action plans for findings and/or discrepancies are reviewed in the context of the overall set of actions and priorities in order to optimize the benefits to the project and/or organization. Interrelated items are analyzed together to obtain a consistent and cost effective resolution.
The SEAC process includes the following steps:
- Monitor and review technical performance and resource use against plans
- Monitor technical risk, escalate significant risks to the project risk register and seek project funding to execute risk mitigation plans
- Hold technical reviews and report outcomes at the project reviews
- Analyze issues and determine appropriate actions
- Manage actions to closure
- Hold a post-delivery assessment (also known as a post-project review) to capture knowledge associated with the project (this may be a separate technical assessment or it may be conducted as part of the project assessment and control process).
The following activities are normally conducted as part of a project assessment and control process:
- Authorization, release and closure of work
- Monitor project performance and resource usage against plan
- Monitor project risk and authorize expenditure of project funds to execute risk mitigation plans
- Hold project reviews
- Analyze issues and determine appropriate actions
- Manage actions to closure
- Hold a post-delivery assessment (also known as a post-project review) to capture knowledge associated with the project
Examples of major technical reviews used in SEAC are shown in Table 1 from DAU (2010).
Name | Description |
---|---|
Alternative Systems Review |
A multi-disciplined review to ensure the resulting set of requirements agrees with the customers' needs and expectations. |
Critical Design Review (CDR) |
A multi-disciplined review establishing the initial product baseline to ensure that the system under review has a reasonable expectation of satisfying the requirements of the capability development document within the currently allocated budget and schedule. |
Functional Configuration Audit |
Formal examination of the as tested characteristics of a configuration item (hardware and software) with the objective of verifying that actual performance complies with design and interface requirements in the functional baseline. |
In-Service Review |
A multi-disciplined product and process assessment to ensure that the system under review is operationally employed with well-understood and managed risk. |
Initial Technical Review |
A multi-disciplined review to support a program's initial program objective memorandum submission. |
Integrated Baseline Review |
A joint assessment conducted by the government program manager and the contractor to establish the performance measurement baseline. |
Operational Test Readiness Review |
A multi-disciplined product and process assessment to ensure that the system can proceed into initial operational test and evaluation with a high probability of success, and that the system is effective and suitable for service introduction. |
Production Readiness Review (PRR) |
Examines a program to determine if the design is ready for production and if the prime contractor and major subcontractors have accomplished adequate production planning without incurring unacceptable risks that will breach thresholds of schedule, performance, cost, or other established criteria. |
Physical Configuration Audit |
Examines the actual configuration of an item being produced around the time of the full-rate production decision. |
Preliminary Design Review (PDR) |
A technical assessment establishing the physically allocated baseline to ensure that the system under review has a reasonable expectation of being judged operationally effective and suitable. |
System Functional Review (SFR) |
A multi-disciplined review to ensure that the system's functional baseline is established and has a reasonable expectation of satisfying the requirements of the initial capabilities document or draft capability development document within the currently allocated budget and schedule. |
System Requirements Review (SRR) |
A multi-disciplined review to ensure that the system under review can proceed into initial systems development, and that all system requirements and performance requirements derived from the initial capabilities document or draft capability development document are defined and testable, and are consistent with cost, schedule, risk, technology readiness, and other system constraints. |
System Verification Review (SVR) |
A multi-disciplined product and process assessment to ensure the system under review can proceed into low-rate initial production and full-rate production within cost (program budget), schedule (program schedule), risk, and other system constraints. |
Technology Readiness Assessment |
A systematic, metrics-based process that assesses the maturity of critical technology elements, including sustainment drivers. |
Test Readiness Review (TRR) |
A multi-disciplined review designed to ensure that the subsystem or system under review is ready to proceed into formal test. |
Linkages to Other Systems Engineering Management Topics
The SE assessment and control process is closely coupled with the measurement, planning, decision management, and risk management processes. The measurement process provides indicators for comparing actuals to plans. planning provides estimates and milestones that constitute plans for monitoring, and the project plan with measures used to monitor progress. Decision management uses the results of project monitoring as decision criteria for making control decisions.
Practical Considerations
Key pitfalls and good practices related to SEAC are described in the next two sections.
Pitfalls
Some of the key pitfalls encountered in planning and performing SE Assessment and Control are shown in Table 2.
Name | Description |
---|---|
No measurement | Since the assessment and control activities are highly dependent on insightful measurement information, it is usually ineffective to proceed independent of the measurement efforts - what you get is what you measure. |
"Something in Time" Culture | Some things are easier to measure than others - for instance, delivery to cost and schedule. Don't focus on these and neglect harder things to measure like quality of the system, Avoid a "something in time" culture where meeting the schedule takes priority over everything else, but what is delivered is not fit for purpose and drives rework into the project. |
No Teeth | Make sure that the technical review gates have "teeth". Sometimes the project manager is given authority (or can appeal to someone with authority) to over-ride a gate decision and allow work to proceed, even when the gate has exposed significant issues with the technical quality of the system or associated work products. This is a major risk if the organization is strongly schedule-driven; it can't afford the time to do it right, but somehow it finds the time to do it again (rework). |
Too Early Baselining | Don't baseline requirements or designs too early. Often there is strong pressure to baseline system requirements and designs before they are fully understood or agreed, in order to start subsystem or component development. This just guarantees high levels of rework. |
Good Practices
Some good practices gathered from the references are shown in Table 3.
Name | Description |
---|---|
Independence | Provide independent (from customer) assessment and recommendations on resources, schedule, technical status, and risk based on experience and trend analysis. |
Peer Reviews | Use peer review to ensure the quality of work products before they are submitted for gate review |
Accept Uncertainty | Communicate uncertainties in requirements or designs and accept that uncertainty is a normal part of developing a system. |
Risk Mitigation Plans | Do not penalize a project at gate review if they admit uncertainty in requirements - ask for their risk mitigation plan to manage the uncertainty. |
Just In-Time Baselining | Baseline requirements and designs only when you need to - when other work is committed based on the stability of the requirement or design. If work has to start and the requirement or design is still uncertain, consider how you can build robustness into the system to handle the uncertainty with minimum rework. |
Communication | Document and communicate status findings and recommendations to stakeholders. |
Full Visibility | Ensure that action items and action-item status, as well as other key status items, are visible to all project participants. |
Leverage Previous Root Cause Analysis | When performing root cause analysis, consider root cause and resolution data documented in previous related findings/discrepancies. |
Concurrent Management | Plan and perform Assessment and Control concurrently with the activities for Measurement and Risk Management. |
Lessons Learned and Post-Mortems | Hold post delivery assessments or post project reviews to capture knowledge associated with the project - for instance, to augment and improve estimation models, lessons learned databases, gate review checklists. |
Additional good practices can be found in (INCOSE 2010, Clause 6.2), (SEG-ITS, 2009, Sections 3.9.3 and 3.9.10), (INCOSE 2010, Section 5.2.1.5), (NASA 2007, Section 6.7).
References
Works Cited
Caltrans and USDOT. 2005. Systems Engineering Guidebook for Intelligent Transportation Systems (ITS). Version 1.1. Sacramento, CA, USA: California Department of Transportation (Caltrans) Division of Reserach & Innovation/U.S. Department of Transportation (USDOT), SEG for ITS 1.1.
DAU. 2010. Defense Acquisition Guidebook (DAG). Ft. Belvoir, VA, USA: Defense Acquisition University (DAU)/U.S. Department of Defense (DoD). February 19, 2010.
INCOSE. 2011. Systems Engineering Handbook: A Guide for System Life Cycle Processes and Activities. Version 3.2.1. San Diego, CA, USA: International Council on Systems Engineering (INCOSE), INCOSE-TP-2003-002-03.2.1.
NASA. 2007. Systems Engineering Handbook. Washington, DC, USA: National Aeronautics and Space Administration (NASA), December 2007. NASA/SP-2007-6105.
SEI. 2007. "Measurement and Analysis Process Area," in Capability Maturity Model Integrated (CMMI) for Development. Version 1.2. Pittsburgh, PA, USA: Software Engineering Institute (SEI)/Carnegie Mellon University (CMU).
Primary References
Caltrans and USDOT. 2005. Systems Engineering Guidebook for Intelligent Transportation Systems (ITS). Version 1.1. Sacramento, CA, USA: California Department of Transportation (Caltrans) Division of Reserach & Innovation/U.S. Department of Transportation (USDOT), SEG for ITS 1.1.
DAU. 2010. Defense Acquisition Guidebook (DAG). Ft. Belvoir, VA, USA: Defense Acquisition University (DAU)/U.S. Department of Defense (DoD). February 19, 2010.
IINCOSE. 2011. Systems Engineering Handbook: A Guide for System Life Cycle Processes and Activities. Version 3.2.1. San Diego, CA, USA: International Council on Systems Engineering (INCOSE), INCOSE-TP-2003-002-03.2.1.
NASA. 2007. Systems Engineering Handbook. Washington, DC, USA: National Aeronautics and Space Administration (NASA), December 2007. NASA/SP-2007-6105
SEI. 2007. "Measurement and Analysis Process Area," in Capability Maturity Model Integrated (CMMI) for Development. Version 1.2. Pittsburgh, PA, USA: Software Engineering Institute (SEI)/Carnegie Mellon University (CMU).
Additional References
ISO/IEC/IEEE. 2009. Systems and Software Engineering - Life Cycle Processes - Project Management. Geneva, Switzerland: International Organization for Standardization (ISO)/International Electronical Commission (IEC)/Institute of Electrical and Electronics Engineers (IEEE), ISO/IEC/IEEE 16326:2009(E).
SEBoK Discussion
Please provide your comments and feedback on the SEBoK below. You will need to log in to DISQUS using an existing account (e.g. Yahoo, Google, Facebook, Twitter, etc.) or create a DISQUS account. Simply type your comment in the text field below and DISQUS will guide you through the login or registration steps. Feedback will be archived and used for future updates to the SEBoK. If you provided a comment that is no longer listed, that comment has been adjudicated. You can view adjudication for comments submitted prior to SEBoK v. 1.0 at SEBoK Review and Adjudication. Later comments are addressed and changes are summarized in the Letter from the Editor and Acknowledgements and Release History.
If you would like to provide edits on this article, recommend new content, or make comments on the SEBoK as a whole, please see the SEBoK Sandbox.
blog comments powered by Disqus