Difference between revisions of "Assessment and Control"

From SEBoK
Jump to navigation Jump to search
Line 248: Line 248:
  
 
--[[User:Dholwell|Dholwell]] 14:32, 2 September 2011 (UTC) core edit
 
--[[User:Dholwell|Dholwell]] 14:32, 2 September 2011 (UTC) core edit
 +
 +
--[[User:Rturner|Rturner]] 13:08, 9 September 2011 (UTC) tech edit
 
[[Category: Part 3]][[Category:Topic]]
 
[[Category: Part 3]][[Category:Topic]]

Revision as of 13:08, 9 September 2011

The purpose of Systems Engineering Assessment and Control (SEAC) is to provide adequate visibility into the project’s actual technical progress and risks with respect to the technical plans (i.e., Systems Engineering Management Plan (SEMP) and subordinate plans). The visibility allows the project team to take timely preventive action when trends are recognized or corrective action when performance deviates beyond established thresholds or expected values. SEAC includes preparing for and conducting reviews and audits to monitor performance. The results of the reviews and measurement analyses are used to identify and record findings/discrepancies and may lead to causal analysis and corrective/preventive action plans. Action plans are implemented, tracked, and monitored to closure. (NASA 2007, Section 6.7) (SEG-ITS, 2009, Section 3.9.3, 3.9.10) (INCOSE, 2010, Clause 6.2) (SEI, 2007)

SE Assessment and Control Process Overview

The Systems Engineering Assessment and Control process includes determining and initiating appropriate handling strategies and actions for findings and/or discrepancies that are uncovered in the enterprise, infrastructure, or life cycle activities associated with the project. Analysis of the causes of the findings/discrepancies aids in the determination of appropriate handling strategies. Implementation of approved preventive, corrective, or improvement actions ensures satisfactory completion of the project within planned technical, schedule, and cost objectives. Potential action plans for findings and/or discrepancies are reviewed in the context of the overall set of actions and priorities in order to optimize the benefits to the project and/or organization. Interrelated items are analyzed together to obtain a consistent and cost effective resolution.

The SE assessment and control process includes the following activities:

  • Monitor and review technical performance and resource usage against plan
  • Monitor technical risk, escalate significant risks to the project risk register and seek project funding to execute risk mitigation plans
  • Hold technical reviews and report outcomes at the project reviews
  • Analyze issues and determine appropriate actions
  • Manage actions to closure
  • Hold a Post Delivery Assessment (also known as a Post Project Review) to capture knowledge associated with the project (this may be a separate technical assessment or it may be conducted as part of the Project Assessment and Control process).

The following activities are normally conducted as part of a Project Assessment and Control process

  • Authorization, release and closure of work
  • Monitor project performance and resource usage against plan
  • Monitor project risk and authorize expenditure of project funds to execute risk mitigation plans
  • Hold Project reviews
  • Analyze issues and determine appropriate actions
  • Manage actions to closure
  • Hold a Post Delivery Assessment (also known as a Post Project Review) to capture knowledge associated with the Project

Examples of major technical reviews used in SEAC from (DAU, 2010) include:

Major Technical Reviews
Name Description
Alternative Systems Review

A multi-disciplined review to ensure the resulting set of requirements agrees with the customers' needs and expectations.

Critical Design Review (CDR)

A multi-disciplined review establishing the initial product baseline to ensure that the system under review has a reasonable expectation of satisfying the requirements of the Capability Development Document within the currently allocated budget and schedule.

Functional Configuration Audit

formal examination of the as tested characteristics of a configuration item (hardware and software) with the objective of verifying that actual performance complies with design and interface requirements in the functional baseline.

In-Service Review

A multi-disciplined product and process assessment to ensure that the system under review is operationally employed with well-understood and managed risk.

Initial Technical Review

A multi-disciplined review to support a program's initial Program Objective Memorandum submission.

Integrated Baseline Review

A joint assessment conducted by the government program manager and the contractor to establish the Performance Measurement Baseline.

Operational Test Readiness Review

A multi-disciplined product and process assessment to ensure that the system can proceed into Initial Operational Test and Evaluation with a high probability of success, and that the system is effective and suitable for service introduction.

Production Readiness Review (PRR)

Examines a program to determine if the design is ready for production and if the prime contractor and major subcontractors have accomplished adequate production planning without incurring unacceptable risks that will breach thresholds of schedule, performance, cost, or other established criteria.

Physical Configuration Audit

Examines the actual configuration of an item being produced around the time of the Full-Rate Production Decision.

Preliminary Design Review (PDR)

A technical assessment establishing the physically allocated baseline to ensure that the system under review has a reasonable expectation of being judged operationally effective and suitable.

System Functional Review (SFR)

A multi-disciplined review to ensure that the system's functional baseline is established and has a reasonable expectation of satisfying the requirements of the Initial Capabilities Document or draft Capability Development Document within the currently allocated budget and schedule.

System Requirements Review (SRR)

A multi-disciplined review to ensure that the system under review can proceed into initial systems development, and that all system requirements and performance requirements derived from the Initial Capabilities Document or draft Capability Development Document are defined and testable, and are consistent with cost, schedule, risk, technology readiness, and other system constraints.

System Verification Review (SVR)

A multi-disciplined product and process assessment to ensure the system under review can proceed into Low-Rate Initial Production and full-rate production within cost (program budget), schedule (program schedule), risk, and other system constraints.

Technology Readiness Assessment

A systematic, metrics-based process that assesses the maturity of critical technology elements, including sustainment drivers.

Test Readiness Review (TRR)

A multi-disciplined review designed to ensure that the subsystem or system under review is ready to proceed into formal test.

Linkages to Other Systems Engineering Management Topics

The Systems Engineering assessment and control process is closely coupled with the Measurement, Planning, Decision Management, and Risk Management processes. The Measurement process provides indicators for comparing actuals to plans. Planning provides estimates and milestones that constitute plans for monitoring, and the project plan with measures used to monitor progress. Decision Management uses the results of project monitoring as decision criteria for making control decisions.

Practical Considerations

Key pitfalls and good practices related to SEAC are described in the next two sections.

Pitfalls

Some of the key pitfalls encountered in planning and performing SE Assessment and Control are:

Name Description
No Measurement
  • Since the assessment and control activities are highly dependent on insightful measurement information, it is usually ineffective to proceed independent of the measurement efforts - what you get is what you measure.
"Something in Time" Culture
  • Some things are easier to measure than others - for instance, delivery to cost and schedule. Don't focus on these and neglect harder things to measure like quality of the system, Avoid a "something in time" culture where meeting the schedule takes priority over everything else, but what is delivered is not fit for purpose and drives rework into the project.
No Teeth
  • Make sure that the technical review gates have "teeth". Sometimes the project manager is given authority (or can appeal to someone with authority) to over-ride a gate decision and allow work to proceed, even when the gate has exposed significant issues with the technical quality of the system or associated work products. This is a major risk if the organization is strongly schedule-driven; it can't afford the time to do it right, but somehow it finds the time to do it again (rework).
Too Early Baselining
  • Don't baseline requirements or designs too early. Often there is strong pressure to baseline system requirements and designs before they are fully understood or agreed, in order to start subsystem or component development. This just guarantees high levels of rework.

Good Practices

Some good practices gathered from the references are:

Name Description
Independence
  • Provide independent (from customer) assessment and recommendations on resources, schedule, technical status, and risk based on experience and trend analysis.
Peer Reviews
  • Use peer review to ensure the quality of work products before they are submitted for gate review
Accept Uncertainty
  • Communicate uncertainties in requirements or designs and accept that uncertainty is a normal part of developing a system.
Risk Mitigation Plans
  • Do not penalize a project at gate review if they admit uncertainty in requirements - ask for their risk mitigation plan to manage the uncertainty.
Just In-Time Baselining
  • Baseline requirements and designs only when you need to - when other work is committed based on the stability of the requirement or design. If work has to start and the requirement or design is still uncertain, consider how you can build robustness into the system to handle the uncertainty with minimum rework.
Communication
  • Document and communicate status findings and recommendations to stakeholders.
Full Visibility
  • Ensure that action items and action-item status, as well as other key status items, are visible to all project participants.
Leverage Previous Root Cause Analysis
  • When performing root cause analysis, consider root cause and resolution data documented in previous related findings/discrepancies.
Concurrent Management
Lessons Learned and Post-Mortems
  • Hold post delivery assessments or post project reviews to capture knowledge associated with the project - for instance, to augment and improve estimation models, lessons learned databases, gate review checklists.


Additional good practices can be found in (INCOSE 2010, Clause 6.2), (SEG-ITS, 2009, Sections 3.9.3 and 3.9.10), (INCOSE, 2010, Section 5.2.1.5), (NASA, 2007, Section 6.7).

References

Citations

Caltrans, and USDOT. 2005. Systems Engineering Guidebook for Intelligent Transportation Systems (ITS), version 1.1. Sacramento, CA, USA: California Department of Transportation (Caltrans) Division of Reserach & Innovation/U.S. Department of Transportation (USDOT), SEG for ITS 1.1.

DAU. February 19, 2010. Defense Acquisition Guidebook (DAG). Ft. Belvoir, VA, USA: Defense Acquisition University (DAU)/U.S. Department of Defense.

INCOSE. 2011. INCOSE Systems Engineering Handbook: A Guide for System Life Cycle Processes and Activities, version 3.2.1. San Diego, CA, USA: International Council on Systems Engineering (INCOSE), INCOSE-TP-2003-002-03.2.1.

NASA. December 2007. NASA Systems Engineering Handbook. Washington, D.C.: National Aeronautics and Space Administration (NASA), NASA/SP-2007-6105.

SEI. 2007. Capability Maturity Model Integrated (CMMI) for Development, version 1.2, measurement and analysis process area. Pittsburgh, PA, USA: Software Engineering Institute (SEI)/Carnegie Mellon University (CMU).

Primary References

DAU. February 19, 2010. Defense Acquisition Guidebook (DAG). Ft. Belvoir, VA, USA: Defense Acquisition University (DAU)/U.S. Department of Defense.

INCOSE. 2011. INCOSE Systems Engineering Handbook: A Guide for System Life Cycle Processes and Activities, version 3.2.1. San Diego, CA, USA: International Council on Systems Engineering (INCOSE), INCOSE-TP-2003-002-03.2.1.

NASA. December 2007. NASA Systems Engineering Handbook. Washington, D.C., USA: National Aeronautics and Space Administration (NASA), NASA/SP-2007-6105.

SEI. 2007. Capability Maturity Model Integrated (CMMI) for Development, version 1.2, measurement and analysis process area. Pittsburgh, PA, USA: Software Engineering Institute (SEI)/Carnegie Mellon University (CMU).

Caltrans, and USDOT. 2005. Systems Engineering Guidebook for Intelligent Transportation Systems (ITS), version 1.1. Sacramento, CA, USA: California Department of Transportation (Caltrans) Division of Reserach & Innovation/U.S. Department of Transportation (USDOT), SEG for ITS 1.1.

Additional References

ISO/IEC/IEEE. 2009. Systems and software engineering - life cycle processes - project management. Geneva, Switzerland: International Organization for Standardization (ISO)/International Electronical Commission (IEC)/Institute of Electrical and Electronics Engineers (IEEE), ISO/IEC/IEEE 16326:2009(E).

Article Discussion

[Go to discussion page]

<- Previous Article | Parent Article | Next Article ->

Signatures

--Groedler 01:16, 30 August 2011 (UTC)

--Dholwell 14:32, 2 September 2011 (UTC) core edit

--Rturner 13:08, 9 September 2011 (UTC) tech edit