SlideShare ist ein Scribd-Unternehmen logo
1 von 148
Master Class
Principles, Processes, Practices, and Tools to
Increase the Probability of Successfully
Complete Project’s On‒Time, On‒Budget, with
Needed Capabilities
Glen B. Alleman
Thursday 16th August 2018
9:00 AM ‒ 5:00 PM
Getting to Done
1
Who Am I?
 I was educated as a Physicist, but practiced software
development for radar, signal processing systems, and
further educated as a Systems Engineer.
 Moved to managing Software Intensive System of Systems
(SISoS) for space, defense, enterprise IT, industrial systems,
process control, and document management companies.
 Worked in heavy construction, electric utilities (conventional
& nuclear), BioPharma, Federal and State IT systems, Policy
and Root Cause Analysis for a DOD FFRDC firm.
 Developed and deployed program management systems
based on the Principles, Processes, and Practices in the this
workshop.
 Some charts here are from Performance‒Based Project
Management®, Glen B. Alleman, American Management
Association, 2014.
2
PGCS 2018 Master Workshop, Canberra Australia
These Principles, Practices, Processes, and
Tools Have Been Applied with Success At
…3
PGCS 2018 Master Workshop, Canberra Australia
TLO’s for the Master Workshop
 Introduce the Five Immutable Principles and
the Processes that implement them
 Show how the 10 Practices enable project
success through the Five Principles and Five
Processes
 Apply these Principles, Processes, and
Practices to a Real project ‒ TSAS
 Or a project selected by the class
 Develop an artifacts package for the project
4
PGCS 2018 Master Workshop, Canberra Australia
5 PGCS 2018 Master Workshop, Canberra Australia
Why is it so hard to accept the truth that Principles
must be in place, before any Practices and
Process can be applied that might increase the
Probability of Success?
1. What Does DONE Look
Like in Units of Measure
meaningful to the Decision
Makers?
2. How Can We Get to
DONE?
3. Is There Enough Time,
Money, and Resources, to
Get to DONE?
4. What Impediments Will We
Encountered Along The
Way to DONE and How can
They be Removed?
5. What Meaningful Units of
All Successful Projects Require Credible
Answers To These Five Questions …
6 PGCS 2018 Master Workshop, Canberra Australia
All Project Success Starts with the First
Principle of the Five Immutable Principles
 The needed Capabilities, stated as Measures of
Effectiveness and Measures of Performance
define what Done looks like.
 These capabilities trace Value to the Strategy.
 Capabilities lay the ground for adapting to change
found on all projects with emerging requirements.
 Features and Functions fulfill the stated
Requirements needed to implement the
Capabilities.
 Capabilities provide the means to address
unstated future requirements.
7
PGCS 2018 Master Workshop, Canberra Australia
1. Capabilities drive
requirements
2. Requirements
identify Technical
and Process
deliverables
3. Work Packages
describes
deliverables
4. IMS arranges
deliverables
5. WP progress
measured as
Physical Percent
Complete
10 Practices
1. Identify Needed
Capabilities
2. Identify
Requirements
Baseline
3. Establish
Performance
Measurement
Baseline
4. Execute the PMB
5. Continuous Risk
Management
5 Processes
There are 5 Process and 10 Practices that
go along with the 5 Principles
1. What Does Done
Look Like?
2. What is the Plan to
reach Done?
3. What Resource
are needed to
reach Done?
4. What Impediments
will be
encountered along
the wat y to Done?
5. What are the
measures of
progress to Plan?
5 Principles
6. WA assures WP
produce
deliverables in
planned order
7. EV/ES
describes
performance to
plan
8. Conformance
with TPM
adjusts EV
9. Feedback
adjusts WP
sequence and
resource
allocation
10. Future
performance
based on TCPI,
IEAC, and
adjusted work
sequence
8
PGCS 2018 Master Workshop, Canberra Australia
1. Capabilities drive
requirements
2. Requirements
identify
Technical and
Process
deliverables
3. Work Packages
describes
deliverables
4. IMS arranges
deliverables
5. WP progress
measured as
Physical Percent
Complete
10 Practices
1. Identify Needed
Capabilities
2. Identify
Requirements
Baseline
3. Establish
Performance
Measurement
Baseline
4. Execute the PMB
5. Continuous Risk
Management
5 Processes
In time allotted for this course, let’s focus on
…
1. What Does Done
Look Like?
2. What is the Plan
to reach Done?
3. What Resource
are needed to
reach Done?
4. What
Impediments will
be encountered
along the way to
Done?
5. What are the
measures of
progress to Plan?
5 Principles
6. WA assures WP
produce
deliverables in
planned order
7. EV/ES
describes
performance
to plan
8. Conformance
with TPM
adjusts EV
9. Feedback
adjusts WP
sequence and
resource
allocation
10. Future
performance
based on TCPI,
IEAC, and
adjusted work
sequence
9
PGCS 2018 Master Workshop, Canberra Australia
Introduction
Every Project Manager in Every Domain is Faced with Balancing
Cost, Schedule, and Technical Performance on the Path to Done
Let’s Learn how Principles, Processes and Practices can
Increase the Probability of Project Success (PoPS)
10 PGCS 2018 Master Workshop, Canberra Australia
Failure to define Done in a
meaningful way, means the
project starts out without a
Mission, Vision, or Goal and
without any associated metrics.
Progress can only then be
measured by the passage of
time and consumption of money.
For success, progress must be
measured by the Effectiveness
and Performance of its
outcomes.
Technical Performance
Measures and Key Performance
Parameters inform the Measures
of Effectiveness (MoE) and
Measures of Performance MoP.
Introduction
Project Success
Starts with the 1st of 5
Immutable Principles
What Does Done
Look Like in Units
of Measure
Meaningful to the
Decision Makers
11 PGCS 2018 Master Workshop, Canberra Australia
Let’s Get Started …
… But First Some Background Needed to Reach The End
PGCS 2018 Master Workshop, Canberra Australia12
PGCS 2018 Master Workshop, Canberra Australia
How Can We Connect The Moving Parts Of
Principles, Processes, and Practices Into A Cohesive Whole?
13
These Connections Start with Three Core
Frameworks…
 Data is the heart of Integrated Project Performance
Management
 Without data, we can’t make decisions about program’s
performance meaningful to the decision makers ‒ Dollars,
Time, and Technical Performance.
 Without data, processes have nothing to work on.
 Processes transform data into information.
 This information is used to make decisions about the
program.
 The primary decision is how to correct or prevent undesirable
variances to stay on plan to deliver needed capabilities.
 People execute processes using data to increase Probability
of Success for the program.
 People varying various roles and responsibilities in the
process and the creation and use of the data.
14
PGCS 2018 Master Workshop, Canberra Australia
Introduction
16 Program Management Activities
Program
Enablers
Program Process
Capabilities
Business Enablers
15
Booz Allen Hamilton US Department of Energy FramewPGCS 2018 Master Workshop, Canberra Australia
Introduction
The 4+1 Questions
Every Successful
Project Must Answer
Capabilities Requirements Plans Execution
2. What technical and operational requirements are
needed to deliver these capabilities?
1. What capabilities are needed to fulfill the Concept of Operations†,
the Mission and Vision, or the Business System Requirements?
3. What schedule delivers the product or
services on time to meet the requirements?
4. What periodic measures of
physical percent complete
assure progress to plan?
What impediments to success, their mitigations, retirement plans,
or “buy downs are in place to increase the probability of success?”
+ Continuous Risk Management
† A Concept of Operations (ConOps) describes the characteristics of a system from the point of view of an individual who will use
that system. It is used to communicate the quantitative and qualitative system characteristics to all stakeholders.





IntroductionPerformance–Based Project Management®, Copyright © Glen B. Alleman, 2012 - 201816 PGCS 2018 Master Workshop, Canberra Australia
The Five Processes Increase the
Maturity of the Project’s Deliverables
Identify
Capabilities
Needed for
Success
Identify
Requirements
to Deliver
Capabilities
Establish
Performance
Measurement
Baseline
Execute
Performance
Measurement
Baseline
 Define Capabilities
 Define Concept of
Operations
 Assess Needs, Cost,
and Risk Impacts
 Define Balanced and
Feasible Alternatives
 Fact Finding
 Gather And Classify
 Evaluate And
Rationalize
 Prioritize Requirements
 Integrate And Validate
 Decompose Scope
 Assign Accountability
 Arrange Work
 Develop Budget
 Assign Performance
 Perform Work
 Accumulate
Performance Measures
 Analyze Performance
 Take Corrective Action
Perform Continuous Risk Management (CRM)
Define the Measurable
Capabilities of each
Project Outcome
Assure All Requirements
Provided In Support of
Capabilities
Define Measures of
Performance and
Effectiveness
Ensure Cost, Schedule,
and Technical
Performance Compliance
17
PGCS 2018 Master Workshop, Canberra Australia
Introduction
What is a Deliverable?
Project Management Case Study, Pierre Bonnal, CNAM IIM MBA Program, Jun
18
PGCS 2018 Master Workshop, Canberra Australia
Introduction
Define the set of capabilities needed to achieve the project objectives or the particular
end state for a specific scenario. Using the Concept of Operations (ConOps), define
the details of who, where, and how this capability is to be accomplished, employed,
and executed.

Define the technical and operational requirements for the system capabilities to be
fulfilled. First, define these requirements in terms isolated from any implementation
details. Only then bind the requirements with technology.

Build a time–phased network of work activities describing the work to be performed,
the budgeted cost for this work, the organizational elements that produce the
deliverables, and the performance measures showing this work is proceeding
according to plan.

Execute work activities, while assuring all performance assessment represent 100%
completion before proceeding. This means – No rework, no forward transfer of
activities to the future. Assure all requirements are traceable to work & all work is
traceable to requirements.

Apply the processes of Continuous Risk Management for each Performance–
Based Project Management® process area to: Identify, Analyze, Plan, Track,
Control, and Communicate programmatic and technical risk.

IntroductionPerformance–Based Project Management®, Copyright © Glen B. Alleman, 2012 - 201819 PGCS 2018 Master Workshop, Canberra Australia
Partition system capabilities into classes of service within operational scenarios.
Connect the capabilities to system requirements using some visual modeling notation.
Define Measures of Effectiveness (MoE) and Measures of Performance (MoP).
Define the delivery schedule for each measure of performance and effectiveness.
Define scenarios for each system capability.
Connect these scenarios to a Value Stream Map of the increasing maturity of the
program.
Assess value flow through the map for each needed capability.
Identify capability mismatches and make corrections to improve overall value flow.
Assign costs to each system element using a value flow model.
Assure risk, probabilistic cost and benefit performance attributes are defined.
Use cost, schedule and technical performance probabilistic models to forecast
potential risks to program performance.
Make tradeoffs that connect cost, schedule, and technical performance in a single
location that compares the tradeoffs and their impacts.
Use Measures of Effectiveness (MoE) and Measures of Performance (MoP) for
these alternative tradeoffs.
Define the capabilities needed to achieve a desired objective or a particular end state
for a specific scenario. Define the details of who, where, and how these capabilities are
to be delivered and employed to fulfill the Mission and Vision
1.1
1.2
1.3
1.4

IntroductionPerformance–Based Project Management®, Copyright © Glen B. Alleman, 2012 - 201820 PGCS 2018 Master Workshop, Canberra Australia
Produce an overall statement of the problem in the operational context.
Develop the overall operational and technical objectives of the target system.
Defined the boundaries and interfaces of the target system.
Gather required system capabilities, functional, nonfunctional and
environmental requirements, and design constraints.
Build the Top Down capabilities and functional decomposition of the
requirements in a Requirements Management System.
Answer the question “why do I need this?” in terms of operational capabilities.
Build a cost / benefit model using probabilistic assessment of all variables,
their dependencies, and impacts.
For all requirements, perform a risk assessment to cost and schedule.
Determine criticality for the functions of the system.
Determine trade off relationships for all requirements to be used when option
decisions must be made.
For all technical items, prioritize their cost and dependency.
Address the completeness of requirements by removing all “TBD” items.
Validate that the requirements are traceable to system capabilities, goals, and
mission.
Resolve any requirements inconsistencies and conflicts.
Define the technical and operational requirements that must be met for the system capabilities
to be delivered. Define these requirements in terms isolated from any technology or
implementation. Assure each requirement is connected to a need system capability.
2.1
2.2
2.3
2.4
2.5

IntroductionPerformance–Based Project Management®, Copyright © Glen B. Alleman, 2012 - 201821 PGCS 2018 Master Workshop, Canberra Australia
Build a time–phased network of activities describing the work to be performed, the budgeted
cost for this work, the organizational elements that produce the deliverables from this work,
and the performance measures showing this work is proceeding according to plan.
Decompose the program Scope into a product based Work Breakdown Structure (WBS), then
further into Work Packages describing the production of the deliverables traceable to the
requirements, and to the needed capabilities.
3.1
Assign responsibility to Work Packages (the groupings of deliverables) to a named owner
accountable for the management of the resource allocations, cost and schedule baseline,
and technical delivery.
3.2
Arrange the Work Packages in a logical network with defined deliverables,
milestones, internal and external dependencies, with credible schedule, cost, and
technical performance margins.
3.3
Develop the Time–Phased Budgeted Cost for Work Scheduled (BCWS) for the labor and
material costs in each Work Package and the Project as a whole. Assure proper resource
allocations can be met and budget profiles match expectations of the program sponsor
3.4
Assign objective Measures of Performance (MoP) and Measures of Effectiveness (MoE) for
each Work Package and summarize these for the Project as a whole.
3.5
Establish a Performance Measurement Baseline (PMB) used to forecast the Work
Package and Project ongoing and completion cost and schedule performance metrics.
3.6

IntroductionPerformance–Based Project Management®, Copyright © Glen B. Alleman, 2012 - 201822 PGCS 2018 Master Workshop, Canberra Australia
Execute the planned work, assuring all work is 100% complete before proceeding to the
next planned work package. No rework, no forward transfer of activities or features.
Assure every requirement is traceable to work and all work is traceable to requirements.
 Using the Work Package sequencing, release work to be performed as planned.
 With the responsibility assignments, identify the accountable delivery manager to guide the
development of the products or services for each Work Package.
4.1
 Using Physical Percent Complete or Apportioned Milestones, capture measures of
progress to plan for each Work Package.
 Report this Physical Percent Complete in a centralized database for each Work Package
and the program as a whole.
4.2
 Compare the Physical Percent Complete against the Planned Percent Complete for each
period of performance.
 Construct cost and schedule performance indices from this information and the Physical
Percent complete measures.
4.3
 With Cost and Schedule performance indices, construct a forecast of future performance
of cost, schedule, and technical performance compliance.
 Take management actions for any Work Packages not performing as planned.
4.4
 Record past performance based on Work Package completion criteria.
 Record past future forecast performance estimates in a historical database.
 Forecast next future performance estimate against the Performance Measurement Baseline.
 Report this next future performance estimate to the program stakeholders.
4.5

IntroductionPerformance–Based Project Management®, Copyright © Glen B. Alleman, 2012 - 201823 PGCS 2018 Master Workshop, Canberra Australia
 Identify and classify risks in a Risk Register. Separate reducible and Irreducible risks
 Manage this Risk Register through a Risk Management Board.
 Connect these risks and their handling and margins in the Master Schedule.
 Convert risk data into risk decision‒making information.
 Use this analysis information as the decision basis for the program manager to work on
the “right” risks.
 Turn risk information into decision making information and actions (both present and
future).
 Develop actions to address individual risks, prioritize risk actions, and create an integrated
risk management plan to retire the risk or handle it when it turned into an issue.
 Monitor the status of risks and actions taken to ameliorate risks.
 Identify and monitor risks to enable the evaluation of the status of risks themselves and
of risk mitigation plans.
 Risk communication lies at the center of the model to emphasize both its pervasiveness and
its criticality.
 Without effective communication, no risk management approach can be viable.
Continuous Risk Management starts with the underlying principles, concepts, and
functions of risk management and provides guidance on how to implement risk
management as a continuous practice in programs and the organizations that
management programs.
5.1
5.2
5.3
5.4
5.5

IntroductionPerformance–Based Project Management®, Copyright © Glen B. Alleman, 2012 - 201824 PGCS 2018 Master Workshop, Canberra Australia
What Does A Capability “Sound” Like?
25
PGCS 2018 Master Workshop, Canberra Australia
Introduction
What Should We Do?
Where Are We Now?
Identifying Needed System Capabilities
Abstracted from:
“Capabilities‒Based Planning – How It Is
Intended To Work And Challenges To Its
Successful Implementation,” Col. Stephen K.
Walker, United States Army, U. S. Army War
College, March 2005
Identifying Needed System Capabilities
26
PGCS 2018 Master Workshop, Canberra Australia
Introduction
What Is a Requirement?
27
PGCS 2018 Master Workshop, Canberra Australia
Introduction
Identifying Requirements†
† Systems Requirements Practices, Jeffery O. Grady, McGraw Hill, 1993
28
PGCS 2018 Master Workshop, Canberra Australia
Introduction
What Does a Credible Based Plan
and Schedule Look Like?
29
PGCS 2018 Master Workshop, Canberra Australia
Introduction
Establishing the Three Elements of
the Performance Measurement
Baseline
Cost Baseline
Schedule Baseline
Technical Baseline
Determine
Scope and
Approach
Develop
Technical
Logic
Develop
Technical
Baseline
Develop
WBS
Define
Activities
Estimate
Time
Durations
Sequence
Activities
Finalize
Schedule
Identify
Apportioned
Milestones
Determine
Resource
Requireme
nt
Prepare
Cost
Estimate
Resource
Load
Schedule
Finalize
Apportioned
Milestones
Determine
Funding
Constraint
s
Approve
PMB
Perform
Functional
Analysis
30
PGCS 2018 Master Workshop, Canberra Australia
Introduction
How Do We Know We Are Making
Progress to Plan?
31
PGCS 2018 Master Workshop, Canberra Australia
Introduction
Executing the Performance
Measurement Baseline (PMB)
32
PGCS 2018 Master Workshop, Canberra Australia
Introduction
Analyze
Plan
Track
Control
Identify Risks, Issues, and
Concerns
Evaluate, classify, and
prioritize risks
Decide what should be done
about each risk
Monitor risk metrics and
verify/validate mitigations
Make risk decisions
Subproject and partner
data/constraints, hazard
analysis, FMEA, FTA, etc.
Risk data: test data,
expert opinion, hazard
analysis, FMEA, FTA,
lessons learned,
technical analysis
Resources
Replan Mitigation
Program/project
data
(metrics
information)
Statement of Risk
Risk classification, Likelihood
Consequence, Timeframe
Risk prioritization
Research, Watch (tracking
requirements)
Acceptance Rationale,
Mitigation Plans
Risk status reports on:
Risks
Risk Mitigation Plans
Close or Accept Risks
Invoke contingency plans
Continue to track
Perform Continuous Risk Management
33
PGCS 2018 Master Workshop, Canberra Australia
Introduction
Integrating all the Parts into a Whole
Risk Management
SOW
SOO
ConOps
WBS
Technical and Operational
Requirements
CWBS &
CWBS Dictionary
Integrated Master Plan
(IMP)
Integrated Master Schedule
(IMS)
Earned Value Management
System (EVMS)
Objective Status and Essential Views to support the proactive management
processes needed to keep the program GREEN
Performance Measurement Baseline (PMB)
Measures of
Effectiveness (MOE)
Measures of
Performance (MOP)
Measures of Cost –
Schedule Progress
JROC
Key Performance Parameters
(KPP)
Program Specific
Key Performance Parameters
(KPP)
Technical Performance
Measures (TPM)
CWBS
34
PGCS 2018 Master Workshop, Canberra Australia
Introduction
From Mission Capabilities to Done
“Coming to Grips with Measures of
Effectiveness,” N. Sproles, Systems
Engineering, Volume 3, Number 1, pp. 50–58
MoE
KPP
MoP TPMMission Need
Acquirer Defines the Needs and Capabilities
in terms of Operational Scenarios
Supplier Defines Physical Solutions that
meet the needs of the Stakeholders
Operational
measures of success
related to the
achievement of the
mission or
operational
objective being
evaluated.
Measures that
characterize
physical or
functional attributes
relating to the
system operation.
Measures used to
assess design
progress,
compliance to
performance
requirements, and
technical risks.
35
PGCS 2018 Master Workshop, Canberra Australia
Introduction
4+1Critical Processes for Success
Capabilities,
Requirements &
Deliverables
Program
Architecture &
Dependencies
Work
Planning and
Sequencing
Performance
Measurement
Baseline
Programmatic
& Technical Risk
Management
Balanced
Scorecard
Concept of
Operations
Statement of
Objectives
Integrated
Master Plan
(IMP)
Gaps
Overlaps
Interfaces
Integrated
Master Schedule
(IMS)
Schedule margin
to protect critical
deliverables
Cost And
Schedule
Baseline
WBS
RAM
Resource
Loaded IMS
Risk Registry
Risk Handling
Plans
Contingency And
Management
Reserve
Requirements
Traceability
Value Stream
Mapping (VSM)
Work Packages
Planning
Packages
Technical
Performance
Measures (TPM)
Risk Integrated
With IMS
Risk Trending
Measures Of
Effectiveness
Design Structure Earned Value
Measures Of
Performance
Monte Carlo
36
PGCS 2018 Master Workshop, Canberra Australia
Introduction
Getting On the Road to Success Means
…
Where are we going?
How do we get there?
Are there enough resources?
What are impediments to progress?
How do we measure progress?
IntroductionPerformance–Based Project Management®, Copyright © Glen B. Alleman, 2012 - 201837 PGCS 2018 Master Workshop, Canberra Australia
Practices of Performance–
Based Project
Management® guide the
application of the 5
Principles and the 5
Process Areas in this
handbook.
These practices define the
reasons for each process,
connect each practice to a
beneficial outcome, and
integrate the processes into
a seamless delivery system.
Connections
Connecting 5
Principles and 5
Processes With 10
Practices
38 PGCS 2018 Master Workshop, Canberra Australia
Connecting Five Process with Five Practices of Project
Success
10PracticesofProjectManagementSuccess
Chapter I
5
Identify Needed
Capabilities
Establish a
Performance
Measurement
Baseline
Execute the
Performance
Measurement
Baseline
Capabilities
Based Plan
Operational
Needs
Earned Value
Performance
0% /100%
Technical
Performance
Measures
Business or
Mission Value
Stream
Technical
Requirements
Identify
Requirements
Baseline Technical
Performance
Measures
PMB
Changes to
Needed Capabilities
Changes to
Requirements Baseline
Changes to
Performance Baseline
ConnectionsPerformance–Based Project Management®, Copyright © Glen B. Alleman, 2012 - 201839 PGCS 2018 Master Workshop, Canberra Australia
10 Practices of Performance–Based
Project Management®
Assess the capabilities being provided through the deliverables
Fulfill the requirements through effort held in the Work Packages
Produce deliverables
from Work Packages
Planned BCWS
Physical % Complete
WP’s contain
deliverables that fulfill
requirements
Capabilities
topology
defines
requirements
flow down
WP flow must describe
the increasing maturity of
the product or service
Producing the deliverables in the
planned sequence maintains the
value stream to the customer
40
PGCS 2018 Master Workshop, Canberra Australia
Connections
Enough Show and Tell, Let’s …
41 PGCS 2018 Master Workshop, Canberra Australia
A Quick Reminder of the 5 Principles
42
Principle Evidence the Principle of Being Implemented
What does Done Look
Like?
Integrated Master Plan (IMP) with Measures of
Effectiveness (MoE) and Measures of Performance
(MoP)
What's the Plan and
Schedule to get Done?
Integrated Master Schedule (IMS) with Technical
Performance Measures (TPM) and Key
Performance Parameters (KPP)
What resources do we
need for Done?
Resource loaded IMS
What impediments well
we encounter along the
way?
Risk adjusted IMS
What are the units of
measure of progress
toward Done?
Earned Value (EV) and Earned Schedule (ES)
informed by compliance with compliance plans for
MoE, MoP, TPM, KPP, Risk Buydown, and Margin
Buydown
PGCS 2018 Master Workshop, Canberra Australia
Connections
One More Reminder, Project Success is
About Doing a Lot Things Right …
… But in the End, it’s always About the Numbers
Cost Numbers, Schedule Numbers, Risk Numbers,
Technical Performance Numbers
43
PGCS 2018 Master Workshop, Canberra Australia
Connections
Integrating the Parts of a Credible
Program Management System (PMS)
Risk
SOW
Cost
WBS
IMP/IMS
TPM
PMB
❺ Deliverables defined
in the SOW, traced to
the WBS, with
narratives and
Measures of
Performance (MoP)
❹ Budget at the Work
Package level, rolled to
the Control Accounts
showing cost spreads for
all work in the IMS
❻ Measures of
Performance (MoP) for
each critical deliverable in
the WBS and identified in
each Work Package in the
IMS, used to assess
maturity in the IMP
❶ The Products and Processes
in a “well structured”
decomposition, traceable to
the deliverables
❷ Schedule contains all
the Work Packages,
Budget, Risk mitigation
plans, with traces to
the Plan measuring
increasing maturity
through Measures of
Effectiveness (MoE)
and KPPs (JROC and
Program)
❸ Technical and Programmatic
Risks Connected through the WBS,
Risk Register, Plan and Schedule
The Baseline is the Document
of Record for the Program
Performance is Measured
through the PMB
44
PGCS 2018 Master Workshop, Canberra Australia
Connections
Before We Start the Hands On Part
What does it mean to be Credible?
The Simplest Answer is …
Our Artifact's Are Believable
A credible
Integrated Master Plan, Integrated Master Schedule,
Cost Baseline, Risk Baseline, Estimate At Completion,
Physical Percent Complete
45 PGCS 2018 Master Workshop, Canberra Australia
That are Connected in this Way
46
PGCS 2018 Master Workshop, Canberra Australia
Connections
The WBS is Paramount
 The WBS defines the deliverables and the
supporting processes that produce them
 The WBS Dictionary describes the technical
and operation behaviors that will be assessed
during the development of the deliverables
 The terminal nodes of the WBS define the
deliverables produced by the Work Packages
in the IMS and assessed through the IMP
Accomplishment Criteria (AC)
47
PGCS 2018 Master Workshop, Canberra Australia
Connections
The IMP Starts with the Buyer
 The IMP defines the measuring of increasing
maturity for the deliverables as the program
moves from left to right
 Significant Accomplishments (SA) are defined
by the Measures of Effectiveness (MoE)
 Accomplishment Criteria (AC) are defined by
the Measures of Performance (MoP)
 Risks are assigned at all levels of the IMP and
IMS
48
PGCS 2018 Master Workshop, Canberra Australia
Connections
Natural Uncertainties
and Event Based Risks†
 Natural uncertainties in cost and schedule
processes create risks to completing on time and
on budget
 Event based risks create impacts to cost,
schedule, and technical performance
 Event based risks are handled through risk
mitigations
 Natural uncertainties are handled through in cost,
schedule, and technical performance margins
 To be credible, the PMB must include both type of
risks with their handling strategies
49
† These are referred to in the literature as Aleatory and Epistemic. We’ll used the naturally occurring and Event
Based in this introduction, but those terms are consider operational rather than mathematical.
PGCS 2018 Master Workshop, Canberra Australia
Connections
Costs Assigned to Packages of
Work
 Labor and material cost are represented in the
Integrated Master Schedule (IMS) and provide
visibility to the probability of program success
 Variances in labor and material costs are
modeled in the same way as work durations
 Event based risks impact both cost and
schedule and are modeled in the PMB
 Risk retirement cost is allocated for the work
effort in response to Event Based risks
50
PGCS 2018 Master Workshop, Canberra Australia
Connections
Statement of Work
 Work in the PMB starts with the Statement of
Work and flows through the Work Breakdown
Structure
 Measures of Effective (MoE) and Measures of
Performance (MoP) can be defined in the
SOW or WBS Dictionary
 Traceability from the IMP to the IMS to all
performance measures in the SOW is the
basis of program performance measurement
51
PGCS 2018 Master Workshop, Canberra Australia
Connections
Technical Performance Measures
 Key Performance Parameters (KPP, both
Acquisition owner and Program specific) and
Technical Performance Measures (TPM)
define how the deliverables complying with the
Statement of Work, Concept of Operations,
and CDRLs
 TPMs inform the measures cost and schedule
for delivered program outcomes
 TPM, MoE, MoP, and KPPs provide
assessment of the cost and schedule
effectiveness
52
PGCS 2018 Master Workshop, Canberra Australia
Connections
TSAS
Command and Data Center
Mobile Sensors
Ground Based Sensors
53
UAV with Airborne Sensors
PGCS 2018 Master Workshop, Canberra Australia
TSAS
Command and Data Center
Mobile Sensors
Ground Based Sensors
54
UAV with Airborne Sensors
PGCS 2018 Master Workshop, Canberra Australia
Top Level Capabilities for Airborne Sensors
55
Capability Program Goal
Maximum Range 2,000 NM
Maximum Altitude 35,000 feet
Maximum endurance 12 hours
SATCOM Link 1.5 – 50 Mbps
LOS Datalink > 50.0 Mbps
Synthetic Aperture Radar
(SAR)
1.0/0.3m resolution (WAS/Spot)
Moving Target Indicator (MTI) 20 -200Km/10m Range resolution
Electro Optical NIIRS 6.5/6.0 (Spot/WAS)
Infrared NIIRS 5.5/5.0 (Spot/WAS)
Wide Area Search 250 Sq. NMI/Day (50 x 50)
Target Coverage 1,000 spot targets / day
PGCS 2018 Master Workshop, Canberra Australia
Connections
TSAS Performance Measures Derived from
those Top Level Capabilities
56
TSAS
Element
Measures Answers the Question Example
SOW,
SOO,
ConOps
Measures of
Effectiveness (MoE)
How Do We Know we are
Accomplishing the Mission?
We need the capability to Increase IED
Placement search capabilities by 50%
WBS
Technical
Performance
Measure (TPM)
What are we building and how do we
know it meets the specifications that
will accomplish the Mission?
Systems, subsystems, and supporting
processes for each deliverable
IMP – PE
and SA
MoE for the Program
Events and
Significant
Accomplishments
How can we measure the increasing
maturity of the deliverables in the
narratives in the Capabilities Based
documents
Sensor payloads capable of IR and UV
detectors within the avionics bay
IMP – AC
Measures of
Performance (MoP)
Technical Performance of the
deliverables derived from the MoEs
100 square miles per hours search
capabilities
IMS
Technical
Performance
Measures (TPM)
How does the work increase the
maturity of the deliverables?
Sensor platform TPMs inside the
bounds, on-time, on-schedule?
Tasks CPI, SPI, TCPI
What work is needed to increase the
maturity of the deliverables?
Cost and schedule matching TPM
progress?
Risk
Register
Identified risks, with
handling strategies
What are the Epistemic risks and how
are they represented in the IMS?
All aleatory risks included in duration
and cost. Epistemic risk retirement
PGCS 2018 Master Workshop, Canberra Australia
Connections
The Work Breakdown
Structure is our starting point
for developing all other
elements needed for the
Performance Measurement
Baseline.
The TSAS WBS is defined
using the a standard. In the
US that is MIL‒STD‒881C,
with an appendix for UAVs.
From this, the details of the
avionics subsystems will be
used for the development of
the Integrated Master Plan
and Integrated Master
Schedule.
WBS
The WBS is
Paramount
57 PGCS 2018 Master Workshop, Canberra Australia
Integrating the Parts of a Credible
Program Management System (PMS)
Risk
SOW
Cost
WBS
IMP/IMS
TPM
PMB
❺ Deliverables defined
in the SOW, traced to
the WBS, with
narratives and
Measures of
Performance (MoP)
❹ Budget at the Work
Package level, rolled to
the Control Accounts
showing cost spreads for
all work in the IMS
❻ Measures of
Performance (MoP) for
each critical deliverable in
the WBS and identified in
each Work Package in the
IMS, used to assess
maturity in the IMP
❶ The Products and Processes
in a “well structured”
decomposition, traceable to
the deliverables
❷ Schedule contains all
the Work Packages,
Budget, Risk mitigation
plans, with traces to
the Plan measuring
increasing maturity
through Measures of
Effectiveness (MoE)
and KPPs (JROC and
Program)
❸ Technical and Programmatic
Risks Connected through the WBS,
Risk Register, Plan and Schedule
The Baseline is the Document
of Record for the Program
Performance is Measured
through the PMB
58
PGCS 2018 Master Workshop, Canberra Australia
WBS
What is the Work Breakdown Structure?
 Defines the total System or the System of Systems
(SoS)
 Provides the framework for planning, prioritizing,
managing, and tracking all work done on the program
 Products
 Supporting services
 Facilities
 Provides the framework for
 Cost Structure for estimating and cost reporting
 Resource allocation
 Status Reporting
 Performance Measurement
 Identify and managing program risk
59
PGCS 2018 Master Workshop, Canberra Australia
WBS
System Architecture Drives the WBS
60
PGCS 2018 Master Workshop, Canberra Australia
WBS
 End Products used to implement the mission
 Based on the System (Physical) Architecture
 Enabling Products and Services
 Products and services required to develop, produce,
and support the end items
 Based on Life Cycle
 The first three (end product) WBS levels:
 Level 1: Overall System
 Level 2: Major Element (or Segment)
 Level 3: Subordinate Components (or Prime Items)
 Levels 4+ continue the decomposition to the
Configuration Item (CI) level
Structure of the TSAS WBS
61
PGCS 2018 Master Workshop, Canberra Australia
WBS
Three WBS’s for TSAS
 High-Level (First 3
levels)
 Provides Program
Structure
 Generally
developed/controlle
d by Customer
(Government)
 Detailed (Levels
4+)
 Provides framework
for Contract Work
Packages and
Costing
 Generally
developed by
Contractor
 Generally follows
Program WBS
 Detailed (Level 4+)
 Provides framework
for Subcontract
Work Packages
and Costing
 Generally
developed by
Subcontractor
 Generally follows
Contract WBS
62
PGCS 2018 Master Workshop, Canberra Australia
WBS
Level 1 (System of Systems) WBS
63
1.0 Tactical Situational Awareness System
(TSAS)
1.1 Unmanned Air Vehicle (UAV)
1.2 Mobile Sensor System
1.3 Ground Based Sensors
1.4 Ground / Host Segment (Command and
Control)
1.5 System of Systems Engineering
1.6 Program Management
1.7 System of Systems Test and Integration
1.8 System of Systems Training
1.9 Systems of Systems Data Packages
1.10 Peculiar Support Equipment
1.11 Common Support Equipment
1.12 Operational and Site Support
1.13 Industrial Facilities
UAV with Airborne
Sensors
PGCS 2018 Master Workshop, Canberra Australia
WBS
Level 2 Unmanned Air Vehicle (UAV)
64
UAV with Airborne Sensors
PGCS 2018 Master Workshop, Canberra Australia
WBS
Level 3 UAV Avionics
65
PGCS 2018 Master Workshop, Canberra Australia
WBS
Level 4 UAV Avionics – GN&C
66
PGCS 2018 Master Workshop, Canberra Australia
WBS
Level 4 UAV Avionics - MMS
67
PGCS 2018 Master Workshop, Canberra Australia
WBS
Building the IMP is a Systems
Engineering activity.
The Integrated Master Plan
(IMP) is a Programmatic
Architecture in the same way
the hardware and software
are a Product Architecture.
Poor, weak, or unstructured
Programmatic Architecture
reduces visibility to the
Product Architecture’s
performance measures of
cost and schedule connected
with Technical Performance
Measures.
IMP
The Measures of
Effectiveness (MoE)
and
Measures of
Performance (MoP)
are held in the
Integrated Master
Plan (IMP)
68 PGCS 2018 Master Workshop, Canberra Australia
Quick View of Building the IMP
 Start with each Program Event and define the
Significant Accomplishments their entry and
exit criteria to assess the needed maturity of
the key deliverables
 Arrange the Significant Accomplishments in
the proper dependency order
 Segregate these Significant Accomplishments
into swim lanes for IPTs
 Define the dependencies between each SA
69
PGCS 2018 Master Workshop, Canberra Australia
IMP
Integrating the Parts of a Credible
Program Management System (PMS)
Risk
SOW
Cost
WBS
IMP/IMS
TPM
PMB
❺ Deliverables defined
in the SOW, traced to
the WBS, with
narratives and
Measures of
Performance (MoP)
❹ Budget at the Work
Package level, rolled to
the Control Accounts
showing cost spreads for
all work in the IMS
❻ Measures of
Performance (MoP) for
each critical deliverable in
the WBS and identified in
each Work Package in the
IMS, used to assess
maturity in the IMP
❶ The Products and Processes
in a “well structured”
decomposition, traceable to
the deliverables
❷ Schedule contains all
the Work Packages,
Budget, Risk mitigation
plans, with traces to
the Plan measuring
increasing maturity
through Measures of
Effectiveness (MoE)
and KPPs (JROC and
Program)
❸ Technical and Programmatic
Risks Connected through the WBS,
Risk Register, Plan and Schedule
The Baseline is the Document
of Record for the Program
Performance is Measured
through the PMB
70
PGCS 2018 Master Workshop, Canberra Australia
IMP
The Critical Purpose of the IMP
71
The IMP defines the connections between the Product maturity – Vertical – and the implementation of this Product
maturity through the Functional activities – the Horizontal
PGCS 2018 Master Workshop, Canberra Australia
IMP
Next Step is the Build an Integrated Master
Plan
 From the WBS define the Significant
Accomplishments (SA) and Accomplishment
Criteria (AC) for each Program Event (PE) for
each of the terminal nodes in the WBS
72
What must be accomplished to complete the
Program Event?
This is where the WBS and the IMP are joined
and are at the same time separate
Both are needed for a credible PMB
PGCS 2018 Master Workshop, Canberra Australia
IMP
73
The IMP tells us where is the program going?
The Plan describes where we are going, the various paths we can take to
reach our destination, and the progress or performance assessment points
along the way to assure we are on the right path.
These assessment points measures the “maturity” of the product or
service against the planned maturity. This is the only real measure of
progress – not the passage of time or consumption of money.
The Integrated Master Plan (IMP) Is A Strategy For
The Successful Completion Of The Project
PGCS 2018 Master Workshop, Canberra Australia
74
The IMP / IMS Structure
IMS
IMP
Describes how program
capabilities will be
delivered and
how these
capabilities will
be recognized
as ready for
delivery
Supplemental Schedules (CAM Notebook)
Work Packages and Tasks
Criteria
Accomplishment
Events
or
Milestones
PGCS 2018 Master Workshop, Canberra Australia
75
 Vertical traceability AC  SA  PE
 Horizontal traceability WP  WP  AC
Program Events
Define the maturity
of a Capability at a point in
time.
Significant Accomplishments
Represent requirements
that enable Capabilities.
Accomplishment Criteria
Exit Criteria for the Work
Packages that fulfill Requirements.
Work
Package
Work
Package
Work
Package
Work
Package
Work
Package
Work
Package
Work
package
Vertical and Horizontal Traceability
PGCS 2018 Master Workshop, Canberra Australia
76
The IMP’s role during Execution
Program ExecutionPMB for IBRProposal SubmittalDRFP & RFP
Performance Measurement Baseline
Tasks (T)
BOE
% Complete
Statement of Work
Program Deliverables
IMP
Accomplishments (A)
Criteria (C)
EVMS
Events (E)
Budget Spreads by CA & WPCAIV
Capabilities Based Requirements
X BCWS =
Probabilistic Risk Analysis
=
Time keeping and ODC =
Technical Performance Measure
BCWP
ACWP
Cost & Schedule Risk Model
BCWS
D ecreasing techni cal and programmatic ri sk usi ng R i sk Management Methods
IMS
Physical % Complete
Continuity and consistency from DRFP through Program Execution
WBS
PGCS 2018 Master Workshop, Canberra Australia
77
The IMP speaks to Measures of Effectiveness (MoE)
and Measures of Performance (MoP)
 This is where
TPMs are
connected with
the MoE’s and
MoP’s
 For each
deliverable from
the program, all
the “measures”
must be defined
in units
meaningful to
the decision
makers.
 Here’s some
“real” examples.
1. Provide Precision
Approach for a 200
FT/0.5 NM DH
2. Provide bearing and
range to AC platform
3. Provide AC surveillance
to GRND platform
Measures of
Effectiveness (MoE)
1. Net Ready
2. Guidance Quality
3. Land Interoperability
4. Manpower
5. Availability
JROC Key Performance
Parameters (KPP)
1. Net Ready
 IPv4/6 compliance
 1Gb Ethernet
2. Guidance quality
 Accuracy threshold p70
@ 6M
 Integrity threshold 4M
@ 10-6 /approach
3. Land interoperability
 Processing capability
meets LB growth matrix
4. Manpower
 MTBC >1000 hrs
 MCM < 2 hrs
5. Availability
 Clear threshold >99%
 Jam threshold >90%
Measures of Performance
(MoP)
1. Net Ready
 Standard message packets
2. Guidance Quality
 Multipath allocation budget
 Multipath bias protection
3. Land Interoperability
 MOSA compliant
 Civil compliant
4. Manpower
 Operating elapsed time
meters
 Standby elapsed time
indicators
5. Availability
 Phase center variations
Technical Performance
Measures (TPM)
Mission Capabilities and Operational Need
Technical Insight – Risk adjusted performance to plan
PGCS 2018 Master Workshop, Canberra Australia
78
 Program Event (PE)
– A PE assess the readiness or completion as a measure
of progress
– First Flight Complete
 Significant Accomplishment (SA)
– The desired result(s) prior to or at completion of an
event demonstrate the level of the program’s progress
– Flight Test Readiness Review Complete
 Accomplishment Criteria (AC)
– Definitive evidence (measures or indicators) that
verify a specific accomplishment has been completed
– SEEK EAGLE Flight Clearance Obtained
F-22 Example
PGCS 2018 Master Workshop, Canberra Australia
79
 Start with the Significant Accomplishments and sequence
them to the maturity flow for each Program Event
 The WBS connections then become orthogonal to this flow
79
Program Event
SRR SDR PDR CDR TRR ATLO
Work Breakdown
Structure
4.920-SDAI A01, A02 B01 C01, C02 D01 E01 F01
4.200-Sys Test A05 B03, B04 D02, D03 E02 F02
4.300-Radar A03 B02 C03 E03
4.330-O&C Sys A06, A07 B05 C04 D04 E04 F03, F04
4.400- I&T A08 C05 E05, E06 F05
4.500-Support A09 D05 E07 F06, F07
The IMP’s connection to the WBS
PGCS 2018 Master Workshop, Canberra Australia
The Primary Role of the IMP is to Describe
what Done Looks Like in MoE’s and MoP’s
80
19 October 1899 Robert Goddard decided that he wanted to "fly without wings" to Moon.
PGCS 2018 Master Workshop, Canberra Australia
IMP
The Integrated Master
Schedule (IMS) is derived
directly from the Integrated
Master Plan’s
Accomplishment Criteria
(AC).
The IMS shows the order in
which the Work Packages
must be performed to assure
the Accomplishment Criteria
are completed within the
define Measures of
Performance, the Key
Performance Parameters,
and the Technical
Performance Measures.
IMS
Developing
the
Integrated Master
Schedule
from
the
Integrated Master
Plan
81 PGCS 2018 Master Workshop, Canberra Australia
Quick View of Building the IMS
 The WBS is Paramount
 The IMP defines the increasing maturity of the
program’s deliverables (end item)
 The IMS sequences the Work Packages
containing the work activities to produce the
End Item Deliverables
 The IMS is built from the IMP, with WBS
coding to assure coverage of all deliverables
82
PGCS 2018 Master Workshop, Canberra Australia
IMS
The Integrated Master Schedule
 The horizontal sequence of work activities that
produce increasing maturity of the product or
services delivered by the program
83
PGCS 2018 Master Workshop, Canberra Australia
IMS
A Credible Integrated Master Plan Must …
84
 Show what Done looks
like through tangible
evidence of success
 Show the order of the
work needed to get to
Done at each stage
 Define the needed
resources to reach
Done
 Identify risks to Done
and their handling
 Measure physical progress toward Done in units
meaningful to the decision makers
PGCS 2018 Master Workshop, Canberra Australia
IMS
Building the Integrated Master Schedule
requires 10 Steps
85
Capture All Activities1
2
3
4
5
6
7
8
9
10
Sequence These Activities
Assign Resources To These Activities
Establish Duration For These Activities
Verify Schedule Is Traceable Horizontally And Vertically
Confirm Valid Critical Path – schedule matches program
Ensure Reasonable Total Float
Conduct Schedule Risk Analysis
Update Schedule With Actual Progress
Maintain Baseline with Repeatable Process
PGCS 2018 Master Workshop, Canberra Australia
IMS
86
Mission Need
Mission Management
Control
1st
Level
RTOS integration
1st
Level
Enroute planning
2nd
Level
Routing upload buffer
2nd Level
Planning verification
2nd
Level
Navigation sensor
interface
2nd Level
1553 data bus drivers
Deliverables defined in WP
Back to the WBS
It connects Work Packages in IMS to IMP
PGCS 2018 Master Workshop, Canberra Australia
The IMS Provides Visibility to …
87
Work Packages
Deliverables
Technical
Capabilities
Mission
Requirements
§ Deliverables represent the required mission
capabilities and their value as defined by the
mission and shared by the development
team.
§ When all deliverables and their Work
Packages are completed, they are not
revisited or reopened.
§ The progression of Work Packages defines the
increasing maturity of the project.
§ Completion of Work Packages is represented
by the Physical Percent Completion of the
program.
PGCS 2018 Master Workshop, Canberra Australia
IMS
Quantifiable Backup Data (QBD)
 QBD is a detailed listing of tasks necessary to complete all scope in
a work package during the defined period of performance.
 It is an approach to objectively measure performance
 Each task on the list is weighted – total weighting equals 100% of the
work package’s Budget at Completion. This weighting should not be
equal weighting on every task.
 The CAM assesses physical percent complete of each QBD task.
 The percent complete is calculated from the cumulative assessments.
 The purpose of the QBD is to help:
 Ensure and demonstrates that all contract work is accounted for
 Ensure the schedule and budget are realistic and achievable
 Mitigate schedule and budget risks
 Provide a basis for objectively assessing progress for discretely
measured work packages.
88
PGCS 2018 Master Workshop, Canberra Australia
IMS
Naturally occurring uncertainties
(Aleatory) in cost, schedule, and
technical performance can be
modeled in a Monte Carlo
Simulation tool.
Event Based uncertainties
(Epistemic) require capturing and
defining the probability of their
occurrence, modeling their
impacts, defining handling
strategies, modeling the
effectiveness of these handling
strategies, and modeling the
residual risks, and the impacts of
both the original risk and the
residual risk on the program.
The management of
Uncertainties in cost, schedule,
and technical performance; and
the Event Based uncertainty and
the resulting risk are both critical
success factors for the programs.
Risk
Risk Management is
Project Management
for Adults ‒ Tim Lister
89 PGCS 2018 Master Workshop, Canberra Australia
Core Elements of Program Risk
Management
 The effectiveness of risk management
depends on the people who set it up and
coordinate the risk management process
 On many program risk management consists
only of having a policy and oversight
 If we treat red flags as false alarms rather than
early warnings of danger this incubates the
threats to program success
 Group think of dominate leaders often inhibits
good thinking about risks
90
“Towards a Contingency Theory of Enterprise Risk Management,” Anette Mikes and Robert Kaplan, Working Paper 13-063 January
13, 2014PGCS 2018 Master Workshop, Canberra Australia
Risk
91
91 PGCS 2018 Master Workshop, Canberra Australia
Some Words About Uncertainty
 When we say uncertainty, we speak about a
future state of a system that is not fixed or
determined
 Uncertainty is related to three aspects of our
program management domain:
 The external world – the activities of the program
 Our knowledge of this world – the planned and
actual behaviors of the program
 Our perception of this world – the data and
information we receive about these behaviors
92
PGCS 2018 Master Workshop, Canberra Australia
Risk
Some Words About the
Risk that Results from Uncertainty
 Risk has Two Dimensions
 The degree of possibility that an state or condition will
take place or occur sometime in the future
 The consequences of that state or condition, once it
has occurred
 The degree of possibility is qualified as the
 Probability of Occurrence (event based)
 Probability Distribution Function (a distribution of the
variability of a random number)
 The consequences are usually taken to be
undesirable and qualified as the magnitude of
harm and the remaining probability of a
recurrence of the same risk.
93
PGCS 2018 Master Workshop, Canberra Australia
Risk
All Program Activities have Naturally
Occurring (Aleatory) Uncertainties
94
 Naturally occurring uncertainty
and its resulting risk, impacts the
probability of a successful
outcome.
 The irreducible statistical
behavior of these activities, their
arrangement in a network of
activities, and correlation
between their behaviors creates
risk.
 Adding margin protects the outcome from the impact of this
naturally occurring uncertainty
 The question is ‒ given the statistical nature of the
Irreducible Uncertainty, what’s the Probability we will be
late, over budget, or the technical outcome won’t work as
needed?PGCS 2018 Master Workshop, Canberra Australia
Risk
The Relationship Between Uncertainty and
Risk
 Uncertainty is present when probabilities cannot
be quantified in a rigorous or valid manner, but
can described as intervals within a probability
distribution function (PDF)
 Risk is present when the Uncertainty of an
outcome can be quantified in terms of
 Probability of Occurrence (Epistemic uncertainty)
 A range of possible values in a Probability Distribution
Function (Aleatory Uncertainty)
 This distinction is important for modeling the
future performance of cost, schedule, and
technical outcomes of a program.
 Risk from Epistemic uncertainty is reducible
 Risk from Aleatory uncertainty is irreducible
95
PGCS 2018 Master Workshop, Canberra Australia
Risk
Taxonomy of Uncertainty and the Risk it
Creates
96
PGCS 2018 Master Workshop, Canberra Australia
Risk
Remember ‒ WBS is Paramount ?
Risk Propagates Through the WBS
97
PGCS 2018 Master Workshop, Canberra Australia
Risk
Connecting Epistemic Risk Retirement in
the IMS
 The work to “Buy down”
risk is planned in the
IMS.
 MoE, MoP, and KPP
defined in the Work
Packages for the critical
measure, e.g. weight.
 If we can’t verify we’ve
succeeded, then the
risk did not get reduced.
 The risk may have
gotten worse.
98
Risk: CEV-037 - Loss of Critical Functions During Descent
Planned Risk Level Planned (Solid=Linked, Hollow =Unlinked, Filled=Complete)RiskScore
24
22
20
18
16
14
12
10
8
6
4
2
0
Conduct Force and Moment Wind
Develop analytical model to de
Conduct focus splinter review
Conduct Block 1 w ind tunnel te
Correlate the analytical model
Conduct w ind tunnel testing of
Conduct w ind tunnel testing of
Flight Application of Spacecra
CEV block 5 w ind tunnel testin
In-Flight development tests of
Damaged TPS flight test
31.Mar.05
5.Oct.05
3.Apr.06
3.Jul.06
15.Sep.06
1.Jun.07
1.Apr.08
1.Aug.08
1.Apr.09
1.Jan.10
16.Dec.10
1.Jul.11
PGCS 2018 Master Workshop, Canberra Australia
Risk
99
The Final Notion of Risk
PGCS 2018 Master Workshop, Canberra Australia
Root Causes of Risk
Some are Reducible some are Irreducible
100
Unrealistic Performance Expectations
missing Measures of Effectiveness
(MOE), Performance (MOP), and
Technical Performance Measures (TPM)
Unrealistic Cost and Schedule Estimates
based on inadequate risk adjusted
growth models
Inadequate assessment of risk and
unmitigated exposure to these risks
without proper handling plans
Unanticipated Technical Issues without
alternative plans and solutions to
maintain effectiveness
Cost,
Schedule,
and
Technical
Impacts
TheLensofthePerformance
Indicators
“Borrowed” with permission from Mr. Gary
Bliss, Director, Performance Assessments
and Root Cause Analyses (PARCA), Office of
Assistant Secretary of Defense for
Acquisition.
PGCS 2018 Master Workshop, Canberra Australia
Risk
101
 Lack of predictive
variance analysis
 Untimely and unrealistic
Latest Revised Estimates
(LRE)
 Progress not monitored
in a regular and
consistent manner
 Lack of vertical and
horizontal traceability
cost and schedule data
for corrective action
 Lack of internal
surveillance and controls
 Managerial actions not
demonstrated using
Earned Value
 Inattention to budgetary
responsibilities
 Work authorizations that
are not always followed
 Issues with Budget and
data reconciliation
 Lack of an integrated
management system
 Baseline fluctuations and
frequent replanning
 Current period and
retroactive changes
 Improper use of
management reserve
 EV techniques that do
not reflect actual
performance
Without Removing these items, Train Wreck of the Program Starts When there is …
Mary K. Evans Picture Library
PGCS 2018 Master Workshop, Canberra Australia
Beware the Black Swan
102 PGCS 2018 Master Workshop, Canberra Australia
Using Integrated
Master Plan and Risk
Adjusted Integrated
Master Schedule,
showing the needed
progress to that plan is
the basis of measuring
physical percent
complete.
These measures start
with Technical
Performance
Measures (TPM)
Progress to
Measuring Progress
to Plan means
Measuring Physical
Percent Complete in
meaningful units of
Measure.
These are Technical
Performance
Measures (TPM)
This Does NOT Start
with EV or ES
103 PGCS 2018 Master Workshop, Canberra Australia
TPM’s Inform EV and ES to show progress
to our destination
 How do we increase visibility into the program’s performance?
 How do we reduce cycle time to deliver the product?
 How do we foster accountability?
 How do we reduce risk?
 How do we start this journey to success?
104
PGCS 2018 Master Workshop, Canberra Australia
Progressof
Plan
To Achieve Success …
105
©gapingvoid ltd www.gapingvoidgallery.com
We Need to …
PGCS 2018 Master Workshop, Canberra Australia
Progressof
Plan
Here’s where TPMs Start
106
PGCS 2018 Master Workshop, Canberra Australia
Progressof
Plan
Guidance for MoE’s, MoP, and TPMs
Belongs to Systems Engineering
The starting point is not EVM, it’s Systems
Engineering
 MOE’s are an essential part of Systems
Engineering, guided by IEEE 1220 and EIA
632.
 System’s Engineers drive the content of all
measurement items, customer or supplier
107
PGCS 2018 Master Workshop, Canberra Australia
Progressof
Plan
Previous Approaches Using EV are Mostly
Unsuccessful Connecting these Measures
 Traditional approaches to program
management are retrospective
 Cost and schedule of Earned Value
 Risk Management
 Systems Engineering
 Reporting past performance
 Sometimes 30 to 60 days old
 Variances are reporting beyond the widow of
opportunity for correction
108
PGCS 2018 Master Workshop, Canberra Australia
Progressof
Plan
TPMs have been around for 34 years
Why Is This Hard To Understand?
 We seem to be focused on EV reporting, not the use of
EV to manage the program.
 Getting the CPR out the door is the end of Program
Planning and Control’s efforts, not the beginning.
109
… the basic tenets of the process are the
need for seamless management tools, that
support an integrated approach … and
“proactive identification and management of
risk” for critical cost, schedule, and technical
performance parameters.
― Secretary of Defense, Perry memo, May
1995 TPM Handbook 1984
PGCS 2018 Master Workshop, Canberra Australia
Progressof
Plan
Technical Performance Measures do what
they say,
Measure the Technical Performance
of the product or service produced by the
program
Back to Technical Performance
Measures
110
PGCS 2018 Master Workshop, Canberra Australia
What’s Our Motivation for
“Connecting the Dots?”
PGCS 2018 Master Workshop, Canberra Australia
Technical Performance Measures …
 Provide program management with information to
make better decisions
 Increase the probability of delivering a solution that
meets both the requirements and mission need
111
TPM
Measure of Effectiveness (MoE)
PGCS 2018 Master Workshop, Canberra Australia
Measures of Effectiveness …
 Are stated by the buyer in units meaningful to the buyer
 Focus on capabilities independent of any technical
implementation
“Technical Measurement,” INCOSE–TP–2003–020–01
112
TPM
Measure of Performance (MoP)
PGCS 2018 Master Workshop, Canberra Australia
Measures of Performance are …
 Attributes that assure the system has the capability to
perform
 Assessment of system to assure it meets design
requirements necessary to satisfy the MOE
“Technical Measurement,” INCOSE–TP–2003–020–01
113
TPM
Key Performance Parameters (KPP)
PGCS 2018 Master Workshop, Canberra Australia
Key Performance Parameters …
 have a threshold or objective value
 Characterize the major drivers of performance
 Are considered Critical to Customer (CTC)
“Technical Measurement,” INCOSE–TP–2003–020–01
114
TPM
Technical Performance Measures (TPM)
PGCS 2018 Master Workshop, Canberra Australia
Technical Performance Measures …
 Assess design progress
 Define compliance to performance requirements
 Identify technical risk
 Are limited to critical thresholds
 Include projected performance
“Technical Measurement,” INCOSE–TP–2003–020–01
115
TPM
Dependencies Between Measures
PGCS 2018 Master Workshop, Canberra Australia
“Coming to Grips with Measures of Effectiveness,” N. Sproles,
Systems Engineering, Volume 3, Number 1, pp. 50–58
116
TPM
“Candidates” for Technical Measures
PGCS 2018 Master Workshop, Canberra Australia
INCOSE Systems Engineering Handbook
Concept Description
Physical Size and Stability
Useful Life
Weight
Volumetric capacity
Functional Correctness
Accuracy
Power performance
All the “ilities”
Supportability
Maintainability
Dependability
Reliability = Mean Time Failure
Efficiency
Utilization
Response time
Throughput
Suitability for Purpose Readiness
117
TPM
“Measures” of Technical Measures
PGCS 2018 Master Workshop, Canberra Australia
INCOSE Systems Engineering Handbook
Attribute Description
Achieved to Date Measured technical progress or estimate of progress
Current Estimate
Value of a technical parameter that is predicted to be
achieved
Milestone
Point in time when an evaluation of a measure is
accomplished
Planned Value Predicted value of the technical parameter
Planned Performance
Profile
Profile representing the project time phased
demonstration of a technical parameter
Tolerance Band Management alert limits
Threshold Limiting acceptable value of a technical parameter
Variances
 Demonstrated technical variance
 Predicted technical variance
118
TPM
A Simple Method of Assembling the TPMs
PGCS 2018 Master Workshop, Canberra Australia
119
TPM
TPMs from
James Webb Space Telescope
PGCS 2018 Master Workshop, Canberra Australia
120
TPM
TPMs from
Chandra X–Ray Telescope
PGCS 2018 Master Workshop, Canberra Australia
121
TPM
TPMs Start With The WBS
122
1.1.2 Airframe
PGCS 2018 Master Workshop, Canberra Australia
TPM
What Do We Need To Know About This
Program Through TPMs
 What WBS elements represent the TPMs?
 What Work Packages produce these WBS
elements?
 Where do these Work Packages live in the
IMS?
 What are the Earned Value baseline values for
these Work Packages?
 How are going to measure all these variables?
 What does the curve look like for these
measurements?
123
PGCS 2018 Master Workshop, Canberra Australia
TPM
Verifying Each TPM at Each Stage in the
Program
Question Answered by the
TPM
Evidence that we’re meeting the TPM
Contract
Award
Do we know what we promised
to deliver, now that we’ve won?
With our submitted ROM what are the values we need to get
through Integrated Baseline Review «how do we measure weight
for each program event»
System
Functional
Requiremen
ts
Can we proceed into preliminary
design?
The contributors to the vehicle weight are confirmed and the upper
limits defined in the product architecture and requirements flow
down database (DOORS) into a model
System
Requiremen
ts Review
Can we proceed into the System
Development and Demonstration
(SDD) phase
Do we know all drivers of vehicle weight? Can we bound their
upper limits? Can the subsystem owners be successful within
these constraints uses a high fidelity model?
Preliminary
Design
Review
Can we start detailed design,
and meet the stated performance
requirements
within cost, schedule,
risk, and other constraints?
Does each subsystem designer have the target component weight
target and have some confidence they can stay below the upper
bound? Can this be verified in some tangible way? Either through
prior examples or a lab model?
Critical
Design
Review
Can the system proceed to
fabrication, demonstration, and
test, with the within cost,
schedule, risk, and other system
constraints.
Do we know all we need to know to start the fabrication of the first
articles of the flight vehicle. Some type of example, maybe a
prototype is used to verify we’re inside the lines
Test Does the assembled vehicle fall within the weight range limits for
124
PGCS 2018 Master Workshop, Canberra Australia
TPM
25kg
23kg
28kg
TPM Trends & Responses
Dr. Falk Chart – modified
26kg
PDRSRRSFRCA TRRCDR
ROM in
Proposal
Design Model
Bench Scale Model
Measurement
Detailed Design Model
Prototype Measurement
Flight 1st
Article
TechnicalPerformanceMeasure
VehicleWeight
125 PGCS 2018 Master Workshop, Canberra Australia
The Assessment Of Weight As A Function
Of Time
 At Contract Award there is a Proposal grade
estimate of vehicle weight
 At System Functional Review, the Concept of
Operations is validated for the weight
 At System Requirements Review the weight
targets are flowed down to the subsystems
components
 At PDR the CAD model starts the verification
process
 At CDR actual measurements are needed to
verify all models
 At Test Readiness Review we need to know
126
PGCS 2018 Master Workshop, Canberra Australia
TPM
Raison d'etre for
Technical Performance Measures
 The real purpose
of Technical
Performance
Measures is to
reveal
Programmatic and
Technical RISK
and engage in the
Risk Management
Process
PGCS 2018 Master Workshop, Canberra Australia
Risk
SOW
Cost
WBS
IMP/IMS
TPM
PMB
127
TPM
Buying Down Risk with TPMs
 “Buying down” risk is
planned in the IMS.
 MoE, MoP, and KPP
defined in the work
package for the
critical measure –
weight.
 If we can’t verify
we’ve succeeded,
then the risk did not
get reduced.
 The risk may have
gotten worse.
Risk: CEV-037 - Loss of Critical Functions During Descent
Planned Risk Level Planned (Solid=Linked, Hollow =Unlinked, Filled=Complete)
24
22
20
18
16
14
12
10
8
6
4
2
0
Conduct Force and Moment Wind
Develop analytical model to de
Conduct focus splinter review
Conduct Block 1 w ind tunnel te
Correlate the analytical model
Conduct w ind tunnel testing of
Conduct w ind tunnel testing of
Flight Application of Spacecra
CEV block 5 w ind tunnel testin
In-Flight development tests of
Damaged TPS flight test
31.Mar.05
5.Oct.05
3.Apr.06
3.Jul.06
15.Sep.06
1.Jun.07
1.Apr.08
1.Aug.08
1.Apr.09
1.Jan.10
16.Dec.10
1.Jul.11
Weight risk
reduced from
RED to Yellow
Weight confirmed
ready to fly – it’s
GREEN at this point
128
PGCS 2018 Master Workshop, Canberra Australia
TPM
Increasing Probability of Success Requires
Risk Management
 Going outside the
TPM limits always
means cost and
schedule impacts
 “Coloring Inside the
Lines” means
knowing the how to
keep the program
GREEN, or at least
stay close to GREEN
PGCS 2018 Master Workshop, Canberra Australia
129
TPM
Technical Performance Measures
Checklist
130
MoE MoP TPM
Traceable to needs, goals,
objectives, and risks
Traceable to applicable
MOEs, KPPs, system level
performance requirements,
and risks
Traceable to applicable MoPs,
system element performance,
requirements, objectives, risks,
and WBS elements
Defined with associated
KPPs
Focused on technical risks
and supports trades between
alternative solutions
Further decomposed, budgeted,
and allocated to lower level
system elements in the WBS
and IMS
Each MoE independent
from others
Provided insight into system
performance
Assigned an owner, the CAM
and Work Package Manager
Each MoE independent of
technical any solution
Decomposed, budgeted and
allocated to system elements
Sources of measure identified
and processes for generating the
measures defined.
Address the required
KPPs
Assigned an “owner,” the
CAM and Technical Manager
Integrated into the program’s
IMS as part of the exit criteria for
the Work Package
PGCS 2018 Master Workshop, Canberra Australia
TPM
All the program
performance data in
the is historical.
This past performance
data – by itself – is like
driving in the rear
view mirror.
What is needed is
Leading Indicators that
can be derived from
this past performance
data.
Integration
Creating an
Integrated Program
Performance
Management System
(IPPMS) starts with
the Five Principles,
their Processes, and
the Practices
131 PGCS 2018 Master Workshop, Canberra Australia
132
Integrated Program Management System (IPMS)
SOW
Risks
Metrics
 Schedule
 Cost
 Performance
CDRLs / DRDs
WBS
OBS
Program Events
Significant
Accomplishments
Accomplishment
Criteria
Work Packages
and Tasks
Weekly project
management
reviews % Cmplt
for PE/SA/AC
Monthly EVMS
Analysis by WBS
CA
WP
Management review
and analysis plan and
updates
Weekly Project
Status
Management
Review
 Critical Path
 Performance
analysis
 Metrics
 Risk analysis
Monthly EVMS
Metrics
 CPI
 SPI
 EAC
Delineated
Success
Measures
Critical efforts
to mature the
product
deliverables
Measures of product
maturity
Work efforts to
mature deliverables
Program Documents IMP / IMS Development Analysis Performance Reporting
Tasks
Product maturity and
traceability analysis
PGCS 2018 Master Workshop, Canberra Australia
133
Ashton Carter,
Secretary of Defense
PGCS 2018 Master Workshop, Canberra Australia
134
Primary Elements of Earned
Value
Cost
Technical
Performance
Schedule
Funding margin for
under performance Schedule margin for over
target baseline (OTB)
Schedule margin for
underperformance or
schedule extension
Over cost or under
performance
Over cost or
over schedule
Over schedule or
under
performing
PGCS 2018 Master Workshop, Canberra Australia
135
Connecting the EVM Variables with Technical
Performance Measures
Integrating Cost, Schedulele, and Technical Performance
Assures Program Management has the needed performance information to deliver
on‒time, on‒budget, and on‒specification
Technical Performance Measures
Cost Schedule
Conventional Earned Value
+
=
 Master Schedule is used to
derive Basis of Estimate
(BOE) not the other way
around.
 Probabilistic cost
estimating uses past
performance and cost risk
modeling.
 Labor, Materiel, and other
direct costs accounted for
in Work Packages.
 Risk adjustments for all
elements of cost.
Cost Baseline
 Earned Value is diluted by
missing technical
performance.
 Earned Value is diluted by
postponed features.
 Earned Value is diluted by
non compliant quality.
 All these dilutions require
adjustments to the
Estimate at Complete
(EAC) and the To Complete
Performance Index (TCPI).
Technical Performance
 Requirements are
decomposed into physical
deliverables.
 Deliverables are produced
through Work Packages.
 Work Packages are
assigned to accountable
manager.
 Work Packages are
sequenced to form the
highest value stream with
the lowest technical and
programmatic risk.
Schedule Baseline
PGCS 2018 Master Workshop, Canberra Australia
Connecting EVM and Technical Performance Measures,
we get the one and only way to measure progress with EV
 This is all that is needed to be successful with EVM
 Measure what has be completed compared what was planned to be
completed in units meaningful the decision makers
 That’s Physical Percent Complete
 Defining Physical Percent Complete starts and ends with
 MoE’s
 MoP’
 TPM’s
 Key Performance Parameters
136
PGCS 2018 Master Workshop, Canberra Australia
Integration
Earned Value and Earned Schedule can
provide Answer these four Critical
Questions137
PGCS 2018 Master Workshop, Canberra Australia
Integration
138 PGCS 2018 Master Workshop, Canberra Australia
Bibliography
139
PGCS 2018 Master Workshop, Canberra Australia
"Knowledge is of two kinds. We know a subject ourselves, or we know where we can
find information upon it.
When we enquire into any subject, the first thing we have to do is to know what
books have treated of it.
This leads us to look at catalogues, and at the backs of books in libraries.“
— Samuel Johnson (Boswell's Life of Johnson)
Bibliography Principles and Processes
 Systems Engineering Handbook: A Guide for System Life Cycle Processes and Activities,
INCOSE–TP–2003–002–03.1, August 2007, www.incose.org
 Systems Engineering: Coping with Complexity, Richard Stevens, Peter Brook, Ken Jackson, and
Stuart Arnold, Prentice Hall, 1998.
 Henri Fayol: Critical Evaluations in Business and Management, John C. Wood and Michael C.
Wood, Routedge, 2002.
 “Technical Performance Measurement, Earned Value, and Risk Management: An Integrated
Diagnostic Tool for Program Management,” Commander N. D. Pisano, SC, USN, Program
Executive Office for Air ASW, Assault, and Special Mission Programs (PEO(A))
140
PGCS 2018 Master Workshop, Canberra Australia
Bibliography Practices
 Modeling Homeland Security: A Value Focused Thinking Approach, Kristopher Pruitt, Air Force
Institute of Technology, March 2003
 Performance Based Earned Value, Paul Solomon.
 Guide to the Project Management Body of Knowledge, Project Management Institute.
 “No Silver Bullet: Essence and Accidents of Software Engineering, Fred Brooks, IEEE Computer,
10‒19, April 1987.
 Systems Requirements Analysis, Jeffry O. Grady, Academic Press, 2006.
 “Further Development in Earned Schedule,” Kym Henderson, The Measurable News, Spring 2004.
 “Schedule is Different,” Walter Lipke, The Measurable News, Summer 2003.
 “A Simulation and Evaluation of Earned Value Metrics to Forecast Project Duration,” M. S.
Vanhoucke, Journal of Operations Research Society, October 2007.
 Standard for Application and Management of the Systems Engineering Process, Institute of
Electrical and Electronics Engineers, 09–Sep–2005
 Systems Engineering: Coping with Complexity, Richard Stevens, Peter Brook, Ken Jackson, and
Stuart Arnold, Prentice Hall, 1998.
 The Requirements Engineering Handbook, Ralph R. Young, Artech House, 2004
141
PGCS 2018 Master Workshop, Canberra Australia
Bibliography Practices
 “Technical Performance Measurement, Earned Value, and Risk Management: An Integrated
Diagnostic Tool for Program Management,” Commander N. D. Pisano, SC, USN, Program
Executive Office for Air ASW, Assault, and Special Mission Programs (PEO(A))
 “Issues with Requirements Elicitation,” Michael G. Christel and Kyo C. Kang, Technical Report,
CMU/SEI–92–TR–12, Software Engineering Institute, Carnegie Mellon University Pittsburgh,
Pennsylvania 15213.
 “Managing complex product development projects with design structure matrices and domain
mapping matrices,” Mike Danilovic and Tyson Browning, International Journal of Project
Management, 25 (2007), pp. 300–314.
 “Architectural optimization using real options theory and Dependency structure matrices,” David M.
Sharman, Ali A. Yassine+, Paul Carlile, Proceedings of DETC ’02 ASME 2002 International Design
Engineering Technical Conferences 28th Design Automation Conference Montreal, Canada,
September 29–October 2, 2002.
 “Modeling and Analyzing Cost, Schedule, and Performance in Complex System Product
Development,” Tyson Browning, Massachusetts Institute of Technology, February 1999.
 The Integrated Project Management Handbook, Dayton Aerospace, 8 Feb 2002, Dayton Ohio.
 “Analytic Architecture for Capabilities–Based Planning, Mission–System Analysis, and
Transformation,” Paul K. Davis, RAND Corporation.
 Systems Engineering Handbook: A Guide for System Life Cycle Processes and Activities,
INCOSE–TP–2003–002–03.1, August 2007, www.incose.org
142
PGCS 2018 Master Workshop, Canberra Australia
Bibliography Capabilities
 “Analytic Architecture for Capabilities–Based Planning, Mission–System Analysis, and Transformation,”
Paul K. Davis, RAND Corporation.
 Portfolio–Analysis Methods for Assessing Capability Options, Paul K. Davis, Russell D. Shaver, and Justin
Beck, Rand Corporation, 2012
 “Architectural optimization using real options theory and Dependency structure matrices,” David M.
Sharman, Ali A. Yassine+, Paul Carlile, Proceedings of DETC ’02 ASME 2002 International Design
Engineering Technical Conferences 28th Design Automation Conference Montreal, Canada, September
29–October 2, 2002.
 “Modeling and Analyzing Cost, Schedule, and Performance in Complex System Product Development,”
Tyson Browning, Massachusetts Institute of Technology, February 1999.
 The Art of Systems Architecting, Mark W. Maier and Eberhardt Rechtin, CRC Press, 2000.
 Systems Engineering: Coping With Complexity, Richard Stevens, Peter Brook, Ken Jackson, and Stuart
Arnold, Prentice Hall, 1998.
 Assumption Based Planning: A Tool for Reducing Avoidable Surprises, James A. Dewar, Cambridge
University Press, 2002.
 Capabilities–Based Planning: A Methodology for Deciphering Commander’s Intent, Peter Kossakowski,
Evidence Based Research, Inc. 1595 Spring Hill Road, Suite 250 Vienna, VA 22182.
 Competing on Capabilities: The New Rules for Corporate Strategy, George Stalk, Philip Evans, and
Lawrence Shulman, Harvard Business Review, No. 92209, March–April 1992.
 Effects–Driven, Capabilities–Based, Planning for Operations, Maj Kira Jeffery, USAF and Mr Robert
Herslow.
 Effects–based Operations: Building Analytical Tools, Desmon Saunder–Newton and Aaron B. Frank,
Defense Horizons, October 2002, pp 1–8.
 Guide to Capability–Based Planning, Joint Systems and Analysis Group, MORS Workshop, October 2004,
Alexandria, VA. http://www.mors.org/
143
PGCS 2018 Master Workshop, Canberra Australia
Bibliography Requirements
 “Issues with Requirements Elicitation,” Michael G. Christel and Kyo C. Kang, Technical Report,
CMU/SEI–92–TR–12, Software Engineering Institute, Carnegie Mellon University Pittsburgh,
Pennsylvania 15213.
 The Requirements Engineering Handbook, Ralph R. Young, ArcTech House, 2004
 Software Risk Management, Barry W. Boehm, IEEE Computer Society Press, 1989.
 Software Requirements Analysis & Specifications, Alan M. Davis, Prentice Hall, 1990.
 Requirements Engineering: A Good Practice Guide, Ian Sommerville and Pete Sawyer, John Wiley
& Sons, 1997.
 Systems Requirements Practices, Jeffery O. Grady, McGraw Hill, 1993
 Four Key Requirements Engineering Techniques, Christof Ebert, IEEE Software, May / June 2006.
 Intent Specifications: An Approach to Building Human–Centered Specifications, Aeronautics and
Astronautics, MIT.
 Sample TCAS Intent Specification, Nancy Leveson and Jon Damon Reese, Software Engineering
Corporation .
144
PGCS 2018 Master Workshop, Canberra Australia
Bibliography Performance Measurement
Baseline
 “Managing complex product development projects with design structure matrices and domain
mapping matrices,” Mike Danilovic and Tyson Browning, International Journal of Project
Management, 25 (2007), pp. 300–314.
 MIL–STD–881A, Work Breakdown Structures.
 The Management of Projects, Peter W. G. Morris, Thomas Telford, 1994
 Modelling Complex Projects, Terry Williams, John Wiley & Sons. 2002.
 The Handbook of Program Management, James T. Brown, McGraw Hill, 2007.
 AntiPatterns in Project Management, William J. Brown, Hays W. McCormick III, and Scott W.
Thomas, John Wiley & Sons, 2000
 Earned Value Management, 3rd Edition, Quentin W. Fleming and Joel M. Koppelman, Project
Management Institute, 2005.
145
PGCS 2018 Master Workshop, Canberra Australia
Bibliography Continuous Risk Management
 “Understanding Risk Management in the DoD,” Mike Bolles, Acquisition Research Journal, Volume 10,
pp. 141–145, 2003.
 Effective Risk Management: Some Keys to Success, 2nd Edition, Edmund H. Conrow, AIAA Press, 2003.
 Managing Risk: methods for Software Systems Development, Elaine M. Hall, Addison Wesley, 1998
 Integrating Risk Management with Earned Value Management, National Defense Industry Association.
 Three point estimates and quantitative risk analysis a process guide for risk practitioners, Acquisitioning
Operating Framework, UK Ministry of Defense, http://www.aof.mod.uk/index.htm
 [Effective Opportunity Management for Project: Exploring Positive Risk, David Hillson, Taylor & Francis,
2004.
 [Catastrophe Disentanglement: Getting Software Projects Back on Track, E. M. Bennatan, Addison
Wesley, 2006.
 Software Engineering Risk Management, Dale Walter Karolak, IEEE Computer Society Press, 1998.
 Assessment and Control of Software Risks, Capers Jones, Prentice Hall, 1994
 Three Point Estimates and Quantitative Risk Analysis – A Process Guide For Risk Practitioners – version
1.2 May 2007 – Risk Management – AOF,
http://www.aof.mod.uk/aofcontent/tactical/risk/downloads/3pepracgude.pdf
 How much risk is too much risk?, Tim Lister, Boston SPIN, January 20th, 2004.
 Risk Management Maturity Level Development, INCOSE Risk Management Working Group, April 2002.
 “A Methodology for Project Risk Analysis using Bayesian Belief Networks within a Monte Carlo Simulation
Environment,” Javier F. Ordóñez Arízaga, University of Maryland, College Park, 2007.
 An Approach to Technology Risk Management, Ricardo Valerdi and Ron J. Kohl, Engineering Systems
Division Symposium, MIT, Cambridge, MA 29–31 March 2004.
146
PGCS 2018 Master Workshop, Canberra Australia
Bibliography Continuous Risk Management
 “An Industry Standard Risk Analysis Technique,” A Terry Bahill and Eric D. Smith, Engineering
Management Journal, Vol. 21 No 4., December 2009.
 Risk Management Process and Implementation, American Systems Corporation, 2003.
 The Basics of Monte Carlo Simulation: A Tutorial, S. Kandaswamy, Proceedings of the Project
Management Institute Seminars & Symposium, 1–10 November 2001.
 Bayesian Inference for NASA Probabilistic Risk and Reliability Analysis, NASA/SP–2009–569, June 2009.
 Common Elements of Risk, Christopher J. Alberts, CMU/SEI–2006–TN–014, April 2006.
 “Development of Risk Management Defense Extensions to the PMI Project Management Body of
Knowledge,” Edmund H. Conrow, Acquisition Review Quarterly, Spring 2003.
 Continuous Risk Management Guidebook, Audrey J. Dorofee, Julie A. Walker, Christopher J. Alberts,
Ronald P. Higuera, Richard L. Murphy, and Ray C. Williams, Software Engineering Institute, 1996.
 Risk Management Guide for DoD Acquisition, Fifth Edition, June 2002.
 Emerging Practice: Joint Cost & Schedule Risk Analysis, Eric Drucker, 2009 St. Louis SCEA Chapter Fall
Symposium, St’ Louis, MO.
 Making Risk Management Tools More Credible: Calibrating the Risk Cube, Richard L. Coleman, Jessica
R. Summerville, and Megan E. Dameron, SCEA 2006, Washington, D.C., 12 June 2006.
 A Theory of Modeling Correlations for Use in Cost–Risk Analysis, Stephen A. Book, Third Annual Project
Management Conference, NASA, Galveston, TX, 21–22 March 2006.
 The Joint Confidence Level Paradox: A History of Denial, 2009 NASA Cost Symposium, 28 April 2009.
147
PGCS 2018 Master Workshop, Canberra Australia
148 PGCS 2018 Master Workshop, Canberra Australia

Weitere ähnliche Inhalte

Was ist angesagt?

Balanced Scorecard Strategies And Hr Metrics Workshop Sept 13 Chicago, Il
Balanced Scorecard Strategies And Hr Metrics   Workshop Sept 13   Chicago, IlBalanced Scorecard Strategies And Hr Metrics   Workshop Sept 13   Chicago, Il
Balanced Scorecard Strategies And Hr Metrics Workshop Sept 13 Chicago, IlFrederick 'Rick' Buchman
 
Anton Rossouw - Strategic Approaches And Tools For Managing Complex Projects
Anton Rossouw - Strategic Approaches And Tools For Managing Complex ProjectsAnton Rossouw - Strategic Approaches And Tools For Managing Complex Projects
Anton Rossouw - Strategic Approaches And Tools For Managing Complex Projectsantonrossouw
 
PMP Flash Card
PMP Flash CardPMP Flash Card
PMP Flash Cardasim78
 
IMP / IMS Step by Step
IMP / IMS Step by StepIMP / IMS Step by Step
IMP / IMS Step by StepGlen Alleman
 
Assessment of Risks in International EPC Projects Reference Current Global Ec...
Assessment of Risks in International EPC Projects Reference Current Global Ec...Assessment of Risks in International EPC Projects Reference Current Global Ec...
Assessment of Risks in International EPC Projects Reference Current Global Ec...HIMADRI BANERJI
 
Risk Assessment, Mitigation And Management In Epc Projects With Case Study By...
Risk Assessment, Mitigation And Management In Epc Projects With Case Study By...Risk Assessment, Mitigation And Management In Epc Projects With Case Study By...
Risk Assessment, Mitigation And Management In Epc Projects With Case Study By...HIMADRI BANERJI
 
#Project Management fundamentals - By SN Panigrahi
#Project Management fundamentals - By SN Panigrahi#Project Management fundamentals - By SN Panigrahi
#Project Management fundamentals - By SN PanigrahiSN Panigrahi, PMP
 
Project Management Professional (PMP) / 42 Processes
Project Management Professional (PMP) / 42 ProcessesProject Management Professional (PMP) / 42 Processes
Project Management Professional (PMP) / 42 Processesedtichy
 
Integrated Master Schedule
Integrated Master ScheduleIntegrated Master Schedule
Integrated Master Scheduleellefsonj
 
Strategic Planning : the best way to manage EPC Contract Risks
Strategic Planning : the best way to manage EPC Contract RisksStrategic Planning : the best way to manage EPC Contract Risks
Strategic Planning : the best way to manage EPC Contract RisksPankaj K Sinha
 
Repositioning the Value of the Architecture Practice
Repositioning the Value of the Architecture PracticeRepositioning the Value of the Architecture Practice
Repositioning the Value of the Architecture PracticeEnterprise Architects
 

Was ist angesagt? (20)

Balanced Scorecard Strategies And Hr Metrics Workshop Sept 13 Chicago, Il
Balanced Scorecard Strategies And Hr Metrics   Workshop Sept 13   Chicago, IlBalanced Scorecard Strategies And Hr Metrics   Workshop Sept 13   Chicago, Il
Balanced Scorecard Strategies And Hr Metrics Workshop Sept 13 Chicago, Il
 
Anton Rossouw - Strategic Approaches And Tools For Managing Complex Projects
Anton Rossouw - Strategic Approaches And Tools For Managing Complex ProjectsAnton Rossouw - Strategic Approaches And Tools For Managing Complex Projects
Anton Rossouw - Strategic Approaches And Tools For Managing Complex Projects
 
080613 Mega-Project Schedule Integration & Management RCF Method-1
080613 Mega-Project Schedule Integration & Management RCF Method-1 080613 Mega-Project Schedule Integration & Management RCF Method-1
080613 Mega-Project Schedule Integration & Management RCF Method-1
 
Forensic schedule analysis acesss
Forensic schedule analysis acesssForensic schedule analysis acesss
Forensic schedule analysis acesss
 
PMP Flash Card
PMP Flash CardPMP Flash Card
PMP Flash Card
 
IMP / IMS Step by Step
IMP / IMS Step by StepIMP / IMS Step by Step
IMP / IMS Step by Step
 
Project Scheduling & Controls
Project Scheduling & ControlsProject Scheduling & Controls
Project Scheduling & Controls
 
The Beyond Budgeting Principles
The Beyond Budgeting PrinciplesThe Beyond Budgeting Principles
The Beyond Budgeting Principles
 
Assessment of Risks in International EPC Projects Reference Current Global Ec...
Assessment of Risks in International EPC Projects Reference Current Global Ec...Assessment of Risks in International EPC Projects Reference Current Global Ec...
Assessment of Risks in International EPC Projects Reference Current Global Ec...
 
Risk Assessment, Mitigation And Management In Epc Projects With Case Study By...
Risk Assessment, Mitigation And Management In Epc Projects With Case Study By...Risk Assessment, Mitigation And Management In Epc Projects With Case Study By...
Risk Assessment, Mitigation And Management In Epc Projects With Case Study By...
 
Pmp study-notes
Pmp study-notesPmp study-notes
Pmp study-notes
 
PMO-Framework
PMO-FrameworkPMO-Framework
PMO-Framework
 
#Project Management fundamentals - By SN Panigrahi
#Project Management fundamentals - By SN Panigrahi#Project Management fundamentals - By SN Panigrahi
#Project Management fundamentals - By SN Panigrahi
 
The PMO journey
The PMO journeyThe PMO journey
The PMO journey
 
Project Management Professional (PMP) / 42 Processes
Project Management Professional (PMP) / 42 ProcessesProject Management Professional (PMP) / 42 Processes
Project Management Professional (PMP) / 42 Processes
 
Emerging Trends in PMO Governance
Emerging Trends in PMO Governance Emerging Trends in PMO Governance
Emerging Trends in PMO Governance
 
Why EPC
Why EPCWhy EPC
Why EPC
 
Integrated Master Schedule
Integrated Master ScheduleIntegrated Master Schedule
Integrated Master Schedule
 
Strategic Planning : the best way to manage EPC Contract Risks
Strategic Planning : the best way to manage EPC Contract RisksStrategic Planning : the best way to manage EPC Contract Risks
Strategic Planning : the best way to manage EPC Contract Risks
 
Repositioning the Value of the Architecture Practice
Repositioning the Value of the Architecture PracticeRepositioning the Value of the Architecture Practice
Repositioning the Value of the Architecture Practice
 

Ähnlich wie Getting To Done - A Master Class Workshop

Heliotropic Abundance
Heliotropic AbundanceHeliotropic Abundance
Heliotropic AbundanceGlen Alleman
 
Getting to Done (PMI Denver)
Getting to Done (PMI Denver)Getting to Done (PMI Denver)
Getting to Done (PMI Denver)Glen Alleman
 
Performance based planning in a nut shell (V5)
Performance based planning in a nut shell (V5)Performance based planning in a nut shell (V5)
Performance based planning in a nut shell (V5)Glen Alleman
 
Increasing the probability of project success using Earned Value Management
Increasing the probability of project success using Earned Value ManagementIncreasing the probability of project success using Earned Value Management
Increasing the probability of project success using Earned Value ManagementGlen Alleman
 
Five Immutable Principles of Program Success and Their Supporting Processes a...
Five Immutable Principles of Program Success and Their Supporting Processes a...Five Immutable Principles of Program Success and Their Supporting Processes a...
Five Immutable Principles of Program Success and Their Supporting Processes a...Glen Alleman
 
The Five Immutable Principles of Project DSuccess
The Five Immutable Principles of Project DSuccessThe Five Immutable Principles of Project DSuccess
The Five Immutable Principles of Project DSuccessGlen Alleman
 
A Guide for Capital Project Mamnagers
A Guide for Capital Project MamnagersA Guide for Capital Project Mamnagers
A Guide for Capital Project MamnagersGlen Alleman
 
Methodologies 1: Managing Agile Projects
Methodologies 1: Managing Agile ProjectsMethodologies 1: Managing Agile Projects
Methodologies 1: Managing Agile ProjectsInflectra
 
Project Success: The Basis of the Five Immutable Principles
Project Success: The Basis of the Five Immutable PrinciplesProject Success: The Basis of the Five Immutable Principles
Project Success: The Basis of the Five Immutable PrinciplesGlen Alleman
 
Five Immutable Principles of Project of Digital Transformation Success
Five Immutable Principles of Project of Digital Transformation SuccessFive Immutable Principles of Project of Digital Transformation Success
Five Immutable Principles of Project of Digital Transformation SuccessGlen Alleman
 
Methodologies 3: Using Spira for Waterfall
Methodologies 3: Using Spira for WaterfallMethodologies 3: Using Spira for Waterfall
Methodologies 3: Using Spira for WaterfallInflectra
 
SE1 - Integrating SE and PPM to Increase the Probability of Success
SE1 - Integrating SE and PPM to Increase the Probability of SuccessSE1 - Integrating SE and PPM to Increase the Probability of Success
SE1 - Integrating SE and PPM to Increase the Probability of SuccessGlen Alleman
 
Showing how to Increase the Probability of Project Success by applying the ...
Showing how to Increase the Probability of Project Success by applying the ...Showing how to Increase the Probability of Project Success by applying the ...
Showing how to Increase the Probability of Project Success by applying the ...Glen Alleman
 
Breaking the Project Failure Cycle
Breaking the Project Failure CycleBreaking the Project Failure Cycle
Breaking the Project Failure CycleGlen Alleman
 
PGCS 2019 Master Class Integrating SE with PPM
PGCS 2019 Master Class Integrating SE with PPMPGCS 2019 Master Class Integrating SE with PPM
PGCS 2019 Master Class Integrating SE with PPMGlen Alleman
 
Principles and Practices of Performance-Based Project Management®
Principles and Practices of Performance-Based Project Management®Principles and Practices of Performance-Based Project Management®
Principles and Practices of Performance-Based Project Management®Glen Alleman
 

Ähnlich wie Getting To Done - A Master Class Workshop (20)

Heliotropic Abundance
Heliotropic AbundanceHeliotropic Abundance
Heliotropic Abundance
 
Getting to Done (PMI Denver)
Getting to Done (PMI Denver)Getting to Done (PMI Denver)
Getting to Done (PMI Denver)
 
Performance based planning in a nut shell (V5)
Performance based planning in a nut shell (V5)Performance based planning in a nut shell (V5)
Performance based planning in a nut shell (V5)
 
Increasing the probability of project success using Earned Value Management
Increasing the probability of project success using Earned Value ManagementIncreasing the probability of project success using Earned Value Management
Increasing the probability of project success using Earned Value Management
 
Five Immutable Principles of Program Success and Their Supporting Processes a...
Five Immutable Principles of Program Success and Their Supporting Processes a...Five Immutable Principles of Program Success and Their Supporting Processes a...
Five Immutable Principles of Program Success and Their Supporting Processes a...
 
The Five Immutable Principles of Project DSuccess
The Five Immutable Principles of Project DSuccessThe Five Immutable Principles of Project DSuccess
The Five Immutable Principles of Project DSuccess
 
A Guide for Capital Project Mamnagers
A Guide for Capital Project MamnagersA Guide for Capital Project Mamnagers
A Guide for Capital Project Mamnagers
 
Methodologies 1: Managing Agile Projects
Methodologies 1: Managing Agile ProjectsMethodologies 1: Managing Agile Projects
Methodologies 1: Managing Agile Projects
 
Project Success: The Basis of the Five Immutable Principles
Project Success: The Basis of the Five Immutable PrinciplesProject Success: The Basis of the Five Immutable Principles
Project Success: The Basis of the Five Immutable Principles
 
Del
DelDel
Del
 
Five Immutable Principles of Project of Digital Transformation Success
Five Immutable Principles of Project of Digital Transformation SuccessFive Immutable Principles of Project of Digital Transformation Success
Five Immutable Principles of Project of Digital Transformation Success
 
Methodologies 3: Using Spira for Waterfall
Methodologies 3: Using Spira for WaterfallMethodologies 3: Using Spira for Waterfall
Methodologies 3: Using Spira for Waterfall
 
SE1 - Integrating SE and PPM to Increase the Probability of Success
SE1 - Integrating SE and PPM to Increase the Probability of SuccessSE1 - Integrating SE and PPM to Increase the Probability of Success
SE1 - Integrating SE and PPM to Increase the Probability of Success
 
Ashish_Revised Draft 2 (1)
Ashish_Revised Draft 2 (1)Ashish_Revised Draft 2 (1)
Ashish_Revised Draft 2 (1)
 
Profile Seema Wadhwa
Profile Seema WadhwaProfile Seema Wadhwa
Profile Seema Wadhwa
 
Showing how to Increase the Probability of Project Success by applying the ...
Showing how to Increase the Probability of Project Success by applying the ...Showing how to Increase the Probability of Project Success by applying the ...
Showing how to Increase the Probability of Project Success by applying the ...
 
Breaking the Project Failure Cycle
Breaking the Project Failure CycleBreaking the Project Failure Cycle
Breaking the Project Failure Cycle
 
PGCS 2019 Master Class Integrating SE with PPM
PGCS 2019 Master Class Integrating SE with PPMPGCS 2019 Master Class Integrating SE with PPM
PGCS 2019 Master Class Integrating SE with PPM
 
Principles and Practices of Performance-Based Project Management®
Principles and Practices of Performance-Based Project Management®Principles and Practices of Performance-Based Project Management®
Principles and Practices of Performance-Based Project Management®
 
Psp Tsp Agile 3 1 En
Psp Tsp Agile 3 1 EnPsp Tsp Agile 3 1 En
Psp Tsp Agile 3 1 En
 

Mehr von Glen Alleman

Managing risk with deliverables planning
Managing risk with deliverables planningManaging risk with deliverables planning
Managing risk with deliverables planningGlen Alleman
 
A Gentle Introduction to the IMP/IMS
A Gentle Introduction to the IMP/IMSA Gentle Introduction to the IMP/IMS
A Gentle Introduction to the IMP/IMSGlen Alleman
 
Increasing the Probability of Project Success
Increasing the Probability of Project SuccessIncreasing the Probability of Project Success
Increasing the Probability of Project SuccessGlen Alleman
 
Process Flow and Narrative for Agile+PPM
Process Flow and Narrative for Agile+PPMProcess Flow and Narrative for Agile+PPM
Process Flow and Narrative for Agile+PPMGlen Alleman
 
Practices of risk management
Practices of risk managementPractices of risk management
Practices of risk managementGlen Alleman
 
Principles of Risk Management
Principles of Risk ManagementPrinciples of Risk Management
Principles of Risk ManagementGlen Alleman
 
Deliverables Based Planning, PMBOK® and 5 Immutable Principles of Project Suc...
Deliverables Based Planning, PMBOK® and 5 Immutable Principles of Project Suc...Deliverables Based Planning, PMBOK® and 5 Immutable Principles of Project Suc...
Deliverables Based Planning, PMBOK® and 5 Immutable Principles of Project Suc...Glen Alleman
 
From Principles to Strategies for Systems Engineering
From Principles to Strategies for Systems EngineeringFrom Principles to Strategies for Systems Engineering
From Principles to Strategies for Systems EngineeringGlen Alleman
 
NAVAIR Integrated Master Schedule Guide guide
NAVAIR Integrated Master Schedule Guide guideNAVAIR Integrated Master Schedule Guide guide
NAVAIR Integrated Master Schedule Guide guideGlen Alleman
 
Building a Credible Performance Measurement Baseline
Building a Credible Performance Measurement BaselineBuilding a Credible Performance Measurement Baseline
Building a Credible Performance Measurement BaselineGlen Alleman
 
Integrated master plan methodology (v2)
Integrated master plan methodology (v2)Integrated master plan methodology (v2)
Integrated master plan methodology (v2)Glen Alleman
 
DHS - Using functions points to estimate agile development programs (v2)
DHS - Using functions points to estimate agile development programs (v2)DHS - Using functions points to estimate agile development programs (v2)
DHS - Using functions points to estimate agile development programs (v2)Glen Alleman
 
Making the impossible possible
Making the impossible possibleMaking the impossible possible
Making the impossible possibleGlen Alleman
 
Capabilities based planning
Capabilities based planningCapabilities based planning
Capabilities based planningGlen Alleman
 
Process Flow and Narrative for Agile
Process Flow and Narrative for AgileProcess Flow and Narrative for Agile
Process Flow and Narrative for AgileGlen Alleman
 
Building the Performance Measurement Baseline
Building the Performance Measurement BaselineBuilding the Performance Measurement Baseline
Building the Performance Measurement BaselineGlen Alleman
 
Program Management Office Lean Software Development and Six Sigma
Program Management Office Lean Software Development and Six SigmaProgram Management Office Lean Software Development and Six Sigma
Program Management Office Lean Software Development and Six SigmaGlen Alleman
 
Policy and Procedure Rollout
Policy and Procedure RolloutPolicy and Procedure Rollout
Policy and Procedure RolloutGlen Alleman
 
Project Management Theory
Project Management TheoryProject Management Theory
Project Management TheoryGlen Alleman
 
Increasing the Probability of Project Success with Five Principles and Practices
Increasing the Probability of Project Success with Five Principles and PracticesIncreasing the Probability of Project Success with Five Principles and Practices
Increasing the Probability of Project Success with Five Principles and PracticesGlen Alleman
 

Mehr von Glen Alleman (20)

Managing risk with deliverables planning
Managing risk with deliverables planningManaging risk with deliverables planning
Managing risk with deliverables planning
 
A Gentle Introduction to the IMP/IMS
A Gentle Introduction to the IMP/IMSA Gentle Introduction to the IMP/IMS
A Gentle Introduction to the IMP/IMS
 
Increasing the Probability of Project Success
Increasing the Probability of Project SuccessIncreasing the Probability of Project Success
Increasing the Probability of Project Success
 
Process Flow and Narrative for Agile+PPM
Process Flow and Narrative for Agile+PPMProcess Flow and Narrative for Agile+PPM
Process Flow and Narrative for Agile+PPM
 
Practices of risk management
Practices of risk managementPractices of risk management
Practices of risk management
 
Principles of Risk Management
Principles of Risk ManagementPrinciples of Risk Management
Principles of Risk Management
 
Deliverables Based Planning, PMBOK® and 5 Immutable Principles of Project Suc...
Deliverables Based Planning, PMBOK® and 5 Immutable Principles of Project Suc...Deliverables Based Planning, PMBOK® and 5 Immutable Principles of Project Suc...
Deliverables Based Planning, PMBOK® and 5 Immutable Principles of Project Suc...
 
From Principles to Strategies for Systems Engineering
From Principles to Strategies for Systems EngineeringFrom Principles to Strategies for Systems Engineering
From Principles to Strategies for Systems Engineering
 
NAVAIR Integrated Master Schedule Guide guide
NAVAIR Integrated Master Schedule Guide guideNAVAIR Integrated Master Schedule Guide guide
NAVAIR Integrated Master Schedule Guide guide
 
Building a Credible Performance Measurement Baseline
Building a Credible Performance Measurement BaselineBuilding a Credible Performance Measurement Baseline
Building a Credible Performance Measurement Baseline
 
Integrated master plan methodology (v2)
Integrated master plan methodology (v2)Integrated master plan methodology (v2)
Integrated master plan methodology (v2)
 
DHS - Using functions points to estimate agile development programs (v2)
DHS - Using functions points to estimate agile development programs (v2)DHS - Using functions points to estimate agile development programs (v2)
DHS - Using functions points to estimate agile development programs (v2)
 
Making the impossible possible
Making the impossible possibleMaking the impossible possible
Making the impossible possible
 
Capabilities based planning
Capabilities based planningCapabilities based planning
Capabilities based planning
 
Process Flow and Narrative for Agile
Process Flow and Narrative for AgileProcess Flow and Narrative for Agile
Process Flow and Narrative for Agile
 
Building the Performance Measurement Baseline
Building the Performance Measurement BaselineBuilding the Performance Measurement Baseline
Building the Performance Measurement Baseline
 
Program Management Office Lean Software Development and Six Sigma
Program Management Office Lean Software Development and Six SigmaProgram Management Office Lean Software Development and Six Sigma
Program Management Office Lean Software Development and Six Sigma
 
Policy and Procedure Rollout
Policy and Procedure RolloutPolicy and Procedure Rollout
Policy and Procedure Rollout
 
Project Management Theory
Project Management TheoryProject Management Theory
Project Management Theory
 
Increasing the Probability of Project Success with Five Principles and Practices
Increasing the Probability of Project Success with Five Principles and PracticesIncreasing the Probability of Project Success with Five Principles and Practices
Increasing the Probability of Project Success with Five Principles and Practices
 

Kürzlich hochgeladen

Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024The Digital Insurer
 
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...apidays
 
Boost PC performance: How more available memory can improve productivity
Boost PC performance: How more available memory can improve productivityBoost PC performance: How more available memory can improve productivity
Boost PC performance: How more available memory can improve productivityPrincipled Technologies
 
GenCyber Cyber Security Day Presentation
GenCyber Cyber Security Day PresentationGenCyber Cyber Security Day Presentation
GenCyber Cyber Security Day PresentationMichael W. Hawkins
 
Partners Life - Insurer Innovation Award 2024
Partners Life - Insurer Innovation Award 2024Partners Life - Insurer Innovation Award 2024
Partners Life - Insurer Innovation Award 2024The Digital Insurer
 
WhatsApp 9892124323 ✓Call Girls In Kalyan ( Mumbai ) secure service
WhatsApp 9892124323 ✓Call Girls In Kalyan ( Mumbai ) secure serviceWhatsApp 9892124323 ✓Call Girls In Kalyan ( Mumbai ) secure service
WhatsApp 9892124323 ✓Call Girls In Kalyan ( Mumbai ) secure servicePooja Nehwal
 
[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdf[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdfhans926745
 
08448380779 Call Girls In Friends Colony Women Seeking Men
08448380779 Call Girls In Friends Colony Women Seeking Men08448380779 Call Girls In Friends Colony Women Seeking Men
08448380779 Call Girls In Friends Colony Women Seeking MenDelhi Call girls
 
🐬 The future of MySQL is Postgres 🐘
🐬  The future of MySQL is Postgres   🐘🐬  The future of MySQL is Postgres   🐘
🐬 The future of MySQL is Postgres 🐘RTylerCroy
 
TrustArc Webinar - Stay Ahead of US State Data Privacy Law Developments
TrustArc Webinar - Stay Ahead of US State Data Privacy Law DevelopmentsTrustArc Webinar - Stay Ahead of US State Data Privacy Law Developments
TrustArc Webinar - Stay Ahead of US State Data Privacy Law DevelopmentsTrustArc
 
Driving Behavioral Change for Information Management through Data-Driven Gree...
Driving Behavioral Change for Information Management through Data-Driven Gree...Driving Behavioral Change for Information Management through Data-Driven Gree...
Driving Behavioral Change for Information Management through Data-Driven Gree...Enterprise Knowledge
 
How to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerHow to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerThousandEyes
 
Raspberry Pi 5: Challenges and Solutions in Bringing up an OpenGL/Vulkan Driv...
Raspberry Pi 5: Challenges and Solutions in Bringing up an OpenGL/Vulkan Driv...Raspberry Pi 5: Challenges and Solutions in Bringing up an OpenGL/Vulkan Driv...
Raspberry Pi 5: Challenges and Solutions in Bringing up an OpenGL/Vulkan Driv...Igalia
 
Tata AIG General Insurance Company - Insurer Innovation Award 2024
Tata AIG General Insurance Company - Insurer Innovation Award 2024Tata AIG General Insurance Company - Insurer Innovation Award 2024
Tata AIG General Insurance Company - Insurer Innovation Award 2024The Digital Insurer
 
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...Drew Madelung
 
Handwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed textsHandwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed textsMaria Levchenko
 
Breaking the Kubernetes Kill Chain: Host Path Mount
Breaking the Kubernetes Kill Chain: Host Path MountBreaking the Kubernetes Kill Chain: Host Path Mount
Breaking the Kubernetes Kill Chain: Host Path MountPuma Security, LLC
 
08448380779 Call Girls In Greater Kailash - I Women Seeking Men
08448380779 Call Girls In Greater Kailash - I Women Seeking Men08448380779 Call Girls In Greater Kailash - I Women Seeking Men
08448380779 Call Girls In Greater Kailash - I Women Seeking MenDelhi Call girls
 
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptxEIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptxEarley Information Science
 
Automating Google Workspace (GWS) & more with Apps Script
Automating Google Workspace (GWS) & more with Apps ScriptAutomating Google Workspace (GWS) & more with Apps Script
Automating Google Workspace (GWS) & more with Apps Scriptwesley chun
 

Kürzlich hochgeladen (20)

Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024
 
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
 
Boost PC performance: How more available memory can improve productivity
Boost PC performance: How more available memory can improve productivityBoost PC performance: How more available memory can improve productivity
Boost PC performance: How more available memory can improve productivity
 
GenCyber Cyber Security Day Presentation
GenCyber Cyber Security Day PresentationGenCyber Cyber Security Day Presentation
GenCyber Cyber Security Day Presentation
 
Partners Life - Insurer Innovation Award 2024
Partners Life - Insurer Innovation Award 2024Partners Life - Insurer Innovation Award 2024
Partners Life - Insurer Innovation Award 2024
 
WhatsApp 9892124323 ✓Call Girls In Kalyan ( Mumbai ) secure service
WhatsApp 9892124323 ✓Call Girls In Kalyan ( Mumbai ) secure serviceWhatsApp 9892124323 ✓Call Girls In Kalyan ( Mumbai ) secure service
WhatsApp 9892124323 ✓Call Girls In Kalyan ( Mumbai ) secure service
 
[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdf[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdf
 
08448380779 Call Girls In Friends Colony Women Seeking Men
08448380779 Call Girls In Friends Colony Women Seeking Men08448380779 Call Girls In Friends Colony Women Seeking Men
08448380779 Call Girls In Friends Colony Women Seeking Men
 
🐬 The future of MySQL is Postgres 🐘
🐬  The future of MySQL is Postgres   🐘🐬  The future of MySQL is Postgres   🐘
🐬 The future of MySQL is Postgres 🐘
 
TrustArc Webinar - Stay Ahead of US State Data Privacy Law Developments
TrustArc Webinar - Stay Ahead of US State Data Privacy Law DevelopmentsTrustArc Webinar - Stay Ahead of US State Data Privacy Law Developments
TrustArc Webinar - Stay Ahead of US State Data Privacy Law Developments
 
Driving Behavioral Change for Information Management through Data-Driven Gree...
Driving Behavioral Change for Information Management through Data-Driven Gree...Driving Behavioral Change for Information Management through Data-Driven Gree...
Driving Behavioral Change for Information Management through Data-Driven Gree...
 
How to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerHow to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected Worker
 
Raspberry Pi 5: Challenges and Solutions in Bringing up an OpenGL/Vulkan Driv...
Raspberry Pi 5: Challenges and Solutions in Bringing up an OpenGL/Vulkan Driv...Raspberry Pi 5: Challenges and Solutions in Bringing up an OpenGL/Vulkan Driv...
Raspberry Pi 5: Challenges and Solutions in Bringing up an OpenGL/Vulkan Driv...
 
Tata AIG General Insurance Company - Insurer Innovation Award 2024
Tata AIG General Insurance Company - Insurer Innovation Award 2024Tata AIG General Insurance Company - Insurer Innovation Award 2024
Tata AIG General Insurance Company - Insurer Innovation Award 2024
 
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
 
Handwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed textsHandwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed texts
 
Breaking the Kubernetes Kill Chain: Host Path Mount
Breaking the Kubernetes Kill Chain: Host Path MountBreaking the Kubernetes Kill Chain: Host Path Mount
Breaking the Kubernetes Kill Chain: Host Path Mount
 
08448380779 Call Girls In Greater Kailash - I Women Seeking Men
08448380779 Call Girls In Greater Kailash - I Women Seeking Men08448380779 Call Girls In Greater Kailash - I Women Seeking Men
08448380779 Call Girls In Greater Kailash - I Women Seeking Men
 
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptxEIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
 
Automating Google Workspace (GWS) & more with Apps Script
Automating Google Workspace (GWS) & more with Apps ScriptAutomating Google Workspace (GWS) & more with Apps Script
Automating Google Workspace (GWS) & more with Apps Script
 

Getting To Done - A Master Class Workshop

  • 1. Master Class Principles, Processes, Practices, and Tools to Increase the Probability of Successfully Complete Project’s On‒Time, On‒Budget, with Needed Capabilities Glen B. Alleman Thursday 16th August 2018 9:00 AM ‒ 5:00 PM Getting to Done 1
  • 2. Who Am I?  I was educated as a Physicist, but practiced software development for radar, signal processing systems, and further educated as a Systems Engineer.  Moved to managing Software Intensive System of Systems (SISoS) for space, defense, enterprise IT, industrial systems, process control, and document management companies.  Worked in heavy construction, electric utilities (conventional & nuclear), BioPharma, Federal and State IT systems, Policy and Root Cause Analysis for a DOD FFRDC firm.  Developed and deployed program management systems based on the Principles, Processes, and Practices in the this workshop.  Some charts here are from Performance‒Based Project Management®, Glen B. Alleman, American Management Association, 2014. 2 PGCS 2018 Master Workshop, Canberra Australia
  • 3. These Principles, Practices, Processes, and Tools Have Been Applied with Success At …3 PGCS 2018 Master Workshop, Canberra Australia
  • 4. TLO’s for the Master Workshop  Introduce the Five Immutable Principles and the Processes that implement them  Show how the 10 Practices enable project success through the Five Principles and Five Processes  Apply these Principles, Processes, and Practices to a Real project ‒ TSAS  Or a project selected by the class  Develop an artifacts package for the project 4 PGCS 2018 Master Workshop, Canberra Australia
  • 5. 5 PGCS 2018 Master Workshop, Canberra Australia Why is it so hard to accept the truth that Principles must be in place, before any Practices and Process can be applied that might increase the Probability of Success?
  • 6. 1. What Does DONE Look Like in Units of Measure meaningful to the Decision Makers? 2. How Can We Get to DONE? 3. Is There Enough Time, Money, and Resources, to Get to DONE? 4. What Impediments Will We Encountered Along The Way to DONE and How can They be Removed? 5. What Meaningful Units of All Successful Projects Require Credible Answers To These Five Questions … 6 PGCS 2018 Master Workshop, Canberra Australia
  • 7. All Project Success Starts with the First Principle of the Five Immutable Principles  The needed Capabilities, stated as Measures of Effectiveness and Measures of Performance define what Done looks like.  These capabilities trace Value to the Strategy.  Capabilities lay the ground for adapting to change found on all projects with emerging requirements.  Features and Functions fulfill the stated Requirements needed to implement the Capabilities.  Capabilities provide the means to address unstated future requirements. 7 PGCS 2018 Master Workshop, Canberra Australia
  • 8. 1. Capabilities drive requirements 2. Requirements identify Technical and Process deliverables 3. Work Packages describes deliverables 4. IMS arranges deliverables 5. WP progress measured as Physical Percent Complete 10 Practices 1. Identify Needed Capabilities 2. Identify Requirements Baseline 3. Establish Performance Measurement Baseline 4. Execute the PMB 5. Continuous Risk Management 5 Processes There are 5 Process and 10 Practices that go along with the 5 Principles 1. What Does Done Look Like? 2. What is the Plan to reach Done? 3. What Resource are needed to reach Done? 4. What Impediments will be encountered along the wat y to Done? 5. What are the measures of progress to Plan? 5 Principles 6. WA assures WP produce deliverables in planned order 7. EV/ES describes performance to plan 8. Conformance with TPM adjusts EV 9. Feedback adjusts WP sequence and resource allocation 10. Future performance based on TCPI, IEAC, and adjusted work sequence 8 PGCS 2018 Master Workshop, Canberra Australia
  • 9. 1. Capabilities drive requirements 2. Requirements identify Technical and Process deliverables 3. Work Packages describes deliverables 4. IMS arranges deliverables 5. WP progress measured as Physical Percent Complete 10 Practices 1. Identify Needed Capabilities 2. Identify Requirements Baseline 3. Establish Performance Measurement Baseline 4. Execute the PMB 5. Continuous Risk Management 5 Processes In time allotted for this course, let’s focus on … 1. What Does Done Look Like? 2. What is the Plan to reach Done? 3. What Resource are needed to reach Done? 4. What Impediments will be encountered along the way to Done? 5. What are the measures of progress to Plan? 5 Principles 6. WA assures WP produce deliverables in planned order 7. EV/ES describes performance to plan 8. Conformance with TPM adjusts EV 9. Feedback adjusts WP sequence and resource allocation 10. Future performance based on TCPI, IEAC, and adjusted work sequence 9 PGCS 2018 Master Workshop, Canberra Australia Introduction
  • 10. Every Project Manager in Every Domain is Faced with Balancing Cost, Schedule, and Technical Performance on the Path to Done Let’s Learn how Principles, Processes and Practices can Increase the Probability of Project Success (PoPS) 10 PGCS 2018 Master Workshop, Canberra Australia
  • 11. Failure to define Done in a meaningful way, means the project starts out without a Mission, Vision, or Goal and without any associated metrics. Progress can only then be measured by the passage of time and consumption of money. For success, progress must be measured by the Effectiveness and Performance of its outcomes. Technical Performance Measures and Key Performance Parameters inform the Measures of Effectiveness (MoE) and Measures of Performance MoP. Introduction Project Success Starts with the 1st of 5 Immutable Principles What Does Done Look Like in Units of Measure Meaningful to the Decision Makers 11 PGCS 2018 Master Workshop, Canberra Australia
  • 12. Let’s Get Started … … But First Some Background Needed to Reach The End PGCS 2018 Master Workshop, Canberra Australia12
  • 13. PGCS 2018 Master Workshop, Canberra Australia How Can We Connect The Moving Parts Of Principles, Processes, and Practices Into A Cohesive Whole? 13
  • 14. These Connections Start with Three Core Frameworks…  Data is the heart of Integrated Project Performance Management  Without data, we can’t make decisions about program’s performance meaningful to the decision makers ‒ Dollars, Time, and Technical Performance.  Without data, processes have nothing to work on.  Processes transform data into information.  This information is used to make decisions about the program.  The primary decision is how to correct or prevent undesirable variances to stay on plan to deliver needed capabilities.  People execute processes using data to increase Probability of Success for the program.  People varying various roles and responsibilities in the process and the creation and use of the data. 14 PGCS 2018 Master Workshop, Canberra Australia Introduction
  • 15. 16 Program Management Activities Program Enablers Program Process Capabilities Business Enablers 15 Booz Allen Hamilton US Department of Energy FramewPGCS 2018 Master Workshop, Canberra Australia Introduction
  • 16. The 4+1 Questions Every Successful Project Must Answer Capabilities Requirements Plans Execution 2. What technical and operational requirements are needed to deliver these capabilities? 1. What capabilities are needed to fulfill the Concept of Operations†, the Mission and Vision, or the Business System Requirements? 3. What schedule delivers the product or services on time to meet the requirements? 4. What periodic measures of physical percent complete assure progress to plan? What impediments to success, their mitigations, retirement plans, or “buy downs are in place to increase the probability of success?” + Continuous Risk Management † A Concept of Operations (ConOps) describes the characteristics of a system from the point of view of an individual who will use that system. It is used to communicate the quantitative and qualitative system characteristics to all stakeholders.      IntroductionPerformance–Based Project Management®, Copyright © Glen B. Alleman, 2012 - 201816 PGCS 2018 Master Workshop, Canberra Australia
  • 17. The Five Processes Increase the Maturity of the Project’s Deliverables Identify Capabilities Needed for Success Identify Requirements to Deliver Capabilities Establish Performance Measurement Baseline Execute Performance Measurement Baseline  Define Capabilities  Define Concept of Operations  Assess Needs, Cost, and Risk Impacts  Define Balanced and Feasible Alternatives  Fact Finding  Gather And Classify  Evaluate And Rationalize  Prioritize Requirements  Integrate And Validate  Decompose Scope  Assign Accountability  Arrange Work  Develop Budget  Assign Performance  Perform Work  Accumulate Performance Measures  Analyze Performance  Take Corrective Action Perform Continuous Risk Management (CRM) Define the Measurable Capabilities of each Project Outcome Assure All Requirements Provided In Support of Capabilities Define Measures of Performance and Effectiveness Ensure Cost, Schedule, and Technical Performance Compliance 17 PGCS 2018 Master Workshop, Canberra Australia Introduction
  • 18. What is a Deliverable? Project Management Case Study, Pierre Bonnal, CNAM IIM MBA Program, Jun 18 PGCS 2018 Master Workshop, Canberra Australia Introduction
  • 19. Define the set of capabilities needed to achieve the project objectives or the particular end state for a specific scenario. Using the Concept of Operations (ConOps), define the details of who, where, and how this capability is to be accomplished, employed, and executed.  Define the technical and operational requirements for the system capabilities to be fulfilled. First, define these requirements in terms isolated from any implementation details. Only then bind the requirements with technology.  Build a time–phased network of work activities describing the work to be performed, the budgeted cost for this work, the organizational elements that produce the deliverables, and the performance measures showing this work is proceeding according to plan.  Execute work activities, while assuring all performance assessment represent 100% completion before proceeding. This means – No rework, no forward transfer of activities to the future. Assure all requirements are traceable to work & all work is traceable to requirements.  Apply the processes of Continuous Risk Management for each Performance– Based Project Management® process area to: Identify, Analyze, Plan, Track, Control, and Communicate programmatic and technical risk.  IntroductionPerformance–Based Project Management®, Copyright © Glen B. Alleman, 2012 - 201819 PGCS 2018 Master Workshop, Canberra Australia
  • 20. Partition system capabilities into classes of service within operational scenarios. Connect the capabilities to system requirements using some visual modeling notation. Define Measures of Effectiveness (MoE) and Measures of Performance (MoP). Define the delivery schedule for each measure of performance and effectiveness. Define scenarios for each system capability. Connect these scenarios to a Value Stream Map of the increasing maturity of the program. Assess value flow through the map for each needed capability. Identify capability mismatches and make corrections to improve overall value flow. Assign costs to each system element using a value flow model. Assure risk, probabilistic cost and benefit performance attributes are defined. Use cost, schedule and technical performance probabilistic models to forecast potential risks to program performance. Make tradeoffs that connect cost, schedule, and technical performance in a single location that compares the tradeoffs and their impacts. Use Measures of Effectiveness (MoE) and Measures of Performance (MoP) for these alternative tradeoffs. Define the capabilities needed to achieve a desired objective or a particular end state for a specific scenario. Define the details of who, where, and how these capabilities are to be delivered and employed to fulfill the Mission and Vision 1.1 1.2 1.3 1.4  IntroductionPerformance–Based Project Management®, Copyright © Glen B. Alleman, 2012 - 201820 PGCS 2018 Master Workshop, Canberra Australia
  • 21. Produce an overall statement of the problem in the operational context. Develop the overall operational and technical objectives of the target system. Defined the boundaries and interfaces of the target system. Gather required system capabilities, functional, nonfunctional and environmental requirements, and design constraints. Build the Top Down capabilities and functional decomposition of the requirements in a Requirements Management System. Answer the question “why do I need this?” in terms of operational capabilities. Build a cost / benefit model using probabilistic assessment of all variables, their dependencies, and impacts. For all requirements, perform a risk assessment to cost and schedule. Determine criticality for the functions of the system. Determine trade off relationships for all requirements to be used when option decisions must be made. For all technical items, prioritize their cost and dependency. Address the completeness of requirements by removing all “TBD” items. Validate that the requirements are traceable to system capabilities, goals, and mission. Resolve any requirements inconsistencies and conflicts. Define the technical and operational requirements that must be met for the system capabilities to be delivered. Define these requirements in terms isolated from any technology or implementation. Assure each requirement is connected to a need system capability. 2.1 2.2 2.3 2.4 2.5  IntroductionPerformance–Based Project Management®, Copyright © Glen B. Alleman, 2012 - 201821 PGCS 2018 Master Workshop, Canberra Australia
  • 22. Build a time–phased network of activities describing the work to be performed, the budgeted cost for this work, the organizational elements that produce the deliverables from this work, and the performance measures showing this work is proceeding according to plan. Decompose the program Scope into a product based Work Breakdown Structure (WBS), then further into Work Packages describing the production of the deliverables traceable to the requirements, and to the needed capabilities. 3.1 Assign responsibility to Work Packages (the groupings of deliverables) to a named owner accountable for the management of the resource allocations, cost and schedule baseline, and technical delivery. 3.2 Arrange the Work Packages in a logical network with defined deliverables, milestones, internal and external dependencies, with credible schedule, cost, and technical performance margins. 3.3 Develop the Time–Phased Budgeted Cost for Work Scheduled (BCWS) for the labor and material costs in each Work Package and the Project as a whole. Assure proper resource allocations can be met and budget profiles match expectations of the program sponsor 3.4 Assign objective Measures of Performance (MoP) and Measures of Effectiveness (MoE) for each Work Package and summarize these for the Project as a whole. 3.5 Establish a Performance Measurement Baseline (PMB) used to forecast the Work Package and Project ongoing and completion cost and schedule performance metrics. 3.6  IntroductionPerformance–Based Project Management®, Copyright © Glen B. Alleman, 2012 - 201822 PGCS 2018 Master Workshop, Canberra Australia
  • 23. Execute the planned work, assuring all work is 100% complete before proceeding to the next planned work package. No rework, no forward transfer of activities or features. Assure every requirement is traceable to work and all work is traceable to requirements.  Using the Work Package sequencing, release work to be performed as planned.  With the responsibility assignments, identify the accountable delivery manager to guide the development of the products or services for each Work Package. 4.1  Using Physical Percent Complete or Apportioned Milestones, capture measures of progress to plan for each Work Package.  Report this Physical Percent Complete in a centralized database for each Work Package and the program as a whole. 4.2  Compare the Physical Percent Complete against the Planned Percent Complete for each period of performance.  Construct cost and schedule performance indices from this information and the Physical Percent complete measures. 4.3  With Cost and Schedule performance indices, construct a forecast of future performance of cost, schedule, and technical performance compliance.  Take management actions for any Work Packages not performing as planned. 4.4  Record past performance based on Work Package completion criteria.  Record past future forecast performance estimates in a historical database.  Forecast next future performance estimate against the Performance Measurement Baseline.  Report this next future performance estimate to the program stakeholders. 4.5  IntroductionPerformance–Based Project Management®, Copyright © Glen B. Alleman, 2012 - 201823 PGCS 2018 Master Workshop, Canberra Australia
  • 24.  Identify and classify risks in a Risk Register. Separate reducible and Irreducible risks  Manage this Risk Register through a Risk Management Board.  Connect these risks and their handling and margins in the Master Schedule.  Convert risk data into risk decision‒making information.  Use this analysis information as the decision basis for the program manager to work on the “right” risks.  Turn risk information into decision making information and actions (both present and future).  Develop actions to address individual risks, prioritize risk actions, and create an integrated risk management plan to retire the risk or handle it when it turned into an issue.  Monitor the status of risks and actions taken to ameliorate risks.  Identify and monitor risks to enable the evaluation of the status of risks themselves and of risk mitigation plans.  Risk communication lies at the center of the model to emphasize both its pervasiveness and its criticality.  Without effective communication, no risk management approach can be viable. Continuous Risk Management starts with the underlying principles, concepts, and functions of risk management and provides guidance on how to implement risk management as a continuous practice in programs and the organizations that management programs. 5.1 5.2 5.3 5.4 5.5  IntroductionPerformance–Based Project Management®, Copyright © Glen B. Alleman, 2012 - 201824 PGCS 2018 Master Workshop, Canberra Australia
  • 25. What Does A Capability “Sound” Like? 25 PGCS 2018 Master Workshop, Canberra Australia Introduction
  • 26. What Should We Do? Where Are We Now? Identifying Needed System Capabilities Abstracted from: “Capabilities‒Based Planning – How It Is Intended To Work And Challenges To Its Successful Implementation,” Col. Stephen K. Walker, United States Army, U. S. Army War College, March 2005 Identifying Needed System Capabilities 26 PGCS 2018 Master Workshop, Canberra Australia Introduction
  • 27. What Is a Requirement? 27 PGCS 2018 Master Workshop, Canberra Australia Introduction
  • 28. Identifying Requirements† † Systems Requirements Practices, Jeffery O. Grady, McGraw Hill, 1993 28 PGCS 2018 Master Workshop, Canberra Australia Introduction
  • 29. What Does a Credible Based Plan and Schedule Look Like? 29 PGCS 2018 Master Workshop, Canberra Australia Introduction
  • 30. Establishing the Three Elements of the Performance Measurement Baseline Cost Baseline Schedule Baseline Technical Baseline Determine Scope and Approach Develop Technical Logic Develop Technical Baseline Develop WBS Define Activities Estimate Time Durations Sequence Activities Finalize Schedule Identify Apportioned Milestones Determine Resource Requireme nt Prepare Cost Estimate Resource Load Schedule Finalize Apportioned Milestones Determine Funding Constraint s Approve PMB Perform Functional Analysis 30 PGCS 2018 Master Workshop, Canberra Australia Introduction
  • 31. How Do We Know We Are Making Progress to Plan? 31 PGCS 2018 Master Workshop, Canberra Australia Introduction
  • 32. Executing the Performance Measurement Baseline (PMB) 32 PGCS 2018 Master Workshop, Canberra Australia Introduction
  • 33. Analyze Plan Track Control Identify Risks, Issues, and Concerns Evaluate, classify, and prioritize risks Decide what should be done about each risk Monitor risk metrics and verify/validate mitigations Make risk decisions Subproject and partner data/constraints, hazard analysis, FMEA, FTA, etc. Risk data: test data, expert opinion, hazard analysis, FMEA, FTA, lessons learned, technical analysis Resources Replan Mitigation Program/project data (metrics information) Statement of Risk Risk classification, Likelihood Consequence, Timeframe Risk prioritization Research, Watch (tracking requirements) Acceptance Rationale, Mitigation Plans Risk status reports on: Risks Risk Mitigation Plans Close or Accept Risks Invoke contingency plans Continue to track Perform Continuous Risk Management 33 PGCS 2018 Master Workshop, Canberra Australia Introduction
  • 34. Integrating all the Parts into a Whole Risk Management SOW SOO ConOps WBS Technical and Operational Requirements CWBS & CWBS Dictionary Integrated Master Plan (IMP) Integrated Master Schedule (IMS) Earned Value Management System (EVMS) Objective Status and Essential Views to support the proactive management processes needed to keep the program GREEN Performance Measurement Baseline (PMB) Measures of Effectiveness (MOE) Measures of Performance (MOP) Measures of Cost – Schedule Progress JROC Key Performance Parameters (KPP) Program Specific Key Performance Parameters (KPP) Technical Performance Measures (TPM) CWBS 34 PGCS 2018 Master Workshop, Canberra Australia Introduction
  • 35. From Mission Capabilities to Done “Coming to Grips with Measures of Effectiveness,” N. Sproles, Systems Engineering, Volume 3, Number 1, pp. 50–58 MoE KPP MoP TPMMission Need Acquirer Defines the Needs and Capabilities in terms of Operational Scenarios Supplier Defines Physical Solutions that meet the needs of the Stakeholders Operational measures of success related to the achievement of the mission or operational objective being evaluated. Measures that characterize physical or functional attributes relating to the system operation. Measures used to assess design progress, compliance to performance requirements, and technical risks. 35 PGCS 2018 Master Workshop, Canberra Australia Introduction
  • 36. 4+1Critical Processes for Success Capabilities, Requirements & Deliverables Program Architecture & Dependencies Work Planning and Sequencing Performance Measurement Baseline Programmatic & Technical Risk Management Balanced Scorecard Concept of Operations Statement of Objectives Integrated Master Plan (IMP) Gaps Overlaps Interfaces Integrated Master Schedule (IMS) Schedule margin to protect critical deliverables Cost And Schedule Baseline WBS RAM Resource Loaded IMS Risk Registry Risk Handling Plans Contingency And Management Reserve Requirements Traceability Value Stream Mapping (VSM) Work Packages Planning Packages Technical Performance Measures (TPM) Risk Integrated With IMS Risk Trending Measures Of Effectiveness Design Structure Earned Value Measures Of Performance Monte Carlo 36 PGCS 2018 Master Workshop, Canberra Australia Introduction
  • 37. Getting On the Road to Success Means … Where are we going? How do we get there? Are there enough resources? What are impediments to progress? How do we measure progress? IntroductionPerformance–Based Project Management®, Copyright © Glen B. Alleman, 2012 - 201837 PGCS 2018 Master Workshop, Canberra Australia
  • 38. Practices of Performance– Based Project Management® guide the application of the 5 Principles and the 5 Process Areas in this handbook. These practices define the reasons for each process, connect each practice to a beneficial outcome, and integrate the processes into a seamless delivery system. Connections Connecting 5 Principles and 5 Processes With 10 Practices 38 PGCS 2018 Master Workshop, Canberra Australia
  • 39. Connecting Five Process with Five Practices of Project Success 10PracticesofProjectManagementSuccess Chapter I 5 Identify Needed Capabilities Establish a Performance Measurement Baseline Execute the Performance Measurement Baseline Capabilities Based Plan Operational Needs Earned Value Performance 0% /100% Technical Performance Measures Business or Mission Value Stream Technical Requirements Identify Requirements Baseline Technical Performance Measures PMB Changes to Needed Capabilities Changes to Requirements Baseline Changes to Performance Baseline ConnectionsPerformance–Based Project Management®, Copyright © Glen B. Alleman, 2012 - 201839 PGCS 2018 Master Workshop, Canberra Australia
  • 40. 10 Practices of Performance–Based Project Management® Assess the capabilities being provided through the deliverables Fulfill the requirements through effort held in the Work Packages Produce deliverables from Work Packages Planned BCWS Physical % Complete WP’s contain deliverables that fulfill requirements Capabilities topology defines requirements flow down WP flow must describe the increasing maturity of the product or service Producing the deliverables in the planned sequence maintains the value stream to the customer 40 PGCS 2018 Master Workshop, Canberra Australia Connections
  • 41. Enough Show and Tell, Let’s … 41 PGCS 2018 Master Workshop, Canberra Australia
  • 42. A Quick Reminder of the 5 Principles 42 Principle Evidence the Principle of Being Implemented What does Done Look Like? Integrated Master Plan (IMP) with Measures of Effectiveness (MoE) and Measures of Performance (MoP) What's the Plan and Schedule to get Done? Integrated Master Schedule (IMS) with Technical Performance Measures (TPM) and Key Performance Parameters (KPP) What resources do we need for Done? Resource loaded IMS What impediments well we encounter along the way? Risk adjusted IMS What are the units of measure of progress toward Done? Earned Value (EV) and Earned Schedule (ES) informed by compliance with compliance plans for MoE, MoP, TPM, KPP, Risk Buydown, and Margin Buydown PGCS 2018 Master Workshop, Canberra Australia Connections
  • 43. One More Reminder, Project Success is About Doing a Lot Things Right … … But in the End, it’s always About the Numbers Cost Numbers, Schedule Numbers, Risk Numbers, Technical Performance Numbers 43 PGCS 2018 Master Workshop, Canberra Australia Connections
  • 44. Integrating the Parts of a Credible Program Management System (PMS) Risk SOW Cost WBS IMP/IMS TPM PMB ❺ Deliverables defined in the SOW, traced to the WBS, with narratives and Measures of Performance (MoP) ❹ Budget at the Work Package level, rolled to the Control Accounts showing cost spreads for all work in the IMS ❻ Measures of Performance (MoP) for each critical deliverable in the WBS and identified in each Work Package in the IMS, used to assess maturity in the IMP ❶ The Products and Processes in a “well structured” decomposition, traceable to the deliverables ❷ Schedule contains all the Work Packages, Budget, Risk mitigation plans, with traces to the Plan measuring increasing maturity through Measures of Effectiveness (MoE) and KPPs (JROC and Program) ❸ Technical and Programmatic Risks Connected through the WBS, Risk Register, Plan and Schedule The Baseline is the Document of Record for the Program Performance is Measured through the PMB 44 PGCS 2018 Master Workshop, Canberra Australia Connections
  • 45. Before We Start the Hands On Part What does it mean to be Credible? The Simplest Answer is … Our Artifact's Are Believable A credible Integrated Master Plan, Integrated Master Schedule, Cost Baseline, Risk Baseline, Estimate At Completion, Physical Percent Complete 45 PGCS 2018 Master Workshop, Canberra Australia
  • 46. That are Connected in this Way 46 PGCS 2018 Master Workshop, Canberra Australia Connections
  • 47. The WBS is Paramount  The WBS defines the deliverables and the supporting processes that produce them  The WBS Dictionary describes the technical and operation behaviors that will be assessed during the development of the deliverables  The terminal nodes of the WBS define the deliverables produced by the Work Packages in the IMS and assessed through the IMP Accomplishment Criteria (AC) 47 PGCS 2018 Master Workshop, Canberra Australia Connections
  • 48. The IMP Starts with the Buyer  The IMP defines the measuring of increasing maturity for the deliverables as the program moves from left to right  Significant Accomplishments (SA) are defined by the Measures of Effectiveness (MoE)  Accomplishment Criteria (AC) are defined by the Measures of Performance (MoP)  Risks are assigned at all levels of the IMP and IMS 48 PGCS 2018 Master Workshop, Canberra Australia Connections
  • 49. Natural Uncertainties and Event Based Risks†  Natural uncertainties in cost and schedule processes create risks to completing on time and on budget  Event based risks create impacts to cost, schedule, and technical performance  Event based risks are handled through risk mitigations  Natural uncertainties are handled through in cost, schedule, and technical performance margins  To be credible, the PMB must include both type of risks with their handling strategies 49 † These are referred to in the literature as Aleatory and Epistemic. We’ll used the naturally occurring and Event Based in this introduction, but those terms are consider operational rather than mathematical. PGCS 2018 Master Workshop, Canberra Australia Connections
  • 50. Costs Assigned to Packages of Work  Labor and material cost are represented in the Integrated Master Schedule (IMS) and provide visibility to the probability of program success  Variances in labor and material costs are modeled in the same way as work durations  Event based risks impact both cost and schedule and are modeled in the PMB  Risk retirement cost is allocated for the work effort in response to Event Based risks 50 PGCS 2018 Master Workshop, Canberra Australia Connections
  • 51. Statement of Work  Work in the PMB starts with the Statement of Work and flows through the Work Breakdown Structure  Measures of Effective (MoE) and Measures of Performance (MoP) can be defined in the SOW or WBS Dictionary  Traceability from the IMP to the IMS to all performance measures in the SOW is the basis of program performance measurement 51 PGCS 2018 Master Workshop, Canberra Australia Connections
  • 52. Technical Performance Measures  Key Performance Parameters (KPP, both Acquisition owner and Program specific) and Technical Performance Measures (TPM) define how the deliverables complying with the Statement of Work, Concept of Operations, and CDRLs  TPMs inform the measures cost and schedule for delivered program outcomes  TPM, MoE, MoP, and KPPs provide assessment of the cost and schedule effectiveness 52 PGCS 2018 Master Workshop, Canberra Australia Connections
  • 53. TSAS Command and Data Center Mobile Sensors Ground Based Sensors 53 UAV with Airborne Sensors PGCS 2018 Master Workshop, Canberra Australia
  • 54. TSAS Command and Data Center Mobile Sensors Ground Based Sensors 54 UAV with Airborne Sensors PGCS 2018 Master Workshop, Canberra Australia
  • 55. Top Level Capabilities for Airborne Sensors 55 Capability Program Goal Maximum Range 2,000 NM Maximum Altitude 35,000 feet Maximum endurance 12 hours SATCOM Link 1.5 – 50 Mbps LOS Datalink > 50.0 Mbps Synthetic Aperture Radar (SAR) 1.0/0.3m resolution (WAS/Spot) Moving Target Indicator (MTI) 20 -200Km/10m Range resolution Electro Optical NIIRS 6.5/6.0 (Spot/WAS) Infrared NIIRS 5.5/5.0 (Spot/WAS) Wide Area Search 250 Sq. NMI/Day (50 x 50) Target Coverage 1,000 spot targets / day PGCS 2018 Master Workshop, Canberra Australia Connections
  • 56. TSAS Performance Measures Derived from those Top Level Capabilities 56 TSAS Element Measures Answers the Question Example SOW, SOO, ConOps Measures of Effectiveness (MoE) How Do We Know we are Accomplishing the Mission? We need the capability to Increase IED Placement search capabilities by 50% WBS Technical Performance Measure (TPM) What are we building and how do we know it meets the specifications that will accomplish the Mission? Systems, subsystems, and supporting processes for each deliverable IMP – PE and SA MoE for the Program Events and Significant Accomplishments How can we measure the increasing maturity of the deliverables in the narratives in the Capabilities Based documents Sensor payloads capable of IR and UV detectors within the avionics bay IMP – AC Measures of Performance (MoP) Technical Performance of the deliverables derived from the MoEs 100 square miles per hours search capabilities IMS Technical Performance Measures (TPM) How does the work increase the maturity of the deliverables? Sensor platform TPMs inside the bounds, on-time, on-schedule? Tasks CPI, SPI, TCPI What work is needed to increase the maturity of the deliverables? Cost and schedule matching TPM progress? Risk Register Identified risks, with handling strategies What are the Epistemic risks and how are they represented in the IMS? All aleatory risks included in duration and cost. Epistemic risk retirement PGCS 2018 Master Workshop, Canberra Australia Connections
  • 57. The Work Breakdown Structure is our starting point for developing all other elements needed for the Performance Measurement Baseline. The TSAS WBS is defined using the a standard. In the US that is MIL‒STD‒881C, with an appendix for UAVs. From this, the details of the avionics subsystems will be used for the development of the Integrated Master Plan and Integrated Master Schedule. WBS The WBS is Paramount 57 PGCS 2018 Master Workshop, Canberra Australia
  • 58. Integrating the Parts of a Credible Program Management System (PMS) Risk SOW Cost WBS IMP/IMS TPM PMB ❺ Deliverables defined in the SOW, traced to the WBS, with narratives and Measures of Performance (MoP) ❹ Budget at the Work Package level, rolled to the Control Accounts showing cost spreads for all work in the IMS ❻ Measures of Performance (MoP) for each critical deliverable in the WBS and identified in each Work Package in the IMS, used to assess maturity in the IMP ❶ The Products and Processes in a “well structured” decomposition, traceable to the deliverables ❷ Schedule contains all the Work Packages, Budget, Risk mitigation plans, with traces to the Plan measuring increasing maturity through Measures of Effectiveness (MoE) and KPPs (JROC and Program) ❸ Technical and Programmatic Risks Connected through the WBS, Risk Register, Plan and Schedule The Baseline is the Document of Record for the Program Performance is Measured through the PMB 58 PGCS 2018 Master Workshop, Canberra Australia WBS
  • 59. What is the Work Breakdown Structure?  Defines the total System or the System of Systems (SoS)  Provides the framework for planning, prioritizing, managing, and tracking all work done on the program  Products  Supporting services  Facilities  Provides the framework for  Cost Structure for estimating and cost reporting  Resource allocation  Status Reporting  Performance Measurement  Identify and managing program risk 59 PGCS 2018 Master Workshop, Canberra Australia WBS
  • 60. System Architecture Drives the WBS 60 PGCS 2018 Master Workshop, Canberra Australia WBS
  • 61.  End Products used to implement the mission  Based on the System (Physical) Architecture  Enabling Products and Services  Products and services required to develop, produce, and support the end items  Based on Life Cycle  The first three (end product) WBS levels:  Level 1: Overall System  Level 2: Major Element (or Segment)  Level 3: Subordinate Components (or Prime Items)  Levels 4+ continue the decomposition to the Configuration Item (CI) level Structure of the TSAS WBS 61 PGCS 2018 Master Workshop, Canberra Australia WBS
  • 62. Three WBS’s for TSAS  High-Level (First 3 levels)  Provides Program Structure  Generally developed/controlle d by Customer (Government)  Detailed (Levels 4+)  Provides framework for Contract Work Packages and Costing  Generally developed by Contractor  Generally follows Program WBS  Detailed (Level 4+)  Provides framework for Subcontract Work Packages and Costing  Generally developed by Subcontractor  Generally follows Contract WBS 62 PGCS 2018 Master Workshop, Canberra Australia WBS
  • 63. Level 1 (System of Systems) WBS 63 1.0 Tactical Situational Awareness System (TSAS) 1.1 Unmanned Air Vehicle (UAV) 1.2 Mobile Sensor System 1.3 Ground Based Sensors 1.4 Ground / Host Segment (Command and Control) 1.5 System of Systems Engineering 1.6 Program Management 1.7 System of Systems Test and Integration 1.8 System of Systems Training 1.9 Systems of Systems Data Packages 1.10 Peculiar Support Equipment 1.11 Common Support Equipment 1.12 Operational and Site Support 1.13 Industrial Facilities UAV with Airborne Sensors PGCS 2018 Master Workshop, Canberra Australia WBS
  • 64. Level 2 Unmanned Air Vehicle (UAV) 64 UAV with Airborne Sensors PGCS 2018 Master Workshop, Canberra Australia WBS
  • 65. Level 3 UAV Avionics 65 PGCS 2018 Master Workshop, Canberra Australia WBS
  • 66. Level 4 UAV Avionics – GN&C 66 PGCS 2018 Master Workshop, Canberra Australia WBS
  • 67. Level 4 UAV Avionics - MMS 67 PGCS 2018 Master Workshop, Canberra Australia WBS
  • 68. Building the IMP is a Systems Engineering activity. The Integrated Master Plan (IMP) is a Programmatic Architecture in the same way the hardware and software are a Product Architecture. Poor, weak, or unstructured Programmatic Architecture reduces visibility to the Product Architecture’s performance measures of cost and schedule connected with Technical Performance Measures. IMP The Measures of Effectiveness (MoE) and Measures of Performance (MoP) are held in the Integrated Master Plan (IMP) 68 PGCS 2018 Master Workshop, Canberra Australia
  • 69. Quick View of Building the IMP  Start with each Program Event and define the Significant Accomplishments their entry and exit criteria to assess the needed maturity of the key deliverables  Arrange the Significant Accomplishments in the proper dependency order  Segregate these Significant Accomplishments into swim lanes for IPTs  Define the dependencies between each SA 69 PGCS 2018 Master Workshop, Canberra Australia IMP
  • 70. Integrating the Parts of a Credible Program Management System (PMS) Risk SOW Cost WBS IMP/IMS TPM PMB ❺ Deliverables defined in the SOW, traced to the WBS, with narratives and Measures of Performance (MoP) ❹ Budget at the Work Package level, rolled to the Control Accounts showing cost spreads for all work in the IMS ❻ Measures of Performance (MoP) for each critical deliverable in the WBS and identified in each Work Package in the IMS, used to assess maturity in the IMP ❶ The Products and Processes in a “well structured” decomposition, traceable to the deliverables ❷ Schedule contains all the Work Packages, Budget, Risk mitigation plans, with traces to the Plan measuring increasing maturity through Measures of Effectiveness (MoE) and KPPs (JROC and Program) ❸ Technical and Programmatic Risks Connected through the WBS, Risk Register, Plan and Schedule The Baseline is the Document of Record for the Program Performance is Measured through the PMB 70 PGCS 2018 Master Workshop, Canberra Australia IMP
  • 71. The Critical Purpose of the IMP 71 The IMP defines the connections between the Product maturity – Vertical – and the implementation of this Product maturity through the Functional activities – the Horizontal PGCS 2018 Master Workshop, Canberra Australia IMP
  • 72. Next Step is the Build an Integrated Master Plan  From the WBS define the Significant Accomplishments (SA) and Accomplishment Criteria (AC) for each Program Event (PE) for each of the terminal nodes in the WBS 72 What must be accomplished to complete the Program Event? This is where the WBS and the IMP are joined and are at the same time separate Both are needed for a credible PMB PGCS 2018 Master Workshop, Canberra Australia IMP
  • 73. 73 The IMP tells us where is the program going? The Plan describes where we are going, the various paths we can take to reach our destination, and the progress or performance assessment points along the way to assure we are on the right path. These assessment points measures the “maturity” of the product or service against the planned maturity. This is the only real measure of progress – not the passage of time or consumption of money. The Integrated Master Plan (IMP) Is A Strategy For The Successful Completion Of The Project PGCS 2018 Master Workshop, Canberra Australia
  • 74. 74 The IMP / IMS Structure IMS IMP Describes how program capabilities will be delivered and how these capabilities will be recognized as ready for delivery Supplemental Schedules (CAM Notebook) Work Packages and Tasks Criteria Accomplishment Events or Milestones PGCS 2018 Master Workshop, Canberra Australia
  • 75. 75  Vertical traceability AC  SA  PE  Horizontal traceability WP  WP  AC Program Events Define the maturity of a Capability at a point in time. Significant Accomplishments Represent requirements that enable Capabilities. Accomplishment Criteria Exit Criteria for the Work Packages that fulfill Requirements. Work Package Work Package Work Package Work Package Work Package Work Package Work package Vertical and Horizontal Traceability PGCS 2018 Master Workshop, Canberra Australia
  • 76. 76 The IMP’s role during Execution Program ExecutionPMB for IBRProposal SubmittalDRFP & RFP Performance Measurement Baseline Tasks (T) BOE % Complete Statement of Work Program Deliverables IMP Accomplishments (A) Criteria (C) EVMS Events (E) Budget Spreads by CA & WPCAIV Capabilities Based Requirements X BCWS = Probabilistic Risk Analysis = Time keeping and ODC = Technical Performance Measure BCWP ACWP Cost & Schedule Risk Model BCWS D ecreasing techni cal and programmatic ri sk usi ng R i sk Management Methods IMS Physical % Complete Continuity and consistency from DRFP through Program Execution WBS PGCS 2018 Master Workshop, Canberra Australia
  • 77. 77 The IMP speaks to Measures of Effectiveness (MoE) and Measures of Performance (MoP)  This is where TPMs are connected with the MoE’s and MoP’s  For each deliverable from the program, all the “measures” must be defined in units meaningful to the decision makers.  Here’s some “real” examples. 1. Provide Precision Approach for a 200 FT/0.5 NM DH 2. Provide bearing and range to AC platform 3. Provide AC surveillance to GRND platform Measures of Effectiveness (MoE) 1. Net Ready 2. Guidance Quality 3. Land Interoperability 4. Manpower 5. Availability JROC Key Performance Parameters (KPP) 1. Net Ready  IPv4/6 compliance  1Gb Ethernet 2. Guidance quality  Accuracy threshold p70 @ 6M  Integrity threshold 4M @ 10-6 /approach 3. Land interoperability  Processing capability meets LB growth matrix 4. Manpower  MTBC >1000 hrs  MCM < 2 hrs 5. Availability  Clear threshold >99%  Jam threshold >90% Measures of Performance (MoP) 1. Net Ready  Standard message packets 2. Guidance Quality  Multipath allocation budget  Multipath bias protection 3. Land Interoperability  MOSA compliant  Civil compliant 4. Manpower  Operating elapsed time meters  Standby elapsed time indicators 5. Availability  Phase center variations Technical Performance Measures (TPM) Mission Capabilities and Operational Need Technical Insight – Risk adjusted performance to plan PGCS 2018 Master Workshop, Canberra Australia
  • 78. 78  Program Event (PE) – A PE assess the readiness or completion as a measure of progress – First Flight Complete  Significant Accomplishment (SA) – The desired result(s) prior to or at completion of an event demonstrate the level of the program’s progress – Flight Test Readiness Review Complete  Accomplishment Criteria (AC) – Definitive evidence (measures or indicators) that verify a specific accomplishment has been completed – SEEK EAGLE Flight Clearance Obtained F-22 Example PGCS 2018 Master Workshop, Canberra Australia
  • 79. 79  Start with the Significant Accomplishments and sequence them to the maturity flow for each Program Event  The WBS connections then become orthogonal to this flow 79 Program Event SRR SDR PDR CDR TRR ATLO Work Breakdown Structure 4.920-SDAI A01, A02 B01 C01, C02 D01 E01 F01 4.200-Sys Test A05 B03, B04 D02, D03 E02 F02 4.300-Radar A03 B02 C03 E03 4.330-O&C Sys A06, A07 B05 C04 D04 E04 F03, F04 4.400- I&T A08 C05 E05, E06 F05 4.500-Support A09 D05 E07 F06, F07 The IMP’s connection to the WBS PGCS 2018 Master Workshop, Canberra Australia
  • 80. The Primary Role of the IMP is to Describe what Done Looks Like in MoE’s and MoP’s 80 19 October 1899 Robert Goddard decided that he wanted to "fly without wings" to Moon. PGCS 2018 Master Workshop, Canberra Australia IMP
  • 81. The Integrated Master Schedule (IMS) is derived directly from the Integrated Master Plan’s Accomplishment Criteria (AC). The IMS shows the order in which the Work Packages must be performed to assure the Accomplishment Criteria are completed within the define Measures of Performance, the Key Performance Parameters, and the Technical Performance Measures. IMS Developing the Integrated Master Schedule from the Integrated Master Plan 81 PGCS 2018 Master Workshop, Canberra Australia
  • 82. Quick View of Building the IMS  The WBS is Paramount  The IMP defines the increasing maturity of the program’s deliverables (end item)  The IMS sequences the Work Packages containing the work activities to produce the End Item Deliverables  The IMS is built from the IMP, with WBS coding to assure coverage of all deliverables 82 PGCS 2018 Master Workshop, Canberra Australia IMS
  • 83. The Integrated Master Schedule  The horizontal sequence of work activities that produce increasing maturity of the product or services delivered by the program 83 PGCS 2018 Master Workshop, Canberra Australia IMS
  • 84. A Credible Integrated Master Plan Must … 84  Show what Done looks like through tangible evidence of success  Show the order of the work needed to get to Done at each stage  Define the needed resources to reach Done  Identify risks to Done and their handling  Measure physical progress toward Done in units meaningful to the decision makers PGCS 2018 Master Workshop, Canberra Australia IMS
  • 85. Building the Integrated Master Schedule requires 10 Steps 85 Capture All Activities1 2 3 4 5 6 7 8 9 10 Sequence These Activities Assign Resources To These Activities Establish Duration For These Activities Verify Schedule Is Traceable Horizontally And Vertically Confirm Valid Critical Path – schedule matches program Ensure Reasonable Total Float Conduct Schedule Risk Analysis Update Schedule With Actual Progress Maintain Baseline with Repeatable Process PGCS 2018 Master Workshop, Canberra Australia IMS
  • 86. 86 Mission Need Mission Management Control 1st Level RTOS integration 1st Level Enroute planning 2nd Level Routing upload buffer 2nd Level Planning verification 2nd Level Navigation sensor interface 2nd Level 1553 data bus drivers Deliverables defined in WP Back to the WBS It connects Work Packages in IMS to IMP PGCS 2018 Master Workshop, Canberra Australia
  • 87. The IMS Provides Visibility to … 87 Work Packages Deliverables Technical Capabilities Mission Requirements § Deliverables represent the required mission capabilities and their value as defined by the mission and shared by the development team. § When all deliverables and their Work Packages are completed, they are not revisited or reopened. § The progression of Work Packages defines the increasing maturity of the project. § Completion of Work Packages is represented by the Physical Percent Completion of the program. PGCS 2018 Master Workshop, Canberra Australia IMS
  • 88. Quantifiable Backup Data (QBD)  QBD is a detailed listing of tasks necessary to complete all scope in a work package during the defined period of performance.  It is an approach to objectively measure performance  Each task on the list is weighted – total weighting equals 100% of the work package’s Budget at Completion. This weighting should not be equal weighting on every task.  The CAM assesses physical percent complete of each QBD task.  The percent complete is calculated from the cumulative assessments.  The purpose of the QBD is to help:  Ensure and demonstrates that all contract work is accounted for  Ensure the schedule and budget are realistic and achievable  Mitigate schedule and budget risks  Provide a basis for objectively assessing progress for discretely measured work packages. 88 PGCS 2018 Master Workshop, Canberra Australia IMS
  • 89. Naturally occurring uncertainties (Aleatory) in cost, schedule, and technical performance can be modeled in a Monte Carlo Simulation tool. Event Based uncertainties (Epistemic) require capturing and defining the probability of their occurrence, modeling their impacts, defining handling strategies, modeling the effectiveness of these handling strategies, and modeling the residual risks, and the impacts of both the original risk and the residual risk on the program. The management of Uncertainties in cost, schedule, and technical performance; and the Event Based uncertainty and the resulting risk are both critical success factors for the programs. Risk Risk Management is Project Management for Adults ‒ Tim Lister 89 PGCS 2018 Master Workshop, Canberra Australia
  • 90. Core Elements of Program Risk Management  The effectiveness of risk management depends on the people who set it up and coordinate the risk management process  On many program risk management consists only of having a policy and oversight  If we treat red flags as false alarms rather than early warnings of danger this incubates the threats to program success  Group think of dominate leaders often inhibits good thinking about risks 90 “Towards a Contingency Theory of Enterprise Risk Management,” Anette Mikes and Robert Kaplan, Working Paper 13-063 January 13, 2014PGCS 2018 Master Workshop, Canberra Australia Risk
  • 91. 91 91 PGCS 2018 Master Workshop, Canberra Australia
  • 92. Some Words About Uncertainty  When we say uncertainty, we speak about a future state of a system that is not fixed or determined  Uncertainty is related to three aspects of our program management domain:  The external world – the activities of the program  Our knowledge of this world – the planned and actual behaviors of the program  Our perception of this world – the data and information we receive about these behaviors 92 PGCS 2018 Master Workshop, Canberra Australia Risk
  • 93. Some Words About the Risk that Results from Uncertainty  Risk has Two Dimensions  The degree of possibility that an state or condition will take place or occur sometime in the future  The consequences of that state or condition, once it has occurred  The degree of possibility is qualified as the  Probability of Occurrence (event based)  Probability Distribution Function (a distribution of the variability of a random number)  The consequences are usually taken to be undesirable and qualified as the magnitude of harm and the remaining probability of a recurrence of the same risk. 93 PGCS 2018 Master Workshop, Canberra Australia Risk
  • 94. All Program Activities have Naturally Occurring (Aleatory) Uncertainties 94  Naturally occurring uncertainty and its resulting risk, impacts the probability of a successful outcome.  The irreducible statistical behavior of these activities, their arrangement in a network of activities, and correlation between their behaviors creates risk.  Adding margin protects the outcome from the impact of this naturally occurring uncertainty  The question is ‒ given the statistical nature of the Irreducible Uncertainty, what’s the Probability we will be late, over budget, or the technical outcome won’t work as needed?PGCS 2018 Master Workshop, Canberra Australia Risk
  • 95. The Relationship Between Uncertainty and Risk  Uncertainty is present when probabilities cannot be quantified in a rigorous or valid manner, but can described as intervals within a probability distribution function (PDF)  Risk is present when the Uncertainty of an outcome can be quantified in terms of  Probability of Occurrence (Epistemic uncertainty)  A range of possible values in a Probability Distribution Function (Aleatory Uncertainty)  This distinction is important for modeling the future performance of cost, schedule, and technical outcomes of a program.  Risk from Epistemic uncertainty is reducible  Risk from Aleatory uncertainty is irreducible 95 PGCS 2018 Master Workshop, Canberra Australia Risk
  • 96. Taxonomy of Uncertainty and the Risk it Creates 96 PGCS 2018 Master Workshop, Canberra Australia Risk
  • 97. Remember ‒ WBS is Paramount ? Risk Propagates Through the WBS 97 PGCS 2018 Master Workshop, Canberra Australia Risk
  • 98. Connecting Epistemic Risk Retirement in the IMS  The work to “Buy down” risk is planned in the IMS.  MoE, MoP, and KPP defined in the Work Packages for the critical measure, e.g. weight.  If we can’t verify we’ve succeeded, then the risk did not get reduced.  The risk may have gotten worse. 98 Risk: CEV-037 - Loss of Critical Functions During Descent Planned Risk Level Planned (Solid=Linked, Hollow =Unlinked, Filled=Complete)RiskScore 24 22 20 18 16 14 12 10 8 6 4 2 0 Conduct Force and Moment Wind Develop analytical model to de Conduct focus splinter review Conduct Block 1 w ind tunnel te Correlate the analytical model Conduct w ind tunnel testing of Conduct w ind tunnel testing of Flight Application of Spacecra CEV block 5 w ind tunnel testin In-Flight development tests of Damaged TPS flight test 31.Mar.05 5.Oct.05 3.Apr.06 3.Jul.06 15.Sep.06 1.Jun.07 1.Apr.08 1.Aug.08 1.Apr.09 1.Jan.10 16.Dec.10 1.Jul.11 PGCS 2018 Master Workshop, Canberra Australia Risk
  • 99. 99 The Final Notion of Risk PGCS 2018 Master Workshop, Canberra Australia
  • 100. Root Causes of Risk Some are Reducible some are Irreducible 100 Unrealistic Performance Expectations missing Measures of Effectiveness (MOE), Performance (MOP), and Technical Performance Measures (TPM) Unrealistic Cost and Schedule Estimates based on inadequate risk adjusted growth models Inadequate assessment of risk and unmitigated exposure to these risks without proper handling plans Unanticipated Technical Issues without alternative plans and solutions to maintain effectiveness Cost, Schedule, and Technical Impacts TheLensofthePerformance Indicators “Borrowed” with permission from Mr. Gary Bliss, Director, Performance Assessments and Root Cause Analyses (PARCA), Office of Assistant Secretary of Defense for Acquisition. PGCS 2018 Master Workshop, Canberra Australia Risk
  • 101. 101  Lack of predictive variance analysis  Untimely and unrealistic Latest Revised Estimates (LRE)  Progress not monitored in a regular and consistent manner  Lack of vertical and horizontal traceability cost and schedule data for corrective action  Lack of internal surveillance and controls  Managerial actions not demonstrated using Earned Value  Inattention to budgetary responsibilities  Work authorizations that are not always followed  Issues with Budget and data reconciliation  Lack of an integrated management system  Baseline fluctuations and frequent replanning  Current period and retroactive changes  Improper use of management reserve  EV techniques that do not reflect actual performance Without Removing these items, Train Wreck of the Program Starts When there is … Mary K. Evans Picture Library PGCS 2018 Master Workshop, Canberra Australia
  • 102. Beware the Black Swan 102 PGCS 2018 Master Workshop, Canberra Australia
  • 103. Using Integrated Master Plan and Risk Adjusted Integrated Master Schedule, showing the needed progress to that plan is the basis of measuring physical percent complete. These measures start with Technical Performance Measures (TPM) Progress to Measuring Progress to Plan means Measuring Physical Percent Complete in meaningful units of Measure. These are Technical Performance Measures (TPM) This Does NOT Start with EV or ES 103 PGCS 2018 Master Workshop, Canberra Australia
  • 104. TPM’s Inform EV and ES to show progress to our destination  How do we increase visibility into the program’s performance?  How do we reduce cycle time to deliver the product?  How do we foster accountability?  How do we reduce risk?  How do we start this journey to success? 104 PGCS 2018 Master Workshop, Canberra Australia Progressof Plan
  • 105. To Achieve Success … 105 ©gapingvoid ltd www.gapingvoidgallery.com We Need to … PGCS 2018 Master Workshop, Canberra Australia Progressof Plan
  • 106. Here’s where TPMs Start 106 PGCS 2018 Master Workshop, Canberra Australia Progressof Plan
  • 107. Guidance for MoE’s, MoP, and TPMs Belongs to Systems Engineering The starting point is not EVM, it’s Systems Engineering  MOE’s are an essential part of Systems Engineering, guided by IEEE 1220 and EIA 632.  System’s Engineers drive the content of all measurement items, customer or supplier 107 PGCS 2018 Master Workshop, Canberra Australia Progressof Plan
  • 108. Previous Approaches Using EV are Mostly Unsuccessful Connecting these Measures  Traditional approaches to program management are retrospective  Cost and schedule of Earned Value  Risk Management  Systems Engineering  Reporting past performance  Sometimes 30 to 60 days old  Variances are reporting beyond the widow of opportunity for correction 108 PGCS 2018 Master Workshop, Canberra Australia Progressof Plan
  • 109. TPMs have been around for 34 years Why Is This Hard To Understand?  We seem to be focused on EV reporting, not the use of EV to manage the program.  Getting the CPR out the door is the end of Program Planning and Control’s efforts, not the beginning. 109 … the basic tenets of the process are the need for seamless management tools, that support an integrated approach … and “proactive identification and management of risk” for critical cost, schedule, and technical performance parameters. ― Secretary of Defense, Perry memo, May 1995 TPM Handbook 1984 PGCS 2018 Master Workshop, Canberra Australia Progressof Plan
  • 110. Technical Performance Measures do what they say, Measure the Technical Performance of the product or service produced by the program Back to Technical Performance Measures 110 PGCS 2018 Master Workshop, Canberra Australia
  • 111. What’s Our Motivation for “Connecting the Dots?” PGCS 2018 Master Workshop, Canberra Australia Technical Performance Measures …  Provide program management with information to make better decisions  Increase the probability of delivering a solution that meets both the requirements and mission need 111 TPM
  • 112. Measure of Effectiveness (MoE) PGCS 2018 Master Workshop, Canberra Australia Measures of Effectiveness …  Are stated by the buyer in units meaningful to the buyer  Focus on capabilities independent of any technical implementation “Technical Measurement,” INCOSE–TP–2003–020–01 112 TPM
  • 113. Measure of Performance (MoP) PGCS 2018 Master Workshop, Canberra Australia Measures of Performance are …  Attributes that assure the system has the capability to perform  Assessment of system to assure it meets design requirements necessary to satisfy the MOE “Technical Measurement,” INCOSE–TP–2003–020–01 113 TPM
  • 114. Key Performance Parameters (KPP) PGCS 2018 Master Workshop, Canberra Australia Key Performance Parameters …  have a threshold or objective value  Characterize the major drivers of performance  Are considered Critical to Customer (CTC) “Technical Measurement,” INCOSE–TP–2003–020–01 114 TPM
  • 115. Technical Performance Measures (TPM) PGCS 2018 Master Workshop, Canberra Australia Technical Performance Measures …  Assess design progress  Define compliance to performance requirements  Identify technical risk  Are limited to critical thresholds  Include projected performance “Technical Measurement,” INCOSE–TP–2003–020–01 115 TPM
  • 116. Dependencies Between Measures PGCS 2018 Master Workshop, Canberra Australia “Coming to Grips with Measures of Effectiveness,” N. Sproles, Systems Engineering, Volume 3, Number 1, pp. 50–58 116 TPM
  • 117. “Candidates” for Technical Measures PGCS 2018 Master Workshop, Canberra Australia INCOSE Systems Engineering Handbook Concept Description Physical Size and Stability Useful Life Weight Volumetric capacity Functional Correctness Accuracy Power performance All the “ilities” Supportability Maintainability Dependability Reliability = Mean Time Failure Efficiency Utilization Response time Throughput Suitability for Purpose Readiness 117 TPM
  • 118. “Measures” of Technical Measures PGCS 2018 Master Workshop, Canberra Australia INCOSE Systems Engineering Handbook Attribute Description Achieved to Date Measured technical progress or estimate of progress Current Estimate Value of a technical parameter that is predicted to be achieved Milestone Point in time when an evaluation of a measure is accomplished Planned Value Predicted value of the technical parameter Planned Performance Profile Profile representing the project time phased demonstration of a technical parameter Tolerance Band Management alert limits Threshold Limiting acceptable value of a technical parameter Variances  Demonstrated technical variance  Predicted technical variance 118 TPM
  • 119. A Simple Method of Assembling the TPMs PGCS 2018 Master Workshop, Canberra Australia 119 TPM
  • 120. TPMs from James Webb Space Telescope PGCS 2018 Master Workshop, Canberra Australia 120 TPM
  • 121. TPMs from Chandra X–Ray Telescope PGCS 2018 Master Workshop, Canberra Australia 121 TPM
  • 122. TPMs Start With The WBS 122 1.1.2 Airframe PGCS 2018 Master Workshop, Canberra Australia TPM
  • 123. What Do We Need To Know About This Program Through TPMs  What WBS elements represent the TPMs?  What Work Packages produce these WBS elements?  Where do these Work Packages live in the IMS?  What are the Earned Value baseline values for these Work Packages?  How are going to measure all these variables?  What does the curve look like for these measurements? 123 PGCS 2018 Master Workshop, Canberra Australia TPM
  • 124. Verifying Each TPM at Each Stage in the Program Question Answered by the TPM Evidence that we’re meeting the TPM Contract Award Do we know what we promised to deliver, now that we’ve won? With our submitted ROM what are the values we need to get through Integrated Baseline Review «how do we measure weight for each program event» System Functional Requiremen ts Can we proceed into preliminary design? The contributors to the vehicle weight are confirmed and the upper limits defined in the product architecture and requirements flow down database (DOORS) into a model System Requiremen ts Review Can we proceed into the System Development and Demonstration (SDD) phase Do we know all drivers of vehicle weight? Can we bound their upper limits? Can the subsystem owners be successful within these constraints uses a high fidelity model? Preliminary Design Review Can we start detailed design, and meet the stated performance requirements within cost, schedule, risk, and other constraints? Does each subsystem designer have the target component weight target and have some confidence they can stay below the upper bound? Can this be verified in some tangible way? Either through prior examples or a lab model? Critical Design Review Can the system proceed to fabrication, demonstration, and test, with the within cost, schedule, risk, and other system constraints. Do we know all we need to know to start the fabrication of the first articles of the flight vehicle. Some type of example, maybe a prototype is used to verify we’re inside the lines Test Does the assembled vehicle fall within the weight range limits for 124 PGCS 2018 Master Workshop, Canberra Australia TPM
  • 125. 25kg 23kg 28kg TPM Trends & Responses Dr. Falk Chart – modified 26kg PDRSRRSFRCA TRRCDR ROM in Proposal Design Model Bench Scale Model Measurement Detailed Design Model Prototype Measurement Flight 1st Article TechnicalPerformanceMeasure VehicleWeight 125 PGCS 2018 Master Workshop, Canberra Australia
  • 126. The Assessment Of Weight As A Function Of Time  At Contract Award there is a Proposal grade estimate of vehicle weight  At System Functional Review, the Concept of Operations is validated for the weight  At System Requirements Review the weight targets are flowed down to the subsystems components  At PDR the CAD model starts the verification process  At CDR actual measurements are needed to verify all models  At Test Readiness Review we need to know 126 PGCS 2018 Master Workshop, Canberra Australia TPM
  • 127. Raison d'etre for Technical Performance Measures  The real purpose of Technical Performance Measures is to reveal Programmatic and Technical RISK and engage in the Risk Management Process PGCS 2018 Master Workshop, Canberra Australia Risk SOW Cost WBS IMP/IMS TPM PMB 127 TPM
  • 128. Buying Down Risk with TPMs  “Buying down” risk is planned in the IMS.  MoE, MoP, and KPP defined in the work package for the critical measure – weight.  If we can’t verify we’ve succeeded, then the risk did not get reduced.  The risk may have gotten worse. Risk: CEV-037 - Loss of Critical Functions During Descent Planned Risk Level Planned (Solid=Linked, Hollow =Unlinked, Filled=Complete) 24 22 20 18 16 14 12 10 8 6 4 2 0 Conduct Force and Moment Wind Develop analytical model to de Conduct focus splinter review Conduct Block 1 w ind tunnel te Correlate the analytical model Conduct w ind tunnel testing of Conduct w ind tunnel testing of Flight Application of Spacecra CEV block 5 w ind tunnel testin In-Flight development tests of Damaged TPS flight test 31.Mar.05 5.Oct.05 3.Apr.06 3.Jul.06 15.Sep.06 1.Jun.07 1.Apr.08 1.Aug.08 1.Apr.09 1.Jan.10 16.Dec.10 1.Jul.11 Weight risk reduced from RED to Yellow Weight confirmed ready to fly – it’s GREEN at this point 128 PGCS 2018 Master Workshop, Canberra Australia TPM
  • 129. Increasing Probability of Success Requires Risk Management  Going outside the TPM limits always means cost and schedule impacts  “Coloring Inside the Lines” means knowing the how to keep the program GREEN, or at least stay close to GREEN PGCS 2018 Master Workshop, Canberra Australia 129 TPM
  • 130. Technical Performance Measures Checklist 130 MoE MoP TPM Traceable to needs, goals, objectives, and risks Traceable to applicable MOEs, KPPs, system level performance requirements, and risks Traceable to applicable MoPs, system element performance, requirements, objectives, risks, and WBS elements Defined with associated KPPs Focused on technical risks and supports trades between alternative solutions Further decomposed, budgeted, and allocated to lower level system elements in the WBS and IMS Each MoE independent from others Provided insight into system performance Assigned an owner, the CAM and Work Package Manager Each MoE independent of technical any solution Decomposed, budgeted and allocated to system elements Sources of measure identified and processes for generating the measures defined. Address the required KPPs Assigned an “owner,” the CAM and Technical Manager Integrated into the program’s IMS as part of the exit criteria for the Work Package PGCS 2018 Master Workshop, Canberra Australia TPM
  • 131. All the program performance data in the is historical. This past performance data – by itself – is like driving in the rear view mirror. What is needed is Leading Indicators that can be derived from this past performance data. Integration Creating an Integrated Program Performance Management System (IPPMS) starts with the Five Principles, their Processes, and the Practices 131 PGCS 2018 Master Workshop, Canberra Australia
  • 132. 132 Integrated Program Management System (IPMS) SOW Risks Metrics  Schedule  Cost  Performance CDRLs / DRDs WBS OBS Program Events Significant Accomplishments Accomplishment Criteria Work Packages and Tasks Weekly project management reviews % Cmplt for PE/SA/AC Monthly EVMS Analysis by WBS CA WP Management review and analysis plan and updates Weekly Project Status Management Review  Critical Path  Performance analysis  Metrics  Risk analysis Monthly EVMS Metrics  CPI  SPI  EAC Delineated Success Measures Critical efforts to mature the product deliverables Measures of product maturity Work efforts to mature deliverables Program Documents IMP / IMS Development Analysis Performance Reporting Tasks Product maturity and traceability analysis PGCS 2018 Master Workshop, Canberra Australia
  • 133. 133 Ashton Carter, Secretary of Defense PGCS 2018 Master Workshop, Canberra Australia
  • 134. 134 Primary Elements of Earned Value Cost Technical Performance Schedule Funding margin for under performance Schedule margin for over target baseline (OTB) Schedule margin for underperformance or schedule extension Over cost or under performance Over cost or over schedule Over schedule or under performing PGCS 2018 Master Workshop, Canberra Australia
  • 135. 135 Connecting the EVM Variables with Technical Performance Measures Integrating Cost, Schedulele, and Technical Performance Assures Program Management has the needed performance information to deliver on‒time, on‒budget, and on‒specification Technical Performance Measures Cost Schedule Conventional Earned Value + =  Master Schedule is used to derive Basis of Estimate (BOE) not the other way around.  Probabilistic cost estimating uses past performance and cost risk modeling.  Labor, Materiel, and other direct costs accounted for in Work Packages.  Risk adjustments for all elements of cost. Cost Baseline  Earned Value is diluted by missing technical performance.  Earned Value is diluted by postponed features.  Earned Value is diluted by non compliant quality.  All these dilutions require adjustments to the Estimate at Complete (EAC) and the To Complete Performance Index (TCPI). Technical Performance  Requirements are decomposed into physical deliverables.  Deliverables are produced through Work Packages.  Work Packages are assigned to accountable manager.  Work Packages are sequenced to form the highest value stream with the lowest technical and programmatic risk. Schedule Baseline PGCS 2018 Master Workshop, Canberra Australia
  • 136. Connecting EVM and Technical Performance Measures, we get the one and only way to measure progress with EV  This is all that is needed to be successful with EVM  Measure what has be completed compared what was planned to be completed in units meaningful the decision makers  That’s Physical Percent Complete  Defining Physical Percent Complete starts and ends with  MoE’s  MoP’  TPM’s  Key Performance Parameters 136 PGCS 2018 Master Workshop, Canberra Australia Integration
  • 137. Earned Value and Earned Schedule can provide Answer these four Critical Questions137 PGCS 2018 Master Workshop, Canberra Australia Integration
  • 138. 138 PGCS 2018 Master Workshop, Canberra Australia
  • 139. Bibliography 139 PGCS 2018 Master Workshop, Canberra Australia "Knowledge is of two kinds. We know a subject ourselves, or we know where we can find information upon it. When we enquire into any subject, the first thing we have to do is to know what books have treated of it. This leads us to look at catalogues, and at the backs of books in libraries.“ — Samuel Johnson (Boswell's Life of Johnson)
  • 140. Bibliography Principles and Processes  Systems Engineering Handbook: A Guide for System Life Cycle Processes and Activities, INCOSE–TP–2003–002–03.1, August 2007, www.incose.org  Systems Engineering: Coping with Complexity, Richard Stevens, Peter Brook, Ken Jackson, and Stuart Arnold, Prentice Hall, 1998.  Henri Fayol: Critical Evaluations in Business and Management, John C. Wood and Michael C. Wood, Routedge, 2002.  “Technical Performance Measurement, Earned Value, and Risk Management: An Integrated Diagnostic Tool for Program Management,” Commander N. D. Pisano, SC, USN, Program Executive Office for Air ASW, Assault, and Special Mission Programs (PEO(A)) 140 PGCS 2018 Master Workshop, Canberra Australia
  • 141. Bibliography Practices  Modeling Homeland Security: A Value Focused Thinking Approach, Kristopher Pruitt, Air Force Institute of Technology, March 2003  Performance Based Earned Value, Paul Solomon.  Guide to the Project Management Body of Knowledge, Project Management Institute.  “No Silver Bullet: Essence and Accidents of Software Engineering, Fred Brooks, IEEE Computer, 10‒19, April 1987.  Systems Requirements Analysis, Jeffry O. Grady, Academic Press, 2006.  “Further Development in Earned Schedule,” Kym Henderson, The Measurable News, Spring 2004.  “Schedule is Different,” Walter Lipke, The Measurable News, Summer 2003.  “A Simulation and Evaluation of Earned Value Metrics to Forecast Project Duration,” M. S. Vanhoucke, Journal of Operations Research Society, October 2007.  Standard for Application and Management of the Systems Engineering Process, Institute of Electrical and Electronics Engineers, 09–Sep–2005  Systems Engineering: Coping with Complexity, Richard Stevens, Peter Brook, Ken Jackson, and Stuart Arnold, Prentice Hall, 1998.  The Requirements Engineering Handbook, Ralph R. Young, Artech House, 2004 141 PGCS 2018 Master Workshop, Canberra Australia
  • 142. Bibliography Practices  “Technical Performance Measurement, Earned Value, and Risk Management: An Integrated Diagnostic Tool for Program Management,” Commander N. D. Pisano, SC, USN, Program Executive Office for Air ASW, Assault, and Special Mission Programs (PEO(A))  “Issues with Requirements Elicitation,” Michael G. Christel and Kyo C. Kang, Technical Report, CMU/SEI–92–TR–12, Software Engineering Institute, Carnegie Mellon University Pittsburgh, Pennsylvania 15213.  “Managing complex product development projects with design structure matrices and domain mapping matrices,” Mike Danilovic and Tyson Browning, International Journal of Project Management, 25 (2007), pp. 300–314.  “Architectural optimization using real options theory and Dependency structure matrices,” David M. Sharman, Ali A. Yassine+, Paul Carlile, Proceedings of DETC ’02 ASME 2002 International Design Engineering Technical Conferences 28th Design Automation Conference Montreal, Canada, September 29–October 2, 2002.  “Modeling and Analyzing Cost, Schedule, and Performance in Complex System Product Development,” Tyson Browning, Massachusetts Institute of Technology, February 1999.  The Integrated Project Management Handbook, Dayton Aerospace, 8 Feb 2002, Dayton Ohio.  “Analytic Architecture for Capabilities–Based Planning, Mission–System Analysis, and Transformation,” Paul K. Davis, RAND Corporation.  Systems Engineering Handbook: A Guide for System Life Cycle Processes and Activities, INCOSE–TP–2003–002–03.1, August 2007, www.incose.org 142 PGCS 2018 Master Workshop, Canberra Australia
  • 143. Bibliography Capabilities  “Analytic Architecture for Capabilities–Based Planning, Mission–System Analysis, and Transformation,” Paul K. Davis, RAND Corporation.  Portfolio–Analysis Methods for Assessing Capability Options, Paul K. Davis, Russell D. Shaver, and Justin Beck, Rand Corporation, 2012  “Architectural optimization using real options theory and Dependency structure matrices,” David M. Sharman, Ali A. Yassine+, Paul Carlile, Proceedings of DETC ’02 ASME 2002 International Design Engineering Technical Conferences 28th Design Automation Conference Montreal, Canada, September 29–October 2, 2002.  “Modeling and Analyzing Cost, Schedule, and Performance in Complex System Product Development,” Tyson Browning, Massachusetts Institute of Technology, February 1999.  The Art of Systems Architecting, Mark W. Maier and Eberhardt Rechtin, CRC Press, 2000.  Systems Engineering: Coping With Complexity, Richard Stevens, Peter Brook, Ken Jackson, and Stuart Arnold, Prentice Hall, 1998.  Assumption Based Planning: A Tool for Reducing Avoidable Surprises, James A. Dewar, Cambridge University Press, 2002.  Capabilities–Based Planning: A Methodology for Deciphering Commander’s Intent, Peter Kossakowski, Evidence Based Research, Inc. 1595 Spring Hill Road, Suite 250 Vienna, VA 22182.  Competing on Capabilities: The New Rules for Corporate Strategy, George Stalk, Philip Evans, and Lawrence Shulman, Harvard Business Review, No. 92209, March–April 1992.  Effects–Driven, Capabilities–Based, Planning for Operations, Maj Kira Jeffery, USAF and Mr Robert Herslow.  Effects–based Operations: Building Analytical Tools, Desmon Saunder–Newton and Aaron B. Frank, Defense Horizons, October 2002, pp 1–8.  Guide to Capability–Based Planning, Joint Systems and Analysis Group, MORS Workshop, October 2004, Alexandria, VA. http://www.mors.org/ 143 PGCS 2018 Master Workshop, Canberra Australia
  • 144. Bibliography Requirements  “Issues with Requirements Elicitation,” Michael G. Christel and Kyo C. Kang, Technical Report, CMU/SEI–92–TR–12, Software Engineering Institute, Carnegie Mellon University Pittsburgh, Pennsylvania 15213.  The Requirements Engineering Handbook, Ralph R. Young, ArcTech House, 2004  Software Risk Management, Barry W. Boehm, IEEE Computer Society Press, 1989.  Software Requirements Analysis & Specifications, Alan M. Davis, Prentice Hall, 1990.  Requirements Engineering: A Good Practice Guide, Ian Sommerville and Pete Sawyer, John Wiley & Sons, 1997.  Systems Requirements Practices, Jeffery O. Grady, McGraw Hill, 1993  Four Key Requirements Engineering Techniques, Christof Ebert, IEEE Software, May / June 2006.  Intent Specifications: An Approach to Building Human–Centered Specifications, Aeronautics and Astronautics, MIT.  Sample TCAS Intent Specification, Nancy Leveson and Jon Damon Reese, Software Engineering Corporation . 144 PGCS 2018 Master Workshop, Canberra Australia
  • 145. Bibliography Performance Measurement Baseline  “Managing complex product development projects with design structure matrices and domain mapping matrices,” Mike Danilovic and Tyson Browning, International Journal of Project Management, 25 (2007), pp. 300–314.  MIL–STD–881A, Work Breakdown Structures.  The Management of Projects, Peter W. G. Morris, Thomas Telford, 1994  Modelling Complex Projects, Terry Williams, John Wiley & Sons. 2002.  The Handbook of Program Management, James T. Brown, McGraw Hill, 2007.  AntiPatterns in Project Management, William J. Brown, Hays W. McCormick III, and Scott W. Thomas, John Wiley & Sons, 2000  Earned Value Management, 3rd Edition, Quentin W. Fleming and Joel M. Koppelman, Project Management Institute, 2005. 145 PGCS 2018 Master Workshop, Canberra Australia
  • 146. Bibliography Continuous Risk Management  “Understanding Risk Management in the DoD,” Mike Bolles, Acquisition Research Journal, Volume 10, pp. 141–145, 2003.  Effective Risk Management: Some Keys to Success, 2nd Edition, Edmund H. Conrow, AIAA Press, 2003.  Managing Risk: methods for Software Systems Development, Elaine M. Hall, Addison Wesley, 1998  Integrating Risk Management with Earned Value Management, National Defense Industry Association.  Three point estimates and quantitative risk analysis a process guide for risk practitioners, Acquisitioning Operating Framework, UK Ministry of Defense, http://www.aof.mod.uk/index.htm  [Effective Opportunity Management for Project: Exploring Positive Risk, David Hillson, Taylor & Francis, 2004.  [Catastrophe Disentanglement: Getting Software Projects Back on Track, E. M. Bennatan, Addison Wesley, 2006.  Software Engineering Risk Management, Dale Walter Karolak, IEEE Computer Society Press, 1998.  Assessment and Control of Software Risks, Capers Jones, Prentice Hall, 1994  Three Point Estimates and Quantitative Risk Analysis – A Process Guide For Risk Practitioners – version 1.2 May 2007 – Risk Management – AOF, http://www.aof.mod.uk/aofcontent/tactical/risk/downloads/3pepracgude.pdf  How much risk is too much risk?, Tim Lister, Boston SPIN, January 20th, 2004.  Risk Management Maturity Level Development, INCOSE Risk Management Working Group, April 2002.  “A Methodology for Project Risk Analysis using Bayesian Belief Networks within a Monte Carlo Simulation Environment,” Javier F. Ordóñez Arízaga, University of Maryland, College Park, 2007.  An Approach to Technology Risk Management, Ricardo Valerdi and Ron J. Kohl, Engineering Systems Division Symposium, MIT, Cambridge, MA 29–31 March 2004. 146 PGCS 2018 Master Workshop, Canberra Australia
  • 147. Bibliography Continuous Risk Management  “An Industry Standard Risk Analysis Technique,” A Terry Bahill and Eric D. Smith, Engineering Management Journal, Vol. 21 No 4., December 2009.  Risk Management Process and Implementation, American Systems Corporation, 2003.  The Basics of Monte Carlo Simulation: A Tutorial, S. Kandaswamy, Proceedings of the Project Management Institute Seminars & Symposium, 1–10 November 2001.  Bayesian Inference for NASA Probabilistic Risk and Reliability Analysis, NASA/SP–2009–569, June 2009.  Common Elements of Risk, Christopher J. Alberts, CMU/SEI–2006–TN–014, April 2006.  “Development of Risk Management Defense Extensions to the PMI Project Management Body of Knowledge,” Edmund H. Conrow, Acquisition Review Quarterly, Spring 2003.  Continuous Risk Management Guidebook, Audrey J. Dorofee, Julie A. Walker, Christopher J. Alberts, Ronald P. Higuera, Richard L. Murphy, and Ray C. Williams, Software Engineering Institute, 1996.  Risk Management Guide for DoD Acquisition, Fifth Edition, June 2002.  Emerging Practice: Joint Cost & Schedule Risk Analysis, Eric Drucker, 2009 St. Louis SCEA Chapter Fall Symposium, St’ Louis, MO.  Making Risk Management Tools More Credible: Calibrating the Risk Cube, Richard L. Coleman, Jessica R. Summerville, and Megan E. Dameron, SCEA 2006, Washington, D.C., 12 June 2006.  A Theory of Modeling Correlations for Use in Cost–Risk Analysis, Stephen A. Book, Third Annual Project Management Conference, NASA, Galveston, TX, 21–22 March 2006.  The Joint Confidence Level Paradox: A History of Denial, 2009 NASA Cost Symposium, 28 April 2009. 147 PGCS 2018 Master Workshop, Canberra Australia
  • 148. 148 PGCS 2018 Master Workshop, Canberra Australia

Hinweis der Redaktion

  1. PGCS 2018 Master Workshop, Canberra Australia
  2. PGCS 2018 Master Workshop, Canberra Australia
  3. PGCS 2018 Master Workshop, Canberra Australia
  4. PGCS 2018 Master Workshop, Canberra Australia
  5. Let’s start with a question that is being asked in every project management engagement on the planet. The answer to this question is simple. We want to start spending our customers' money. We want to start building something, that’s why we got hired ‒ to produce something tangible. But as Stephen Covey reminds us the 2nd habit of highly effective people is Begin with the End in Mind. The 1st habit is Be Proactive by taking responsibility for the success of the project. You can't keep blaming everything on the system, the tools, the bad requirements, the confusion of the project. Proactive project managers recognize that they are "response-able."  They start on the road to success by proactively develop a description of Done in Units of Measure Meaningful to the Decision Makers.
  6. PGCS 2018 Master Workshop, Canberra Australia
  7. PGCS 2018 Master Workshop, Canberra Australia
  8. PGCS 2018 Master Workshop, Canberra Australia
  9. PGCS 2018 Master Workshop, Canberra Australia
  10. We’ve all been here before. The question is how do we get out of this situation while increasing the probability of project success?
  11. PGCS 2018 Master Workshop, Canberra Australia
  12. 12
  13. PGCS 2018 Master Workshop, Canberra Australia
  14. PGCS 2018 Master Workshop, Canberra Australia
  15. PGCS 2018 Master Workshop, Canberra Australia
  16. PGCS 2018 Master Workshop, Canberra Australia
  17. PGCS 2018 Master Workshop, Canberra Australia
  18. PGCS 2018 Master Workshop, Canberra Australia
  19. PGCS 2018 Master Workshop, Canberra Australia
  20. PGCS 2018 Master Workshop, Canberra Australia
  21. PGCS 2018 Master Workshop, Canberra Australia
  22. PGCS 2018 Master Workshop, Canberra Australia
  23. PGCS 2018 Master Workshop, Canberra Australia
  24. PGCS 2018 Master Workshop, Canberra Australia
  25. PGCS 2018 Master Workshop, Canberra Australia
  26. Here’s a high level process flow for defining the needed system capabilities. This process flow comes from the United States Army War College paper written in 2005. It’s still applicable today but rarely used. Because it is rarely used we can attribute this missing process a root cause of many program failures because they failed to define what done looks like. This is a theme that will be repeated throughout this course and likely repeated throughout your experiences on defense and critical infrastructure programs.
  27. PGCS 2018 Master Workshop, Canberra Australia
  28. PGCS 2018 Master Workshop, Canberra Australia
  29. PGCS 2018 Master Workshop, Canberra Australia
  30. There are three baselines for every program: The technical baseline showing what will be delivered to the customer. The cost baseline showing what each deliverables and all it’s components will cost as a function of time. The schedule baseline showing when those deliverables will arrive and what is the sequence of work needed to arrive at the needed time. All three of these baselines must be integrated into a cohesive plan for the project to be have a chance of being successful. With this baseline, the challenge becomes how to measure progress to the plan.
  31. PGCS 2018 Master Workshop, Canberra Australia
  32. PGCS 2018 Master Workshop, Canberra Australia
  33. PGCS 2018 Master Workshop, Canberra Australia
  34. Here’s how these pieces are connected in the Integrated Program Performance Management System. We start with the statement of work and a Statement of Objectives. Along with those is the Concept Of Operations (ConOps). These three documents say what done looks like in some meaningful way to the end-user. From there we can build the work breakdown structure. And derived the technical and operational requirements. With those requirements in hand we can start building the Integrated Master Plan. As the program moves left to right the increasing maturity is measured in units of effectiveness and performance. In the Integrated Master Schedule progress to plan is measured with the Technical Performance Measures, showing how the produced deliverables comply with the needed Technical Performance Measures (TPM) to fulfill the Measures of Performance (MoP) that enable the Measures of Effectiveness (MoE).
  35. Starting with the Mission Need, here is the flow for Measures of Effectiveness, Measures a Performance, Key Performance Parameters, and the Technical Performance Measures used to assess Physical Percent Complete for our program. The increasing maturity of The Definition of Done goes from left to right. For our program to be successful we need to convert the vague mission needs into capabilities defined through measures of effectiveness. Then convert those Measures of Effectiveness to Measures of Performance and the Key Performance Parameters needed to ensure success. With those measures in place our Technical Performance Measures assess Physical Percent Complete on fine‒grained boundaries to inform Earned Value and Earned Schedule to calculate our Estimate to Complete. These measures integrate the technical baseline to schedule baseline and the cost baseline into an Integrated Program Performance Management System (IPPMS).
  36. PGCS 2018 Master Workshop, Canberra Australia
  37. PGCS 2018 Master Workshop, Canberra Australia
  38. PGCS 2018 Master Workshop, Canberra Australia
  39. Using the five immutable principles of Project success, we need to connect them with processes and practices. Here’s a picture of that integration. The five processes are Identify needed capabilities Identify the requirements baseline Establish the performance Management Baseline Execute that baseline Perform continuous risk management To the left of the chart are numbers 1 through 10 these of the practices supporting the 5 immutable principles
  40. Here’s a quick look at the 10 practices. This is too much detail for our short one day workshop but these practices must be in place during the execution of your actual program.
  41. Let’s get started on the notional program or a program of your choice.
  42. But let’s have one more quick look at the principles. They start with: What does Done look like? How do we get to done? What resources do we need along the way? What could possibly go wrong along the way and how are we going to fix it? How do we measure physical percent complete towards done?
  43. All of these questions have answers that are stated in Numbers. We need the numbers to make decisions. Those numbers have to have confidence precision and accuracy.
  44. Here’s where all those numbers come together Remember we had our three baselines? The cost baseline the schedule baseline and the technical baseline? There are some other components to those three baselines. We have our statement of work. We have the risk management plan. We have the technical performance measures for each of the deliverables. When those are all collected in a single integrated performance management system we’ve now established our performance measurement baseline.
  45. I’ve used the word credible several times let’s all agree what that means.
  46. The performance measurement baseline shown a few charts back integrates all of the data of her integrated performance management system. Here’s a picture from an actual program of that integration. So let’s start with the first element of that integrated program performance management system.
  47. It always starts with the Work Breakdown Structure (WBS). The WBS is paramount is a phrase we use in our defense acquisition community. What’s the WBS tells us is What are the deliverables? What is the relationship between each deliverable to its parent into its children. The WBS tells us what are the elements of done. These elements don’t have measures yet but those can be applied through the measures of effectiveness and measures up performance. During the performance of the work to produce the WBS element the technical performance measures assure us we’re complying with the measures of performance and the measures of effectiveness.
  48. I’m using the term Integrated Master Plan (IMP) here to mean any master plan that shows the capabilities when they are delivered and how they are meaningful to the decision-makers. In our US Community the Integrated Master Plan has three elements: The program events that assess the increasing maturity of the deliverables The Significant Accomplishments that implement the Program Events The Acceptance Criteria for each Work Package outcome that implement the Significant Accomplishment
  49. I’m going to introduce a notion of risk that is not commonly used in our defense community. But this notion of risk started in the scientific community for assessing the quality of the data from the field or form models. It can directly applied in our defense community for success. This starts with the notion that all risk comes from uncertainty. Risk it self does not stand alone. It is only the product of uncertainty. Uncertainty comes in only two forms. Reducible uncertainty which has the name Epistemic uncertainty. And irreducible uncertainty which has the name Aleatory uncertainty. Epistemic uncertainty comes from the term a Epistemology which means the study of knowledge Aleatory uncertainty comes from the term Alia which in Greek means a single die. Will get two more details of these uncertainties later.
  50. In our integration process we need to assign cost to packages of work. This cost assignment starts of course with the WBS. But then it needs to be flow to our schedule. This means a resource loaded cost for material Integrated Master Schedule is needed. Mini object to the effort required to produce this resource loaded schedule but without it we can’t see how we are measuring progress toward done
  51. In our Statement Of Work, it must clearly and concisely say the measures of effectiveness and measures of performance needed for program success.
  52. The last element of our Integrated Program Performance Management Systems are the Technical Performance Measures. These tell us how fast how big how much capacity how reliable how maintainable in fact all of the illities must be present for our product to be a success. These need to be written down before we start work otherwise we won’t know what the goal is when we’re doing that work.
  53. Let’s go back and look at the major components of our sample program.
  54. This program is much too big for our one-day workshop. Let’s pick one of the components of this system. And then will pick one of the elements of this component for our analysis.
  55. Going back to our starting point of defining the needed capabilities for mission success here’s an example of the airborne center platform and it’s needed capabilities. For your actual programs you should be able to find a similar document. If you don’t have one you do not have a description of what done looks like.
  56. Here’s where those measures of a performance measures of effectiveness in fact all of the measures live.
  57. PGCS 2018 Master Workshop, Canberra Australia
  58. Let’s go back to our WBS
  59. Here’s a summary of what I good WBS looks like. We have a standard in the US Mill standard 881C that describes the elements of a WBS. You have a similar standard in Australia.
  60. The WBS starts with the architecture of the system that is to be developed. The WBS of course is not actually work its products. It should be called the product breakdown structure. There is an urban wedge and that the WBS was developed in world war one by the Germans for tank production. And when translated into English we confuse the word product with work.
  61. Here’s the structure of the TSAS WBS
  62. And like all good WBS’s it has multiple layers
  63. So let’s look at the level one WBS for this program.
  64. And then the WBS for the unmanned air vehicle
  65. And then the WBS for the unmanned error vehicles flight avionics.
  66. And then the WBS for the get guidance navigation and control computer
  67. And then the WBS for the mission management system.
  68. PGCS 2018 Master Workshop, Canberra Australia
  69. In our Integrated Master Plan we collect the measures of performance and the measures of effectiveness. I was unable to find any references that the Integrated Master Plan wish used in Australian defense. So let me give a quick overview of the concept of the Integrated Master Plan The IMP is the strategy for success It shows the increasing maturity of the deliverables through accomplishments and accomplishments criteria.
  70. It’s the Integrated Master Plan and the Integrated Master Schedule where the technical performance measures live. Each of these performance measures effectiveness technical and operational are connected in the Integrated Master Plan and the Integrated Master Schedule.
  71. The Integrated Master Plan connects product with functions that produce the product.
  72. The Integrated Master Plan is not a replica of the work breakdown structure. The Integrated Master Plan shows how the work breakdown structure elements will be developed. That is what must be accomplished to increase the maturity of a program event.
  73. The Integrated Master Plan tells us where the program is going. The information provided by the Integrated Master Plan comes in units of effectiveness and performance. When we measure progress in compliance with our planned effectiveness and performance we have a notion of physical percent complete that is much more informative than earned value can ever tell us. Earned value now becomes the collection point for cost and schedule data that connects the project performance with this measures of effectiveness and measures of performance.
  74. Here’s the structure of the Integrated Master Plan and the Integrated Master Schedule. At the top of this pyramid are the events that assess the increasing maturity of the program. Sometimes these advances are referred to as milestones. But milestones are rocks on the side of the road measuring distance back to Rome. You only know how far you’ve come after you’ve passed that milestone. For each program of vent we need significant accomplishments that speak to the progress to plan in measures of effectiveness of the outcome of the program. These significant accomplishments collected the output of work packages. The outcomes of work packages are assessed through accomplishment criteria. The outcomes of work packages are measured through technical performance measures. These TPM’s show the technical compliance with the needed performance for each outcome of the work package.
  75. Here’s how the elements of the Integrated Master Plan and Integrated Master Schedule are connected. The IMP is the vertical representation of the program in measures of increasing maturity. The IMS is the horizontal representation of the work performed. Both vertical and horizontal representations are needed. In the US we have a guide that requires this DID 81681 but it’s not always performed.
  76. Here’s the role of the Integrated Master Plan during the execution of the program. as the program progresses with compliance with the technical performance measures. this compliance at the integrated Master schedule level is informed by the progress to plan for the work packages using the Technical Performance Measures. Each TPM is assessed at the task level for compliance with the exit criteria of the task using quantifiable back back data.
  77. Here’s an actual example of an Integrated Master Plan with the measures of effectiveness and measures her performance. This program is an autonomous landing system for the Navy’s F-18. We have measures of effectiveness measures of performance key performance parameters and the technical performance measures all defined at the proposal stage before any contract work starts. these measures of effectiveness in these measures of performance to the flight avionics systems that implement the joint precision autonomous landing Landing system.
  78. here’s an example of an Integrated Master Plan for the F-22 Raptor. It starts with the program event to assess the readiness or completion of the progress of the program. In this case first flight complete. The significant accomplishments needed to conduct the first flight complete or a flight readiness review. The criteria for the flight readiness review is that the seek eagle has flight clearance. The purpose of the Integrated Master Plan is to describe the increasing maturity of the deliverables of the program.
  79. Once we have the Integrated Master Plan we can now connect the increasing maturity of the program with the elements of the work breakdown structure. here’s an example from a spacecraft whose Integrated Master Plan starts with the systems requirements review goes to the systems design review, preliminary design review, critical design review, test readiness review and assembly test and launch operations. The dependencies between each WBS element are shown as the program matures
  80. The primary purpose of the Integrated Master Plan is to show what Don looks like in measures of effectiveness and measures of performance. Robert Goddard knew this when he first proposed flying to the moon
  81. PGCS 2018 Master Workshop, Canberra Australia
  82. We’re back to the WBS is Paramount as the starting point for constructing an credible Integrated Master Schedule.
  83. This is an actual sketch from the Redstone Arsenal and Huntsville Alabama where Warner von Braun defined how to fly astronauts to the moon. What’s interesting is he did not write down how to get them home.
  84. Here’s an example of an Integrated Master Plan and the associated Integrated Master Schedules.
  85. Building a credible Integrated Master Schedule requires 10 steps. Each step should be performed in the defined order. As each new staff is performed in the previous steps should be assessed to confirm they are in fact correct.
  86. We are back again to the WBS being paramount. It is the glue that holds the Ims together.
  87. With the completed I am asked we now have an integrated picture into the elements we need to assess the increasing maturity of the program.
  88. 88
  89. PGCS 2018 Master Workshop, Canberra Australia
  90. Managing risk is the primary role of Project Management. Technical and Operational management can be handled by others Project Management IS Risk Management.
  91. But risk is not a standalone entity All risk comes from uncertainty. Uncertainty creates risk. It is managing in the presence of uncertainty that is the primary role of project management.
  92. We need to have a clear understanding of what we mean when we say uncertainty.
  93. The traditional approach to risk management states there are two dimensions of risk. This is technically correct for the two classes of uncertainty that will be described in the next charts.
  94. Many times there is confusion between the difference of statistics and probability. Aleatory uncertainties have statistical behaviors. Remember Aleatory comes from Alea, means a single die. When tossed there is a statistical processes of determining the number that will appear. In the risk management business, we must separate irreducible uncertainty and he risk it creates from the reducible uncertainty and the risk it creates. Research shows that the majority of risks on development programs comes from Aleatory uncertainties. The traditional risk register only captures the reducible risks. The epistemic uncertainties that created those risks. While some risk management tools can’t handle both we must understand the difference between reducible and irreducible uncertainty. It’s the naturally occurring uncertainties and the resulting risk that will create the most problems for our program if we do not have sufficient cost schedule and technical margin to protect from those risk.
  95. I want to restate that risk comes from uncertainties. Risk it does not stand alone. Risk management starts with determining which uncertainties impact the project. There are other models of risk which do not based themselves on this principle. The principles stated in this chart come from NASA and similar mission critical agencies. Here’s a sample of the resources that can be found on the web Risk Informed Decision Making, NASA/SP-2010-576 Risk Management Handbook, NASA/SP-2011-3422 NASA Procedural Requirements ‒ Technical Probabilistic Risk Assessment (PRA) Procedures for Safety and Mission Success for NASA Programs and Projects, NPR 8505.5A Considering Risk and Resilience in Decision Making, NASA/TM-2015-218777 Probabilistic Risk Assessment Procedures Guide for NASA Managers and Practitioners, NASA/SP-2011-3421 Second Edition “A Study Of Uncertainty And Risk Management Practice Relative To Perceived Project Complexity,” A thesis submitted for the degree of Doctor of Philosophy by Craig Michael Harvett Bond University May 2013. “Characterizing Uncertainty to Manage Risk in Spacecraft Development with Application to Structures and Mass,” Emily Baker Clements, Massachusetts Institute of Technology, June 2013. Several of these documents have a subtitle that says compliance is mandatory
  96. Here is the taxonomy of uncertainty and the risks it creates. Print this out and put it in your notebook. Anytime anyone talks to you about risk gets handling and mitigation strategies take out this chart and have them point what type of risk they are talking about.
  97. This entire topic of risk management can be connected back to the work breakdown structure. One of the missing concepts in programs that have failed or are in trouble is the notion of risk of propagation through the work breakdown structure. Static risk registers with static risk management ignore this important principle. Risks propagate. A simple risk may create a complete failure of the program if it is not revealed handled and mitigated. This proverb tells it all For want of a nail the shoe was lost. For want of a shoe the horse was lost. For want of a horse the rider was lost. For want of a rider the message was lost. For want of a message the battle was lost. For want of a battle the kingdom was lost. And all for the want of a horseshoe nail.
  98. We then need to connect to the risks and their buydown or mitigation strategies to the Integrated Master Schedule. Risk buydown becomes a technical performance measure. If we are not buying down risks or providing proper schedule cost and technical margin as planned we cannot possibly meet the technical performance measures of the program as planned. Therefore we are late over budget and have a probability of our project not working and we don’t know it until it’s too late.
  99. Here is our final thought for this workshop on risk. But just remember as Tim Lister tells us Risk management is project management for adults.
  100. Mr. Gary bless is the director of performance assessment and root cause analysis division of the office of Secretary of Defense for acquisition technology and logistics. The root cause analysis commission to buy his division came up with for primary root causes for program poor performance.
  101. Here’s a collection of observations from root cause analysis that impacted the probability of success for all programs.
  102. But as always beware of the Black Swan. And I know in Australia black swans are common so be aware of the white Swan.
  103. PGCS 2018 Master Workshop, Canberra Australia
  104. Progress to plan should only be measured with tangible evidentiary materials. These measures must compare actual evidence with the planned values for that evidence. If the actuals do not meet to plan we cannot be on schedule we cannot be on budget and it is unlikely the technical outcome will be what we expected.
  105. This notion of measuring actuals to plan and taking corrective action is called closed loop control. All dynamic systems make use of closed loop control to stay inside the bounds of acceptable outcomes. This is the primary purpose of program planning and controls. This principle isn’t applicable to acquisition programs in the defense and space community critical infrastructure programs and hockey. We need to know where the program is going to be where we are now and how to close that gap. Earned value tells us where the program was and gives us a hint about where the program might be going. But I earned value and its companion earns schedule or not risk-adjusted. They don’t measure risk buydown they don’t measure schedule margin buydown and they don’t measure the impact of on mitigated risks. Risk adjusted earned value in earned schedule takes place when the the IMS, risk management process technical performance measures are integrated I a seamless Integrated Program Performance Management System (PPMS).
  106. Here’s a reminder what technical performance means.
  107. Here are some sources for technical performance measures, measures of effectiveness and measures of performance. These measures come from systems engineering. Without a systems engineering process the probability of success is dramatically reduced.
  108. In the past earned value has not addressed the systems engineering aspects of program management.
  109. 109
  110. 110
  111. 111
  112. 112
  113. 113
  114. 114
  115. 115
  116. 116
  117. 117
  118. 118
  119. 119
  120. 120
  121. 121
  122. 122
  123. 123
  124. 124
  125. 125
  126. 126
  127. 127
  128. 128
  129. 129
  130. 130
  131. PGCS 2018 Master Workshop, Canberra Australia
  132. Here’s the result of our integrated performance management system.
  133. Well earned value management Mutt is the best tool for managing large complex acquisition programs it must be informed by risk management technical performance measures all living inside of a integrated Master plan.
  134. The three elements of earned value management and its use on program performance management are cost schedule and technical performance. When we have only cost and schedule numbers from the earned value management system we are missing the critical success factor for all programs success. And that critical success factor is are we building the right thing for the right cost at the right time?
  135. The traditional earned value management system only addresses cost and schedule. In fact the traditional current value management system only addresses cost since the units of measure of schedule variance or dollars. Earned schedule provides additional insight into the programs performance. Without technical performance measures the assessment of these measures to plan earn schedule and earned value cannot tell us the probability of success for the program. All three are needed.
  136. As we’re nearing the end of our workshop, let’s revisit some important concepts. We can only measure progress to plan with assessment of Physical Percent Complete. To define Physical Percent Complete we must define the Measures of Effectiveness and Measures of Performance of the outcomes of the project. The Technical Performance Measures of the incremental work along the way to done. At the lowest level we need to define the Quantifiable Backup Data used to assess the completeness of the individual work elements. Then these measures can be collected and analyzed to assessment ‒ in a risk adjusted manner ‒ what the Estimate to Complete might be. Then we can use this information to produce the Earned Value and Earned Schedule data for management. But we must remember EV and ES are just numbers, the MoEs, MoPs, TPMs, and KPPs are the evidence supported by those and many other numbers needed to increase the probability of program success.
  137. 137
  138. PGCS 2018 Master Workshop, Canberra Australia