SlideShare ist ein Scribd-Unternehmen logo
1 von 32
Downloaden Sie, um offline zu lesen
Project Management Challenge 2007


Track: Case Study Insights at NASA

            Case Study:
Extracting Trends and Lessons Learned
 From Project Problems and Solutions
   NASA Goddard Space Flight Center

              February 6, 2007

                                     Mike Rackley          Catherine Traffanstedt
                            NASA GSFC Code 170          NASA GSFC – SGT, Inc
                                     301-614-7058                  301-925-1125
                      Michael.W.Rackley@nasa.gov    ctraffanstedt@sgt-mets.com
Presentation Outline

   The Vision
What are we trying to
     achieve?



                  Implementation
                   How do we achieve
                      the Vision?



                                   Current Status
                                       Where are we
                                         today?



                                                      Gap Analysis
                                                      How do we close
                                                         the gap?
                                                                        2
The Anomaly Management Vision

   The Vision
What are we trying to
     achieve?



                  Implementation
                   How do we achieve
                      the Vision?



                                   Current Status
                                       Where are we
                                         today?



                                                      Gap Analysis
                                                      How do we close
                                                         the gap?
                                                                        3
Anomaly Management Vision

•   GSFC has well documented requirements, procedures, and best
    practices for mission-related problem management, reporting,
    analysis, and trending

•   GSFC efficiently verifies adherence to the established problem
    management requirements and established best practices

•   All GSFC development and on-orbit missions use a consistent
    problem management process and philosophy
     – Flexible enough to handle the diversity of spacecraft missions in the
       organization, enforcing certain fundamental “rules” for all missions, but
       allowing some differences where appropriate in other areas
         • Example: HST vs. Small Explorer Program
         • Recognizes that risks and budgets can be very different among programs
     – Applies to all missions that the organization is responsible for,
       regardless of whether or not the missions are built and/or flown “in-
       house”
         • NASA responsible for mission success regardless of approach to
           development or operations (i.e., in or out of house)
         • Out-of-house missions do pose unique challenges
                                                                                    4
Anomaly Management Vision (cont)

•   GSFC provides online access to accurate mission configuration
    information for all Goddard managed missions in development and
    operations
    – Example: Subsystems, components, orbit configuration, duty cycles

•   GSFC analyzes and trends problems across development and on-
    orbit missions
    – Individual missions, across mission “families”, across all missions
    – Consistent use of a variety of Categorizations (e.g., failure type, root
      cause, impact)

•   GSFC extracts appropriate “Lessons Learned” information from the
    problem and solution space and apply these back into the
    development and operations processes
    – Examples: improvements in Rules/Best Practices for development,
      testing, and operations; planned vs. actual hardware & software
      performance



                                                                                 5
Problem Management
                                    Information Flow

                          Mission 1       . . .                   Mission N
                                                                                     • In-House
                                                                                       Development
                                                                                     • Out-of-House
                                                                                       Procurement


                                   Mission Anomaly Information                                  Impediments:
                                                                                                • Project focus
  Mission                                • Indirect (export)                                    • Heterogeneous
Configuration                            • Direct                                                 missions
                                                                • Record
 Definitions                                                      anomaly                          –   Design
                                                                  information                      –   Timeframe
                                           Centralized
                                          Centralized           • Provide                          –   Orbit
 Examples:                                                                                         –   Ops/science
                                            Problem
                                           Problem                search/query
 • HW & SW                                                                                             profile
   descriptions                       Information System
                                     Information System           capability
                                                                                                • Contracts
 • Design life info                                             • Generate
                                                                  reports                          – End-item
 • Failure                                                                                         – Proprietary
   handling                     • Trending                      • Support                            data
   approach (e.g.,              • Analysis                        individual and                • Culture
   single vs. dual                                                multi-mission
                             • Data Mining                        “views”                       • Sustaining over
   string)                                                                                        time


                                                                                     Cross-
                  Multi-mission                      Ops/         Lessons
   Output                           Failure                                         mission
                   Summary                        Engineering     Learned
                                   Analysis                                        Trending &
 Products           Reports                        Research      Extraction
                                                                                    Analysis

                                                                                                                     6
Lessons Learned Feedback Process


                                 Knowledge Management

     Ops process or
 procedure proves to be                                                          Feedback /
  successful over time                                                        recommendations
                                                                                 provided to
                                            Lessons                            spacecraft flight
 Ops experiences actual                                                            projects
 behavior of spacecraft
                                            Learned
                                           established
                                                                              Best practices
 Anomaly resolved with                                                         established
defined corrective action                                                          and
                                                                               documented
     Anomaly
      occurs
                                       Influencing Variables
    • Spacecraft design / technology       • Risk tolerance           • Personnel changeover
    • Ground system design / technology    • Organizational culture   • Mission models/profiles
                                                                                                   7
Implementing The Vision

   The Vision
What are we trying to
     achieve?



                  Implementation
                   How do we achieve
                      the Vision?



                                   Current Status
                                       Where are we
                                         today?



                                                      Gap Analysis
                                                      How do we close
                                                         the gap?
                                                                        8
Implementation Approach

•   Ensure appropriate requirements are officially documented so that
    they apply to all missions in development and operations
     – Goddard Procedures and Requirements Documents (GPR’s)
     – Mission Assurance Requirements (MAR) Document (per mission), with
       the Mission Assurance Guideline (MAG) document as a template


•   Establish and document rules/best practices for how to document
    anomalies
     – Main goal is to achieve the needed consistency and robustness across
       missions
     – Anomaly Management Best Practices Handbook (or comparable)


•   Utilize a centralized, independent operations mission assurance
    team to verify adherence across missions to requirements and best
    practices
     – Provides ability to recognize and correct problems early
                                                                              9
Implementation Approach (cont)

•   Utilize a centralized anomaly trending and analysis team to perform
    various trending and analysis functions within and across missions
     – Assist in development of Mission Configurations
     – Analyze trends
     – Extract “lessons learned” and work to apply them across the relevant
       organizations, documents, etc.
         • Implement a “Lessons Learned CCB” with broad participation to disposition
           potential/proposed lessons learned and best practices

•   Utilize a centralized database to capture problem report data that
    provides ability:
     – For individual missions to document, track, and close out their problems
     – For the Center to perform various analysis and trend functions across
       missions
     – To handle development/test (pre-launch) and operations phases
     – To handle flight and ground problems
     – To bring in data from outside users
     – To provide broad access, but with appropriate access protections (for
       ITAR sensitive and proprietary data)

                                                                                       10
Implementation Approach (cont)

•   The centralized database is the Goddard Problem Reporting System
    (GPRS), which contains two modules for documenting and
    managing ground and flight segment problems
    – Both provide multi-mission, web-based, secure user access
    – Other features include Mission Configuration definition, search/query
      capabilities, user notification (via email), user privilege control, user
      problem reporting, and attachments

•   SOARS -- Spacecraft Orbital Anomaly Reporting System
    – Used to document flight and ground problems during operations
    – Problems categorized via a common approach, using various
      Categorization Fields (Anomaly Classification, Impacts, Root Cause)

•   PR/PFR -- Problem Reporting/Problem Failure Reporting System
    – Used to capture problems encountered during the development,
      integration, and test
    – Similar to SOARS, but tailored to the I&T environment (e.g., the
      signature process)
                                                                                  11
SOARS Model
                                                  Maximize Consistency &
                                                   Commonality in how
          Support Mission Anomaly               Anomalies are Documented &
                                                       Categorized
               Management
          (document, track, report, notify)

                                                    Document Mission
                                                       Configurations
                                                (mission-unique and common
                                                 subsystems & components)
Anomaly
 Event                     SOARS
                           SOARS
 Occurs                                            Perform Cross-Mission
                                                    Trending & Analysis


                                                Support Independent Mission
                                                   Assurance Activities
                                                     (analysis, audits)
            Support Cross-Mission
                  Analysis
             (have others had a similar
                                                 Provide Common Approach to
                     problem?)
                                                  Maintenance, Security, etc.


                                                                                12
Current Status:
                        Where Are We Today?

   The Vision
What are we trying to
     achieve?



                  Implementation
                   How do we achieve
                      the Vision?



                                   Current Status
                                       Where are we
                                         today?

                                                  s
                                            lenge
                                       C hal          Gap Analysis
                                                      How do we close
                                                         the gap?
                                                                        13
Current Status

•   Requirements & Best Practices Development/Compliance

    – High-level requirements associated with using GPRS and performing
      quality assurance, trending, and analysis are in place
        • GPR 1710.H - Corrective and Preventative Action
        • GPR 5340.2 - Control of Nonconformances

    – Mission Assurance Requirements Documents, with the Mission
      Assurance Guideline (MAG) document as a template also generally
      cover the requirements
        • But, capturing problem information in the central PR/PFR System limited to
          in-house missions for pre-launch phase

    – Operations Rules/Best Practices not yet formally documented in a
      central location that would apply across all missions
        • Exist in pockets within missions, families of missions, and Programs
        • Working with missions and Programs to begin development



                                                                                       14
Current Status (cont)

•   Operational Improvements

    – Major GPRS enhancements made to SOARS and PR/PFR software
      over the last year to improve operational usability, etc.
        • SOARS now considered an operationally useful tool, with ongoing activities
          to continue incremental improvements
        • PR/PFR has more recently undergone improvement, largely driven by two
          Goddard in-house missions (SDO and LRO)
            – PR/PFR needs much more improvement to “catch-up” to SOARS


    – SOARS use by the operational missions has increased dramatically
      over the last year
        • One year ago very few missions were using SOARS at Goddard
        • Almost all in-house missions (~20) now using SOARS effectively, including
          RXTE, Terra, Aqua, EO-1, and TDRSS
        • Only one out-of-house mission currently populating SOARS with anomaly
          report data (Swift at Penn State University, as a direct user)
            – Other out-of-house “operations vendors” include APL, Orbital, LASP, and
              University of California at Berkeley (UCB)
            – Currently working with these organizations to develop plan for populating SOARS
              (e.g., make use of data import)


                                                                                                15
Current Status (cont)

•   Anomaly Analysis and Trending

    – Centralized teams in place within Office of Mission Assurance (Code
      300) for performing roles of operations mission assurance and cross-
      mission anomaly analysis and trending
        • Development phase handled by individual Code 300 System Assurance
          Managers (SAM’s) assigned to missions

    – GSFC (Code 300) using SOARS data extraction to:
        • Provide monthly Center-wide snapshot of mission anomalies, characterized
          by Mission Impact, Operations Impact, and Root Cause
        • Develop annual “year-in-review” anomaly summary report (OAGS)
        • Trend and analyze problems across missions, subsystems, failure types,
          etc.
        • Support long term reliability studies on flight hardware
        • Identify missions that are not adequately documenting anomalies and work
          with them to improve

                                                                                     16
SOARS Data Extraction – Example 1
                            SOAR Records - All active Goddard Missions                                                             Root Cause Categorization
                                      (12 month summary)                                                                    for Cum CY06 "Ops Workaround" SOAR's
50
                                                                                          Ground Segment SOARs
45                                                                                        Flight Segment SOARs                                         Design, 8
40
35                                                                                                                                                             Workmanship, 4
30
25                                                                                                                     Unknown, 28
                                                                                                                                                                     Process, 4
20
                                                                                                                                                                      Environmental, 2
15
10
                                                                                                                                                                     Degradation, 3
    5
    0
                                      6




                                                                                               6
         06




                                                        06




                                                                           06


                                                                                     06




                                                                                                        06


                                                                                                                  06
                                                                  6
                   06



                             6




                                               6
                                  r-0




                                                                                              -0
                                                              l- 0
                            -0




                                              -0
        n-




                                                    n-




                                                                       g-


                                                                                 p-




                                                                                                    v-


                                                                                                              c-
                                                                                                                                                              Hardware Failure, 18
               b-




                                                                                           ct
                                          ay
                         ar




                                                             Ju
                                 Ap




                                                                                                   No


                                                                                                             De
                                                                      Au


                                                                                Se
    Ja




                                                   Ju
              Fe




                                                                                          O
                        M




                                          M




                                                                                                                              Human Error, 3
                                                                                                                                    Training, 1
                    Significant Nov’06 On-orbit Anomalies
                                                                                                                                 Root Cause Categorization
•            Msn-X power load shed anomaly
                                                                                                                             for Cum CY06 "Data Loss" SOAR's
              –         All science instrument main voltages shed in accordance to the
                                                                                                                                                  Design, 3
                        Under Voltage protection Circuitry                                                                                                Workmanship, 2
              –         Under voltage condition, possibly indicating a failure (short)                                                                      Process, 0
                        somewhere in the affected loads                                                                                                         Environmental, 2
                                                                                                                                                                    Degradation, 2
•            Msn-Y Star Tracker and Sun Sensor Boresight anomaly
              –         Star Tracker alignment calibration analysis indicates over a 2 arc-
                        minute shift in the alignment between the Star Tracker and FSS                                                                              Hardware Failure, 7

•            Msn-Z ACS Quaternion Spike Error anomaly
              –         Star Tracker Quaternion element was corrupted
                                                                                                                                                                   Training, 0
              –         Spacecraft commanded to Sun/Mag pointing and the Star Tracker                                      Unknown, 26
                                                                                                                                                            Human Error, 2
                        stabilized
SOARS Data Extraction – Example 2
                                                            Root Cause Categorization                                                              CY06 Cumulative Root Cause Categorizations
  1                Design Related
0.9
                   Hardware Degradation/Failure                                                                                                                                     Design Related, 16%
0.8
                   Operations Related                                                                                                                                                        Hardware
0.7
0.6
                   Environmentally Related                                                                                                                                             Degradation/Failure, 20%
0.5
                   Unknown
0.4
0.3
                                                                                                                                                                                      Operations Related, 4%
0.2                                                                                                                                                 Unknown, 50%
0.1
  0
                                                                                                                                                                                     Environmentally Related,




                                                                                                                          6
       06




                                                                   06



                                                                               6
                   06



                                  6


                                               6




                                                                                        06


                                                                                                  06




                                                                                                                                      6
                                                            6




                                                                                                             6
                                                                           l -0
                                  -0




                                                                                                                       v-0
                                           r- 0



                                                           -0




                                                                                                                                  c-0
                                                                                                           t- 0
                                                                                                                                                                                             10%
       n-




                                                                   n-
                   b-




                                                                                    g-


                                                                                               p-
                              ar




                                                       ay




                                                                          Ju
                                          Ap




                                                                                                       Oc


                                                                                                                   No
      Ja




                                                                  Ju
                  Fe




                                                                                   Au


                                                                                             Se




                                                                                                                              De
                              M




                                                       M




                                                 Anomaly Classification Categorization                                                    CY06 Cumulative Anomaly Classification Categorization
100%
                                                                                                                           Hardware                       Other, 5%
 90%
 80%
                                                                                                                           Software
 70%
                                                                                                                           Operations            Operations, 13%
                                                                                                                           Other
60%                                                                                                                                                                                          Hardware, 41%
50%
40%
30%
20%
10%
 0%
                                                                     06




                                                                                                                       06
                                                                               6
                                     6




                                                                                                              6
             06




                                                                                        6
                                                             6
                     06




                                               6




                                                                                                  6




                                                                                                                                  06
                                                                           l- 0
                                  -0




                                                           -0




                                                                                                           -0
                                                                                       0
                                           r-0




                                                                                                 0




                                                                                                                                                     Software, 41%
                                                                   n-




                                                                                                                      -
           n-


                   b-




                                                                                    g-


                                                                                              p-




                                                                                                                                 -
                                                                                                                   ov
                              ar




                                                                                                       ct
                                                       ay




                                                                                                                              ec
                                                                          Ju
                                          Ap




                                                                  Ju
      Ja




                                                                                   Au
                  Fe




                                                                                             Se


                                                                                                       O
                              M




                                                                                                                   N
                                                       M




                                                                                                                              D




                                                Operations Impact Categorization                                                                     CY06 Cumulative Mission Impact Categorization
100%
                                                                                                  Mission Data Loss
 90%                                                                                                                                               Permanent Loss of                      Mission Data Loss,
                                                                                                  Operational Workaround
 80%                                                                                                                                                 Capacity, 0%                                28%
 70%
                                                                                                  No Impact
 60%                                                                                              Permanent Loss of Capacity
 50%
                                                                                                                                                  No Impact, 42%
 40%
 30%
 20%
 10%
  0%
                                                                                                                                                                                           Operational
                                                                                                                       06
                                                                     06



                                                                               6
                                      6




                                                                                                               6
                                                                                        6
             06




                                                              6
                         06




                                                   6




                                                                                                  6




                                                                                                                                  06
                                                                            l- 0
                                   -0




                                                            -0




                                                                                                            -0
                                                                                       0
                                               r-0




                                                                                                 0
                                                                   n-




                                                                                                                       -
           n-


                       b-




                                                                                    g-


                                                                                              p-




                                                                                                                                  -
                                                                                                                    ov
                                  ar




                                                                                                           ct
                                                           ay




                                                                                                                                                                                         Workaround, 30%
                                                                                                                               ec
                                                                          Ju
                                          Ap




                                                                  Ju
       Ja




                                                                                   Au
                   Fe




                                                                                             Se



                                                                                                       O
                              M




                                                                                                                   N
                                                       M




                                                                                                                              D
Current Status (cont)

•   Inserting “Lessons Learned” into the Process

    – Implementing an “Operations Lessons Learned” process, but still in the
      very early stages
    – Plan is to implement a process where a Lessons Learned
      Committee/Board is presented with potential lessons
         • Committee determines if lesson valid, and if so, the actions that are appropriate
           to take it forward
         • Working to engage broad participation (mission assurance, missions, and
           engineering)

       Lesson extracted                 Lesson vetted
                                                                         Action
       from Operations                       by
                                                                      taken on LL
       event, activity, etc.            LL Committee
                                                                Examples:
                                                                •  Ops best practice update
                                                                •  Improvement in testing
                                                                •  Feedback on hardware
                                                                   performance issues
                                                                •  Recommendation to flight
                                                                   software group
                                                                •  Recommendation for “Gold
                                                                   Rules” update
                                                                                               19
Sample Lessons Learned Summary

          Anomaly          Anomaly                      Proposed
Mission
            Date                                                                       Action
                          Description                Lesson Learned
Msn_X     10/3/06    Spacecraft clock           Establish best practice of      Work directly with Ops
                     rollover to an illegal     evaluating registers prior to   Projects to document
                     1998 date; on-board        launch, and as part of the      as Ops Best Practice.
                     clock register capacity    on-orbit ops procedures.        Work with Flight
                     was exceeded.              Include FSW capabilities to     Software Branch to
                                                issue warnings when             determine if design
                                                appropriate.                    best practice needed.

Msn_Y     10/19/06   UV instrument              Lesson tbd, but would apply     Investigate further.
                     experiencing channel       to improving test approach.     Discuss with Msn_Y
                     noise, apparently          Seems this should have          SAM and I&T Manager.
                     associated with heater     been caught in instrument       Generate independent
                     activation and             or observatory I&T. Why         analysis report if
                     environmental              was this not seen during        warranted.
                     conditions. Causing        thermal vac? Does this
                     failure of instrument to   have implications to other
                     meet signal to noise       missions? Could or should
                     ratio requirements for     this instrument sensitivity
                     the data channel.          have been better tested
                                                during instrument I&T?
Implementation Challenges

•   Achieving consistency across missions in how anomalies are
    documented (e.g., terminology, level of detail) and categorized is
    proving to be very difficult
     – Many missions with different ops teams have different taxonomies,
       operations cultures, etc.
     – Missions themselves also very different in terms of bus design, payload
       suite, operations profile, age, and risk tolerance

•   Mission Configuration definitions in SOARS not well-defined across
    all missions, making it difficult to perform cross-mission analyses
     – Not all information is readily available to use in SOARS definitions,
       especially for older missions (e.g., component manufacturers and
       design life)
     – Missions have struggled to put in the extra time to populate SOARS with
       their Mission Configurations
     – Working with the existing missions to improve the definitions, but effort
       is time consuming and challenging
     – Working with upcoming missions to get the definitions in place prior to
       launch (e.g., GLAST, SDO, and LRO)

                                                                                   21
Implementation Challenges (cont)

•   Proving very difficult to get out-of-house operations centers/teams to
    document problems in SOARS
     – Only one direct user so far (Swift/PSU)
     – Working with others to take advantage of enhanced SOARS data import
       capability
     – Finding that the critical factor is getting the agreement more explicitly
       planned well before launch

•   Only getting pre-launch problem report data into PR/PFR for in-
    house missions
     – Leaves a big “information gap” given the high number of out-of-house
       missions
     – Vendors cannot use PR/PFR directly, but working to determine if certain
       types of problem report data can be imported into PR/PFR




                                                                                   22
Gap Analysis

   The Vision
What are we trying to
     achieve?



                  Implementation
                   How do we achieve
                      the Vision?



                                   Current Status
                                       Where are we
                                         today?



                                                      Gap Analysis
                                                      How do we close
                                                         the gap?
                                                                        23
Gap Analysis:
             How Would You Complete “The Job”?

1. How would you approach achieving and maintaining more
   consistency across missions in how anomalies are documented?
    – Getting them to do it consistently over time
    – Working to measure/enforce consistency over time


2. Do you think it is possible to develop a set of Operations Anomaly
   Management Best Practices that would reasonably apply across a
   set of heterogeneous missions, that are a mix of in-house and out-of-
   house operations approaches?
    – If so, how would you approach getting this defined and keeping it
      updated?
    – If not, what can be done? Do you just give up?


3. What type of information besides the basic anomaly information
   would you want to capture for your missions to support long term
   trending, reliability/failure analysis, etc?
                                                                           24
Gap Analysis:
             How Would You Complete “The Job”?

4. How would you ensure that the relevant Mission Configuration
   information for upcoming missions is documented so that it is
   usable by the problem management system?
    –   Address the in-house vs. out-of-house challenges

5. How would you approach getting more out-of-house operations
   centers/teams to document their anomalies in the centralized
   problem information system (SOARS)?

6. How do you extract problem information from out-of-house mission
   vendors for use in the centralized problem information system
   (PR/PFR) during the pre-launch development phase?
    –   Will likely have proprietary data issues

7. How do you integrate lessons learned into projects/programs so
   that they will actually affect mission development and operations,
   and maintain benefit over time?
                                                                        25
BACKUP SLIDES




                26
SOARS Anomaly Report Form




                            27
SOARS Anomaly Report Form (cont)




                                   28
SOARS Anomaly Report Form
         Observation area includes all problem descriptions
         including orbit and attitude data, and other user
         defined fields

         Investigation area provides a place for the
         operations team to document the problem
         investigation and status. Includes an investigative
         log, subsystem and component identification,
         anomaly criticality, status, and responsibility
         assignment.

         Resolution area provides the operations team an text
         area to capture a synopsis of the problem cause and
         resolution.

         Assessment area is used to identify follow-on
         activities or recommendations and categorize the
         SOAR for generic use.

         Notification area is used to send email notifications
         to personnel (either other SOARS users or other
         contacts identified by missions)
                                                             29
SOARS Data Summary

•   All SOARS anomaly reports will include the following information:
     –   Anomaly occurrence time
     –   Anomaly title
     –   Anomaly category (flight or ground segment specification)
     –   Anomaly Subsystem and Component identification
     –   Anomaly description
     –   Closure information
     –   Follow-on recommendations
     –   Anomaly categorization

•   SOARS may include the following types of information:
     –   Anomaly assignment and responsible party
     –   Investigation information
     –   Orbital information
     –   Mission unique phase and mode
     –   Additional file attachments as appropriate



                                                                        30
SOAR Data Categorization

•   SOAR Categorization Fields are the key fields used for trending and
    analysis across missions for the Center. The fields include:

     – Mission Impact: describes the actual anomaly impact on the ability to meet
       Science and/or Mission Goals or requirements
         • No Effect, Minor, Substantial, Major, Catastrophic, Undetermined

     – Impact Severity on the Anomalous Item: describes the actual impact to the
       subsystem or component where the anomaly occurred
         • No Effect, Minor, Major, Catastrophic, Undetermined, Not applicable

     – Operations Impact: describes the effect of the anomaly on mission and/or
       science operations
         • Data Loss, Service Loss, Mission Degradation/Loss, Undetermined, Ops Workaround,
           None

     – Anomaly Classification: describes where within the mission systems the failure
       occurred
         • Hardware, Software, Operations, Other

     – Anomaly Root Cause: describes the root cause of the anomaly.
         • Design, Workmanship/Implementation, Process, Environmental, Degradation, Hardware
           Failure, Training, Human Error, Unknown

                                                                                               31
One Size Fits All Anomaly Management (AM) System:
                               Does It Exist?



                              Mission Phases
           (S/C I&T, Observatory I&T, L&EO, Normal Ops, Disposal)


                  •   AM System must meet the needs of individual
Missions




                      missions and of the broader multiple mission
                      perspective (e.g., looking for cross mission trends)

                  •   Mission may want one discrepancy/anomaly
                      management system through all mission phases, but
                      the phases are very different environments (pre and
                      post-launch)

                  •   AM System may have to handle spacecraft and
                      ground anomalies

                  •   AM System inevitably must be able to interface with
                      other AM Systems used by individual missions

                                                                             32

Weitere ähnliche Inhalte

Was ist angesagt?

Ray hines stegeman
Ray hines stegemanRay hines stegeman
Ray hines stegemanNASAPMC
 
Vonnie simonsen
Vonnie simonsenVonnie simonsen
Vonnie simonsenNASAPMC
 
Thomas.mc vittie
Thomas.mc vittieThomas.mc vittie
Thomas.mc vittieNASAPMC
 
Rovinsky final
Rovinsky finalRovinsky final
Rovinsky finalNASAPMC
 
Hihn.jarius
Hihn.jariusHihn.jarius
Hihn.jariusNASAPMC
 
Reliability Engineering Blended Learning Training
Reliability Engineering Blended Learning TrainingReliability Engineering Blended Learning Training
Reliability Engineering Blended Learning TrainingRicky Smith CMRP, CMRT
 
Matt.gonzales
Matt.gonzalesMatt.gonzales
Matt.gonzalesNASAPMC
 
Hr Transformation Overview
Hr Transformation OverviewHr Transformation Overview
Hr Transformation OverviewMark Gavora
 
Osterkamp jeff
Osterkamp jeffOsterkamp jeff
Osterkamp jeffNASAPMC
 
Planificación del proyecto análisis de riesgo
Planificación del proyecto   análisis de riesgoPlanificación del proyecto   análisis de riesgo
Planificación del proyecto análisis de riesgoProColombia
 
Phillip.hall
Phillip.hallPhillip.hall
Phillip.hallNASAPMC
 
NG BB 05 Roles and Responsibilities
NG BB 05 Roles and ResponsibilitiesNG BB 05 Roles and Responsibilities
NG BB 05 Roles and ResponsibilitiesLeanleaders.org
 
Ejecución del proyecto gestión de problemas
Ejecución del proyecto    gestión de problemasEjecución del proyecto    gestión de problemas
Ejecución del proyecto gestión de problemasProColombia
 
Jim.free
Jim.freeJim.free
Jim.freeNASAPMC
 
Bobby.german
Bobby.germanBobby.german
Bobby.germanNASAPMC
 
Simulación del proyecto
Simulación del proyectoSimulación del proyecto
Simulación del proyectoProColombia
 
Change Management
Change ManagementChange Management
Change Managementpromoking
 

Was ist angesagt? (20)

Ray hines stegeman
Ray hines stegemanRay hines stegeman
Ray hines stegeman
 
Vonnie simonsen
Vonnie simonsenVonnie simonsen
Vonnie simonsen
 
Thomas.mc vittie
Thomas.mc vittieThomas.mc vittie
Thomas.mc vittie
 
Rovinsky final
Rovinsky finalRovinsky final
Rovinsky final
 
Hihn.jarius
Hihn.jariusHihn.jarius
Hihn.jarius
 
Reliability Engineering Blended Learning Training
Reliability Engineering Blended Learning TrainingReliability Engineering Blended Learning Training
Reliability Engineering Blended Learning Training
 
Matt.gonzales
Matt.gonzalesMatt.gonzales
Matt.gonzales
 
Hr Transformation Overview
Hr Transformation OverviewHr Transformation Overview
Hr Transformation Overview
 
Osterkamp jeff
Osterkamp jeffOsterkamp jeff
Osterkamp jeff
 
Grieman
GriemanGrieman
Grieman
 
Planificación del proyecto análisis de riesgo
Planificación del proyecto   análisis de riesgoPlanificación del proyecto   análisis de riesgo
Planificación del proyecto análisis de riesgo
 
Phillip.hall
Phillip.hallPhillip.hall
Phillip.hall
 
NG BB 05 Roles and Responsibilities
NG BB 05 Roles and ResponsibilitiesNG BB 05 Roles and Responsibilities
NG BB 05 Roles and Responsibilities
 
Ejecución del proyecto gestión de problemas
Ejecución del proyecto    gestión de problemasEjecución del proyecto    gestión de problemas
Ejecución del proyecto gestión de problemas
 
Symons
SymonsSymons
Symons
 
Jim.free
Jim.freeJim.free
Jim.free
 
Bobby.german
Bobby.germanBobby.german
Bobby.german
 
Simulación del proyecto
Simulación del proyectoSimulación del proyecto
Simulación del proyecto
 
NG BB 01 Introduction
NG BB 01 IntroductionNG BB 01 Introduction
NG BB 01 Introduction
 
Change Management
Change ManagementChange Management
Change Management
 

Andere mochten auch

Mc namara.karen
Mc namara.karenMc namara.karen
Mc namara.karenNASAPMC
 
Rudolphi mike
Rudolphi mikeRudolphi mike
Rudolphi mikeNASAPMC
 
Wood frank
Wood frankWood frank
Wood frankNASAPMC
 
Paradis william
Paradis williamParadis william
Paradis williamNASAPMC
 
Law.richard
Law.richardLaw.richard
Law.richardNASAPMC
 
O'keefe william
O'keefe williamO'keefe william
O'keefe williamNASAPMC
 
Baniszewski john
Baniszewski johnBaniszewski john
Baniszewski johnNASAPMC
 
Trahan stuart
Trahan stuartTrahan stuart
Trahan stuartNASAPMC
 
Sampietro marco
Sampietro marcoSampietro marco
Sampietro marcoNASAPMC
 
Vellinga joe
Vellinga joeVellinga joe
Vellinga joeNASAPMC
 
Seftas krage
Seftas krageSeftas krage
Seftas krageNASAPMC
 
Roberts karlene
Roberts karleneRoberts karlene
Roberts karleneNASAPMC
 
Snow lee
Snow leeSnow lee
Snow leeNASAPMC
 
Yew manson
Yew mansonYew manson
Yew mansonNASAPMC
 
Stock gahm
Stock gahmStock gahm
Stock gahmNASAPMC
 
Wessen randi (cd)
Wessen randi (cd)Wessen randi (cd)
Wessen randi (cd)NASAPMC
 
Wood frank
Wood frankWood frank
Wood frankNASAPMC
 
Humphreys.gary
Humphreys.garyHumphreys.gary
Humphreys.garyNASAPMC
 
Muller ralf
Muller ralfMuller ralf
Muller ralfNASAPMC
 
Bejmuk bo
Bejmuk boBejmuk bo
Bejmuk boNASAPMC
 

Andere mochten auch (20)

Mc namara.karen
Mc namara.karenMc namara.karen
Mc namara.karen
 
Rudolphi mike
Rudolphi mikeRudolphi mike
Rudolphi mike
 
Wood frank
Wood frankWood frank
Wood frank
 
Paradis william
Paradis williamParadis william
Paradis william
 
Law.richard
Law.richardLaw.richard
Law.richard
 
O'keefe william
O'keefe williamO'keefe william
O'keefe william
 
Baniszewski john
Baniszewski johnBaniszewski john
Baniszewski john
 
Trahan stuart
Trahan stuartTrahan stuart
Trahan stuart
 
Sampietro marco
Sampietro marcoSampietro marco
Sampietro marco
 
Vellinga joe
Vellinga joeVellinga joe
Vellinga joe
 
Seftas krage
Seftas krageSeftas krage
Seftas krage
 
Roberts karlene
Roberts karleneRoberts karlene
Roberts karlene
 
Snow lee
Snow leeSnow lee
Snow lee
 
Yew manson
Yew mansonYew manson
Yew manson
 
Stock gahm
Stock gahmStock gahm
Stock gahm
 
Wessen randi (cd)
Wessen randi (cd)Wessen randi (cd)
Wessen randi (cd)
 
Wood frank
Wood frankWood frank
Wood frank
 
Humphreys.gary
Humphreys.garyHumphreys.gary
Humphreys.gary
 
Muller ralf
Muller ralfMuller ralf
Muller ralf
 
Bejmuk bo
Bejmuk boBejmuk bo
Bejmuk bo
 

Ähnlich wie Rackley mike

Removing the Systemic Project Barriers
Removing the Systemic Project BarriersRemoving the Systemic Project Barriers
Removing the Systemic Project BarriersJorvig Consulting Inc.
 
Clarify And Connect 4 12 10
Clarify And Connect 4 12 10Clarify And Connect 4 12 10
Clarify And Connect 4 12 10gomezgrp
 
Agile Estimation And Planning Part I
Agile Estimation And Planning Part IAgile Estimation And Planning Part I
Agile Estimation And Planning Part IKevin Zamora
 
Planificación del proyecto estimación
Planificación del proyecto   estimaciónPlanificación del proyecto   estimación
Planificación del proyecto estimaciónProColombia
 
Blending Methods To Succeed Comparing Prince2 S Agility With Scrum Within The...
Blending Methods To Succeed Comparing Prince2 S Agility With Scrum Within The...Blending Methods To Succeed Comparing Prince2 S Agility With Scrum Within The...
Blending Methods To Succeed Comparing Prince2 S Agility With Scrum Within The...thavo001
 
It's the organisation, Stupid
It's the organisation, StupidIt's the organisation, Stupid
It's the organisation, StupidJohn M Walsh
 
Introduction to Agile and Scrum (Montana Programmers Meetup Jan 2012).pptx
Introduction to Agile and Scrum (Montana Programmers Meetup Jan 2012).pptxIntroduction to Agile and Scrum (Montana Programmers Meetup Jan 2012).pptx
Introduction to Agile and Scrum (Montana Programmers Meetup Jan 2012).pptxDesigned Culture
 
Build Your NGO: Monitoring & Evaluation
Build Your NGO: Monitoring & Evaluation Build Your NGO: Monitoring & Evaluation
Build Your NGO: Monitoring & Evaluation Allie Hoffman
 
NG BB 43 Standardized Work
NG BB 43 Standardized WorkNG BB 43 Standardized Work
NG BB 43 Standardized WorkLeanleaders.org
 
uTest and Crowdsortium Webinar: Scope & Briefs
uTest and Crowdsortium Webinar: Scope & BriefsuTest and Crowdsortium Webinar: Scope & Briefs
uTest and Crowdsortium Webinar: Scope & BriefsCrowdsortium
 
OCC Group Capability Presentation Linkedin 2010
OCC Group Capability Presentation   Linkedin 2010OCC Group Capability Presentation   Linkedin 2010
OCC Group Capability Presentation Linkedin 2010harrychristiansen
 
Key Considerations for a Successful Hyperion Planning Implementation
Key Considerations for a Successful Hyperion Planning ImplementationKey Considerations for a Successful Hyperion Planning Implementation
Key Considerations for a Successful Hyperion Planning ImplementationAlithya
 
Customer Development Past Present Future Steve Blank 111909
Customer Development Past Present Future Steve Blank 111909Customer Development Past Present Future Steve Blank 111909
Customer Development Past Present Future Steve Blank 111909Stanford University
 
Eileen may webcast
Eileen may webcastEileen may webcast
Eileen may webcastdhadobas
 
¿Cómo gestionar un proyecto aspectos clave
¿Cómo gestionar un proyecto   aspectos clave¿Cómo gestionar un proyecto   aspectos clave
¿Cómo gestionar un proyecto aspectos claveProColombia
 

Ähnlich wie Rackley mike (20)

Removing the Systemic Project Barriers
Removing the Systemic Project BarriersRemoving the Systemic Project Barriers
Removing the Systemic Project Barriers
 
Clarify And Connect 4 12 10
Clarify And Connect 4 12 10Clarify And Connect 4 12 10
Clarify And Connect 4 12 10
 
Agile Estimation And Planning Part I
Agile Estimation And Planning Part IAgile Estimation And Planning Part I
Agile Estimation And Planning Part I
 
Planificación del proyecto estimación
Planificación del proyecto   estimaciónPlanificación del proyecto   estimación
Planificación del proyecto estimación
 
Blending Methods To Succeed Comparing Prince2 S Agility With Scrum Within The...
Blending Methods To Succeed Comparing Prince2 S Agility With Scrum Within The...Blending Methods To Succeed Comparing Prince2 S Agility With Scrum Within The...
Blending Methods To Succeed Comparing Prince2 S Agility With Scrum Within The...
 
User
UserUser
User
 
It's the organisation, Stupid
It's the organisation, StupidIt's the organisation, Stupid
It's the organisation, Stupid
 
ARCTEK2012SZornetzer-1
ARCTEK2012SZornetzer-1ARCTEK2012SZornetzer-1
ARCTEK2012SZornetzer-1
 
Introduction to Agile and Scrum (Montana Programmers Meetup Jan 2012).pptx
Introduction to Agile and Scrum (Montana Programmers Meetup Jan 2012).pptxIntroduction to Agile and Scrum (Montana Programmers Meetup Jan 2012).pptx
Introduction to Agile and Scrum (Montana Programmers Meetup Jan 2012).pptx
 
Session 1
Session 1Session 1
Session 1
 
Build Your NGO: Monitoring & Evaluation
Build Your NGO: Monitoring & Evaluation Build Your NGO: Monitoring & Evaluation
Build Your NGO: Monitoring & Evaluation
 
NG BB 43 Standardized Work
NG BB 43 Standardized WorkNG BB 43 Standardized Work
NG BB 43 Standardized Work
 
uTest and Crowdsortium Webinar: Scope & Briefs
uTest and Crowdsortium Webinar: Scope & BriefsuTest and Crowdsortium Webinar: Scope & Briefs
uTest and Crowdsortium Webinar: Scope & Briefs
 
OCC Group Capability Presentation Linkedin 2010
OCC Group Capability Presentation   Linkedin 2010OCC Group Capability Presentation   Linkedin 2010
OCC Group Capability Presentation Linkedin 2010
 
Key Considerations for a Successful Hyperion Planning Implementation
Key Considerations for a Successful Hyperion Planning ImplementationKey Considerations for a Successful Hyperion Planning Implementation
Key Considerations for a Successful Hyperion Planning Implementation
 
Customer Development Past Present Future Steve Blank 111909
Customer Development Past Present Future Steve Blank 111909Customer Development Past Present Future Steve Blank 111909
Customer Development Past Present Future Steve Blank 111909
 
Eileen may webcast
Eileen may webcastEileen may webcast
Eileen may webcast
 
NG BB 04 DEFINE Roadmap
NG BB 04 DEFINE RoadmapNG BB 04 DEFINE Roadmap
NG BB 04 DEFINE Roadmap
 
Requirements Management
Requirements ManagementRequirements Management
Requirements Management
 
¿Cómo gestionar un proyecto aspectos clave
¿Cómo gestionar un proyecto   aspectos clave¿Cómo gestionar un proyecto   aspectos clave
¿Cómo gestionar un proyecto aspectos clave
 

Mehr von NASAPMC

Mulenburg jerry
Mulenburg jerryMulenburg jerry
Mulenburg jerryNASAPMC
 
Mitskevich amanda
Mitskevich amandaMitskevich amanda
Mitskevich amandaNASAPMC
 
Martt anne
Martt anneMartt anne
Martt anneNASAPMC
 
Manzer fred
Manzer fredManzer fred
Manzer fredNASAPMC
 
Manthos jeff
Manthos jeffManthos jeff
Manthos jeffNASAPMC
 
Krahula john
Krahula johnKrahula john
Krahula johnNASAPMC
 
Kroeger lin
Kroeger linKroeger lin
Kroeger linNASAPMC
 
Krage jousha
Krage joushaKrage jousha
Krage joushaNASAPMC
 
Kotnour tim
Kotnour timKotnour tim
Kotnour timNASAPMC
 

Mehr von NASAPMC (9)

Mulenburg jerry
Mulenburg jerryMulenburg jerry
Mulenburg jerry
 
Mitskevich amanda
Mitskevich amandaMitskevich amanda
Mitskevich amanda
 
Martt anne
Martt anneMartt anne
Martt anne
 
Manzer fred
Manzer fredManzer fred
Manzer fred
 
Manthos jeff
Manthos jeffManthos jeff
Manthos jeff
 
Krahula john
Krahula johnKrahula john
Krahula john
 
Kroeger lin
Kroeger linKroeger lin
Kroeger lin
 
Krage jousha
Krage joushaKrage jousha
Krage jousha
 
Kotnour tim
Kotnour timKotnour tim
Kotnour tim
 

Kürzlich hochgeladen

Vertex AI Gemini Prompt Engineering Tips
Vertex AI Gemini Prompt Engineering TipsVertex AI Gemini Prompt Engineering Tips
Vertex AI Gemini Prompt Engineering TipsMiki Katsuragi
 
What's New in Teams Calling, Meetings and Devices March 2024
What's New in Teams Calling, Meetings and Devices March 2024What's New in Teams Calling, Meetings and Devices March 2024
What's New in Teams Calling, Meetings and Devices March 2024Stephanie Beckett
 
Leverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage Cost
Leverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage CostLeverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage Cost
Leverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage CostZilliz
 
Are Multi-Cloud and Serverless Good or Bad?
Are Multi-Cloud and Serverless Good or Bad?Are Multi-Cloud and Serverless Good or Bad?
Are Multi-Cloud and Serverless Good or Bad?Mattias Andersson
 
Bun (KitWorks Team Study 노별마루 발표 2024.4.22)
Bun (KitWorks Team Study 노별마루 발표 2024.4.22)Bun (KitWorks Team Study 노별마루 발표 2024.4.22)
Bun (KitWorks Team Study 노별마루 발표 2024.4.22)Wonjun Hwang
 
Anypoint Exchange: It’s Not Just a Repo!
Anypoint Exchange: It’s Not Just a Repo!Anypoint Exchange: It’s Not Just a Repo!
Anypoint Exchange: It’s Not Just a Repo!Manik S Magar
 
DevoxxFR 2024 Reproducible Builds with Apache Maven
DevoxxFR 2024 Reproducible Builds with Apache MavenDevoxxFR 2024 Reproducible Builds with Apache Maven
DevoxxFR 2024 Reproducible Builds with Apache MavenHervé Boutemy
 
Advanced Test Driven-Development @ php[tek] 2024
Advanced Test Driven-Development @ php[tek] 2024Advanced Test Driven-Development @ php[tek] 2024
Advanced Test Driven-Development @ php[tek] 2024Scott Keck-Warren
 
Integration and Automation in Practice: CI/CD in Mule Integration and Automat...
Integration and Automation in Practice: CI/CD in Mule Integration and Automat...Integration and Automation in Practice: CI/CD in Mule Integration and Automat...
Integration and Automation in Practice: CI/CD in Mule Integration and Automat...Patryk Bandurski
 
Streamlining Python Development: A Guide to a Modern Project Setup
Streamlining Python Development: A Guide to a Modern Project SetupStreamlining Python Development: A Guide to a Modern Project Setup
Streamlining Python Development: A Guide to a Modern Project SetupFlorian Wilhelm
 
Unleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubUnleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubKalema Edgar
 
Nell’iperspazio con Rocket: il Framework Web di Rust!
Nell’iperspazio con Rocket: il Framework Web di Rust!Nell’iperspazio con Rocket: il Framework Web di Rust!
Nell’iperspazio con Rocket: il Framework Web di Rust!Commit University
 
Human Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR SystemsHuman Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR SystemsMark Billinghurst
 
Search Engine Optimization SEO PDF for 2024.pdf
Search Engine Optimization SEO PDF for 2024.pdfSearch Engine Optimization SEO PDF for 2024.pdf
Search Engine Optimization SEO PDF for 2024.pdfRankYa
 
Vector Databases 101 - An introduction to the world of Vector Databases
Vector Databases 101 - An introduction to the world of Vector DatabasesVector Databases 101 - An introduction to the world of Vector Databases
Vector Databases 101 - An introduction to the world of Vector DatabasesZilliz
 
CloudStudio User manual (basic edition):
CloudStudio User manual (basic edition):CloudStudio User manual (basic edition):
CloudStudio User manual (basic edition):comworks
 
Story boards and shot lists for my a level piece
Story boards and shot lists for my a level pieceStory boards and shot lists for my a level piece
Story boards and shot lists for my a level piececharlottematthew16
 
Unraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdfUnraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdfAlex Barbosa Coqueiro
 
SAP Build Work Zone - Overview L2-L3.pptx
SAP Build Work Zone - Overview L2-L3.pptxSAP Build Work Zone - Overview L2-L3.pptx
SAP Build Work Zone - Overview L2-L3.pptxNavinnSomaal
 
Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 365Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 3652toLead Limited
 

Kürzlich hochgeladen (20)

Vertex AI Gemini Prompt Engineering Tips
Vertex AI Gemini Prompt Engineering TipsVertex AI Gemini Prompt Engineering Tips
Vertex AI Gemini Prompt Engineering Tips
 
What's New in Teams Calling, Meetings and Devices March 2024
What's New in Teams Calling, Meetings and Devices March 2024What's New in Teams Calling, Meetings and Devices March 2024
What's New in Teams Calling, Meetings and Devices March 2024
 
Leverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage Cost
Leverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage CostLeverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage Cost
Leverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage Cost
 
Are Multi-Cloud and Serverless Good or Bad?
Are Multi-Cloud and Serverless Good or Bad?Are Multi-Cloud and Serverless Good or Bad?
Are Multi-Cloud and Serverless Good or Bad?
 
Bun (KitWorks Team Study 노별마루 발표 2024.4.22)
Bun (KitWorks Team Study 노별마루 발표 2024.4.22)Bun (KitWorks Team Study 노별마루 발표 2024.4.22)
Bun (KitWorks Team Study 노별마루 발표 2024.4.22)
 
Anypoint Exchange: It’s Not Just a Repo!
Anypoint Exchange: It’s Not Just a Repo!Anypoint Exchange: It’s Not Just a Repo!
Anypoint Exchange: It’s Not Just a Repo!
 
DevoxxFR 2024 Reproducible Builds with Apache Maven
DevoxxFR 2024 Reproducible Builds with Apache MavenDevoxxFR 2024 Reproducible Builds with Apache Maven
DevoxxFR 2024 Reproducible Builds with Apache Maven
 
Advanced Test Driven-Development @ php[tek] 2024
Advanced Test Driven-Development @ php[tek] 2024Advanced Test Driven-Development @ php[tek] 2024
Advanced Test Driven-Development @ php[tek] 2024
 
Integration and Automation in Practice: CI/CD in Mule Integration and Automat...
Integration and Automation in Practice: CI/CD in Mule Integration and Automat...Integration and Automation in Practice: CI/CD in Mule Integration and Automat...
Integration and Automation in Practice: CI/CD in Mule Integration and Automat...
 
Streamlining Python Development: A Guide to a Modern Project Setup
Streamlining Python Development: A Guide to a Modern Project SetupStreamlining Python Development: A Guide to a Modern Project Setup
Streamlining Python Development: A Guide to a Modern Project Setup
 
Unleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubUnleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding Club
 
Nell’iperspazio con Rocket: il Framework Web di Rust!
Nell’iperspazio con Rocket: il Framework Web di Rust!Nell’iperspazio con Rocket: il Framework Web di Rust!
Nell’iperspazio con Rocket: il Framework Web di Rust!
 
Human Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR SystemsHuman Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR Systems
 
Search Engine Optimization SEO PDF for 2024.pdf
Search Engine Optimization SEO PDF for 2024.pdfSearch Engine Optimization SEO PDF for 2024.pdf
Search Engine Optimization SEO PDF for 2024.pdf
 
Vector Databases 101 - An introduction to the world of Vector Databases
Vector Databases 101 - An introduction to the world of Vector DatabasesVector Databases 101 - An introduction to the world of Vector Databases
Vector Databases 101 - An introduction to the world of Vector Databases
 
CloudStudio User manual (basic edition):
CloudStudio User manual (basic edition):CloudStudio User manual (basic edition):
CloudStudio User manual (basic edition):
 
Story boards and shot lists for my a level piece
Story boards and shot lists for my a level pieceStory boards and shot lists for my a level piece
Story boards and shot lists for my a level piece
 
Unraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdfUnraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdf
 
SAP Build Work Zone - Overview L2-L3.pptx
SAP Build Work Zone - Overview L2-L3.pptxSAP Build Work Zone - Overview L2-L3.pptx
SAP Build Work Zone - Overview L2-L3.pptx
 
Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 365Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 365
 

Rackley mike

  • 1. Project Management Challenge 2007 Track: Case Study Insights at NASA Case Study: Extracting Trends and Lessons Learned From Project Problems and Solutions NASA Goddard Space Flight Center February 6, 2007 Mike Rackley Catherine Traffanstedt NASA GSFC Code 170 NASA GSFC – SGT, Inc 301-614-7058 301-925-1125 Michael.W.Rackley@nasa.gov ctraffanstedt@sgt-mets.com
  • 2. Presentation Outline The Vision What are we trying to achieve? Implementation How do we achieve the Vision? Current Status Where are we today? Gap Analysis How do we close the gap? 2
  • 3. The Anomaly Management Vision The Vision What are we trying to achieve? Implementation How do we achieve the Vision? Current Status Where are we today? Gap Analysis How do we close the gap? 3
  • 4. Anomaly Management Vision • GSFC has well documented requirements, procedures, and best practices for mission-related problem management, reporting, analysis, and trending • GSFC efficiently verifies adherence to the established problem management requirements and established best practices • All GSFC development and on-orbit missions use a consistent problem management process and philosophy – Flexible enough to handle the diversity of spacecraft missions in the organization, enforcing certain fundamental “rules” for all missions, but allowing some differences where appropriate in other areas • Example: HST vs. Small Explorer Program • Recognizes that risks and budgets can be very different among programs – Applies to all missions that the organization is responsible for, regardless of whether or not the missions are built and/or flown “in- house” • NASA responsible for mission success regardless of approach to development or operations (i.e., in or out of house) • Out-of-house missions do pose unique challenges 4
  • 5. Anomaly Management Vision (cont) • GSFC provides online access to accurate mission configuration information for all Goddard managed missions in development and operations – Example: Subsystems, components, orbit configuration, duty cycles • GSFC analyzes and trends problems across development and on- orbit missions – Individual missions, across mission “families”, across all missions – Consistent use of a variety of Categorizations (e.g., failure type, root cause, impact) • GSFC extracts appropriate “Lessons Learned” information from the problem and solution space and apply these back into the development and operations processes – Examples: improvements in Rules/Best Practices for development, testing, and operations; planned vs. actual hardware & software performance 5
  • 6. Problem Management Information Flow Mission 1 . . . Mission N • In-House Development • Out-of-House Procurement Mission Anomaly Information Impediments: • Project focus Mission • Indirect (export) • Heterogeneous Configuration • Direct missions • Record Definitions anomaly – Design information – Timeframe Centralized Centralized • Provide – Orbit Examples: – Ops/science Problem Problem search/query • HW & SW profile descriptions Information System Information System capability • Contracts • Design life info • Generate reports – End-item • Failure – Proprietary handling • Trending • Support data approach (e.g., • Analysis individual and • Culture single vs. dual multi-mission • Data Mining “views” • Sustaining over string) time Cross- Multi-mission Ops/ Lessons Output Failure mission Summary Engineering Learned Analysis Trending & Products Reports Research Extraction Analysis 6
  • 7. Lessons Learned Feedback Process Knowledge Management Ops process or procedure proves to be Feedback / successful over time recommendations provided to Lessons spacecraft flight Ops experiences actual projects behavior of spacecraft Learned established Best practices Anomaly resolved with established defined corrective action and documented Anomaly occurs Influencing Variables • Spacecraft design / technology • Risk tolerance • Personnel changeover • Ground system design / technology • Organizational culture • Mission models/profiles 7
  • 8. Implementing The Vision The Vision What are we trying to achieve? Implementation How do we achieve the Vision? Current Status Where are we today? Gap Analysis How do we close the gap? 8
  • 9. Implementation Approach • Ensure appropriate requirements are officially documented so that they apply to all missions in development and operations – Goddard Procedures and Requirements Documents (GPR’s) – Mission Assurance Requirements (MAR) Document (per mission), with the Mission Assurance Guideline (MAG) document as a template • Establish and document rules/best practices for how to document anomalies – Main goal is to achieve the needed consistency and robustness across missions – Anomaly Management Best Practices Handbook (or comparable) • Utilize a centralized, independent operations mission assurance team to verify adherence across missions to requirements and best practices – Provides ability to recognize and correct problems early 9
  • 10. Implementation Approach (cont) • Utilize a centralized anomaly trending and analysis team to perform various trending and analysis functions within and across missions – Assist in development of Mission Configurations – Analyze trends – Extract “lessons learned” and work to apply them across the relevant organizations, documents, etc. • Implement a “Lessons Learned CCB” with broad participation to disposition potential/proposed lessons learned and best practices • Utilize a centralized database to capture problem report data that provides ability: – For individual missions to document, track, and close out their problems – For the Center to perform various analysis and trend functions across missions – To handle development/test (pre-launch) and operations phases – To handle flight and ground problems – To bring in data from outside users – To provide broad access, but with appropriate access protections (for ITAR sensitive and proprietary data) 10
  • 11. Implementation Approach (cont) • The centralized database is the Goddard Problem Reporting System (GPRS), which contains two modules for documenting and managing ground and flight segment problems – Both provide multi-mission, web-based, secure user access – Other features include Mission Configuration definition, search/query capabilities, user notification (via email), user privilege control, user problem reporting, and attachments • SOARS -- Spacecraft Orbital Anomaly Reporting System – Used to document flight and ground problems during operations – Problems categorized via a common approach, using various Categorization Fields (Anomaly Classification, Impacts, Root Cause) • PR/PFR -- Problem Reporting/Problem Failure Reporting System – Used to capture problems encountered during the development, integration, and test – Similar to SOARS, but tailored to the I&T environment (e.g., the signature process) 11
  • 12. SOARS Model Maximize Consistency & Commonality in how Support Mission Anomaly Anomalies are Documented & Categorized Management (document, track, report, notify) Document Mission Configurations (mission-unique and common subsystems & components) Anomaly Event SOARS SOARS Occurs Perform Cross-Mission Trending & Analysis Support Independent Mission Assurance Activities (analysis, audits) Support Cross-Mission Analysis (have others had a similar Provide Common Approach to problem?) Maintenance, Security, etc. 12
  • 13. Current Status: Where Are We Today? The Vision What are we trying to achieve? Implementation How do we achieve the Vision? Current Status Where are we today? s lenge C hal Gap Analysis How do we close the gap? 13
  • 14. Current Status • Requirements & Best Practices Development/Compliance – High-level requirements associated with using GPRS and performing quality assurance, trending, and analysis are in place • GPR 1710.H - Corrective and Preventative Action • GPR 5340.2 - Control of Nonconformances – Mission Assurance Requirements Documents, with the Mission Assurance Guideline (MAG) document as a template also generally cover the requirements • But, capturing problem information in the central PR/PFR System limited to in-house missions for pre-launch phase – Operations Rules/Best Practices not yet formally documented in a central location that would apply across all missions • Exist in pockets within missions, families of missions, and Programs • Working with missions and Programs to begin development 14
  • 15. Current Status (cont) • Operational Improvements – Major GPRS enhancements made to SOARS and PR/PFR software over the last year to improve operational usability, etc. • SOARS now considered an operationally useful tool, with ongoing activities to continue incremental improvements • PR/PFR has more recently undergone improvement, largely driven by two Goddard in-house missions (SDO and LRO) – PR/PFR needs much more improvement to “catch-up” to SOARS – SOARS use by the operational missions has increased dramatically over the last year • One year ago very few missions were using SOARS at Goddard • Almost all in-house missions (~20) now using SOARS effectively, including RXTE, Terra, Aqua, EO-1, and TDRSS • Only one out-of-house mission currently populating SOARS with anomaly report data (Swift at Penn State University, as a direct user) – Other out-of-house “operations vendors” include APL, Orbital, LASP, and University of California at Berkeley (UCB) – Currently working with these organizations to develop plan for populating SOARS (e.g., make use of data import) 15
  • 16. Current Status (cont) • Anomaly Analysis and Trending – Centralized teams in place within Office of Mission Assurance (Code 300) for performing roles of operations mission assurance and cross- mission anomaly analysis and trending • Development phase handled by individual Code 300 System Assurance Managers (SAM’s) assigned to missions – GSFC (Code 300) using SOARS data extraction to: • Provide monthly Center-wide snapshot of mission anomalies, characterized by Mission Impact, Operations Impact, and Root Cause • Develop annual “year-in-review” anomaly summary report (OAGS) • Trend and analyze problems across missions, subsystems, failure types, etc. • Support long term reliability studies on flight hardware • Identify missions that are not adequately documenting anomalies and work with them to improve 16
  • 17. SOARS Data Extraction – Example 1 SOAR Records - All active Goddard Missions Root Cause Categorization (12 month summary) for Cum CY06 "Ops Workaround" SOAR's 50 Ground Segment SOARs 45 Flight Segment SOARs Design, 8 40 35 Workmanship, 4 30 25 Unknown, 28 Process, 4 20 Environmental, 2 15 10 Degradation, 3 5 0 6 6 06 06 06 06 06 06 6 06 6 6 r-0 -0 l- 0 -0 -0 n- n- g- p- v- c- Hardware Failure, 18 b- ct ay ar Ju Ap No De Au Se Ja Ju Fe O M M Human Error, 3 Training, 1 Significant Nov’06 On-orbit Anomalies Root Cause Categorization • Msn-X power load shed anomaly for Cum CY06 "Data Loss" SOAR's – All science instrument main voltages shed in accordance to the Design, 3 Under Voltage protection Circuitry Workmanship, 2 – Under voltage condition, possibly indicating a failure (short) Process, 0 somewhere in the affected loads Environmental, 2 Degradation, 2 • Msn-Y Star Tracker and Sun Sensor Boresight anomaly – Star Tracker alignment calibration analysis indicates over a 2 arc- minute shift in the alignment between the Star Tracker and FSS Hardware Failure, 7 • Msn-Z ACS Quaternion Spike Error anomaly – Star Tracker Quaternion element was corrupted Training, 0 – Spacecraft commanded to Sun/Mag pointing and the Star Tracker Unknown, 26 Human Error, 2 stabilized
  • 18. SOARS Data Extraction – Example 2 Root Cause Categorization CY06 Cumulative Root Cause Categorizations 1 Design Related 0.9 Hardware Degradation/Failure Design Related, 16% 0.8 Operations Related Hardware 0.7 0.6 Environmentally Related Degradation/Failure, 20% 0.5 Unknown 0.4 0.3 Operations Related, 4% 0.2 Unknown, 50% 0.1 0 Environmentally Related, 6 06 06 6 06 6 6 06 06 6 6 6 l -0 -0 v-0 r- 0 -0 c-0 t- 0 10% n- n- b- g- p- ar ay Ju Ap Oc No Ja Ju Fe Au Se De M M Anomaly Classification Categorization CY06 Cumulative Anomaly Classification Categorization 100% Hardware Other, 5% 90% 80% Software 70% Operations Operations, 13% Other 60% Hardware, 41% 50% 40% 30% 20% 10% 0% 06 06 6 6 6 06 6 6 06 6 6 06 l- 0 -0 -0 -0 0 r-0 0 Software, 41% n- - n- b- g- p- - ov ar ct ay ec Ju Ap Ju Ja Au Fe Se O M N M D Operations Impact Categorization CY06 Cumulative Mission Impact Categorization 100% Mission Data Loss 90% Permanent Loss of Mission Data Loss, Operational Workaround 80% Capacity, 0% 28% 70% No Impact 60% Permanent Loss of Capacity 50% No Impact, 42% 40% 30% 20% 10% 0% Operational 06 06 6 6 6 6 06 6 06 6 6 06 l- 0 -0 -0 -0 0 r-0 0 n- - n- b- g- p- - ov ar ct ay Workaround, 30% ec Ju Ap Ju Ja Au Fe Se O M N M D
  • 19. Current Status (cont) • Inserting “Lessons Learned” into the Process – Implementing an “Operations Lessons Learned” process, but still in the very early stages – Plan is to implement a process where a Lessons Learned Committee/Board is presented with potential lessons • Committee determines if lesson valid, and if so, the actions that are appropriate to take it forward • Working to engage broad participation (mission assurance, missions, and engineering) Lesson extracted Lesson vetted Action from Operations by taken on LL event, activity, etc. LL Committee Examples: • Ops best practice update • Improvement in testing • Feedback on hardware performance issues • Recommendation to flight software group • Recommendation for “Gold Rules” update 19
  • 20. Sample Lessons Learned Summary Anomaly Anomaly Proposed Mission Date Action Description Lesson Learned Msn_X 10/3/06 Spacecraft clock Establish best practice of Work directly with Ops rollover to an illegal evaluating registers prior to Projects to document 1998 date; on-board launch, and as part of the as Ops Best Practice. clock register capacity on-orbit ops procedures. Work with Flight was exceeded. Include FSW capabilities to Software Branch to issue warnings when determine if design appropriate. best practice needed. Msn_Y 10/19/06 UV instrument Lesson tbd, but would apply Investigate further. experiencing channel to improving test approach. Discuss with Msn_Y noise, apparently Seems this should have SAM and I&T Manager. associated with heater been caught in instrument Generate independent activation and or observatory I&T. Why analysis report if environmental was this not seen during warranted. conditions. Causing thermal vac? Does this failure of instrument to have implications to other meet signal to noise missions? Could or should ratio requirements for this instrument sensitivity the data channel. have been better tested during instrument I&T?
  • 21. Implementation Challenges • Achieving consistency across missions in how anomalies are documented (e.g., terminology, level of detail) and categorized is proving to be very difficult – Many missions with different ops teams have different taxonomies, operations cultures, etc. – Missions themselves also very different in terms of bus design, payload suite, operations profile, age, and risk tolerance • Mission Configuration definitions in SOARS not well-defined across all missions, making it difficult to perform cross-mission analyses – Not all information is readily available to use in SOARS definitions, especially for older missions (e.g., component manufacturers and design life) – Missions have struggled to put in the extra time to populate SOARS with their Mission Configurations – Working with the existing missions to improve the definitions, but effort is time consuming and challenging – Working with upcoming missions to get the definitions in place prior to launch (e.g., GLAST, SDO, and LRO) 21
  • 22. Implementation Challenges (cont) • Proving very difficult to get out-of-house operations centers/teams to document problems in SOARS – Only one direct user so far (Swift/PSU) – Working with others to take advantage of enhanced SOARS data import capability – Finding that the critical factor is getting the agreement more explicitly planned well before launch • Only getting pre-launch problem report data into PR/PFR for in- house missions – Leaves a big “information gap” given the high number of out-of-house missions – Vendors cannot use PR/PFR directly, but working to determine if certain types of problem report data can be imported into PR/PFR 22
  • 23. Gap Analysis The Vision What are we trying to achieve? Implementation How do we achieve the Vision? Current Status Where are we today? Gap Analysis How do we close the gap? 23
  • 24. Gap Analysis: How Would You Complete “The Job”? 1. How would you approach achieving and maintaining more consistency across missions in how anomalies are documented? – Getting them to do it consistently over time – Working to measure/enforce consistency over time 2. Do you think it is possible to develop a set of Operations Anomaly Management Best Practices that would reasonably apply across a set of heterogeneous missions, that are a mix of in-house and out-of- house operations approaches? – If so, how would you approach getting this defined and keeping it updated? – If not, what can be done? Do you just give up? 3. What type of information besides the basic anomaly information would you want to capture for your missions to support long term trending, reliability/failure analysis, etc? 24
  • 25. Gap Analysis: How Would You Complete “The Job”? 4. How would you ensure that the relevant Mission Configuration information for upcoming missions is documented so that it is usable by the problem management system? – Address the in-house vs. out-of-house challenges 5. How would you approach getting more out-of-house operations centers/teams to document their anomalies in the centralized problem information system (SOARS)? 6. How do you extract problem information from out-of-house mission vendors for use in the centralized problem information system (PR/PFR) during the pre-launch development phase? – Will likely have proprietary data issues 7. How do you integrate lessons learned into projects/programs so that they will actually affect mission development and operations, and maintain benefit over time? 25
  • 28. SOARS Anomaly Report Form (cont) 28
  • 29. SOARS Anomaly Report Form Observation area includes all problem descriptions including orbit and attitude data, and other user defined fields Investigation area provides a place for the operations team to document the problem investigation and status. Includes an investigative log, subsystem and component identification, anomaly criticality, status, and responsibility assignment. Resolution area provides the operations team an text area to capture a synopsis of the problem cause and resolution. Assessment area is used to identify follow-on activities or recommendations and categorize the SOAR for generic use. Notification area is used to send email notifications to personnel (either other SOARS users or other contacts identified by missions) 29
  • 30. SOARS Data Summary • All SOARS anomaly reports will include the following information: – Anomaly occurrence time – Anomaly title – Anomaly category (flight or ground segment specification) – Anomaly Subsystem and Component identification – Anomaly description – Closure information – Follow-on recommendations – Anomaly categorization • SOARS may include the following types of information: – Anomaly assignment and responsible party – Investigation information – Orbital information – Mission unique phase and mode – Additional file attachments as appropriate 30
  • 31. SOAR Data Categorization • SOAR Categorization Fields are the key fields used for trending and analysis across missions for the Center. The fields include: – Mission Impact: describes the actual anomaly impact on the ability to meet Science and/or Mission Goals or requirements • No Effect, Minor, Substantial, Major, Catastrophic, Undetermined – Impact Severity on the Anomalous Item: describes the actual impact to the subsystem or component where the anomaly occurred • No Effect, Minor, Major, Catastrophic, Undetermined, Not applicable – Operations Impact: describes the effect of the anomaly on mission and/or science operations • Data Loss, Service Loss, Mission Degradation/Loss, Undetermined, Ops Workaround, None – Anomaly Classification: describes where within the mission systems the failure occurred • Hardware, Software, Operations, Other – Anomaly Root Cause: describes the root cause of the anomaly. • Design, Workmanship/Implementation, Process, Environmental, Degradation, Hardware Failure, Training, Human Error, Unknown 31
  • 32. One Size Fits All Anomaly Management (AM) System: Does It Exist? Mission Phases (S/C I&T, Observatory I&T, L&EO, Normal Ops, Disposal) • AM System must meet the needs of individual Missions missions and of the broader multiple mission perspective (e.g., looking for cross mission trends) • Mission may want one discrepancy/anomaly management system through all mission phases, but the phases are very different environments (pre and post-launch) • AM System may have to handle spacecraft and ground anomalies • AM System inevitably must be able to interface with other AM Systems used by individual missions 32