1 / 26

7/22/04 Report Back: Performance Analysis Track

7/22/04 Report Back: Performance Analysis Track. Dr. Carol Smidts Wes Deadrick. Track Members. Carol Smidts (UMD) – Track Chair Integrating Software into PRA Ted Bennett and Paul Wennberg (Triakis) Empirical Assurance of Embedded Software Using Realistic Simulated Failure Modes

jeb
Download Presentation

7/22/04 Report Back: Performance Analysis Track

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. 7/22/04 Report Back:Performance Analysis Track Dr. Carol SmidtsWes Deadrick

  2. Track Members • Carol Smidts (UMD) – Track Chair • Integrating Software into PRA • Ted Bennett and Paul Wennberg (Triakis) • Empirical Assurance of Embedded Software Using Realistic Simulated Failure Modes • Dolores Wallace (GSFC) • System and Software Reliability • Bojan Cukic (WVU) • Compositional Approach to Formal Models • Kalynnda Berens (GRC) • Software Safety Assurance of Programmable Logic • Injecting Faults for Software Error Evaluation of Flight Software • Hany Ammar (WVU) • Risk Assessment of Software Architectures

  3. Agenda • Characterization of the Field • Problem Statement • Benefits of Performance Analysis • Future Directions • Limitations • Technology Readiness Levels

  4. Characterization of Field • Goal: Prediction and Assessment of Software Risk/Assurance Level (Mitigation optimization) • System Characteristics of interest • Risk (Off-nominal situations) • Reliability, availability, maintainability = Dependability • Failures - general sense • Performance Analysis Techniques - modeling and simulation, data analysis, failure analysis, design analysis focused on criticality

  5. Problem Statement • Why should NASA do performance analysis? - We care if things fail! • Successfully conducting SW and System Performance Analysis gives us the data necessary to make informed decisions in order to improve performance and overall quality • Performance analysis permits: • Ability to determine if/when system meets requirements • Risk reduction and quantification • Application of new knowledge to future systems • A better understanding of the processes by which systems are developed and therefore enables NASA to exercise continual improvement

  6. Benefits of Performance Analysis • Reduced development and operating costs • Manage and optimize current processes thereby resulting in more efficient and effective processes • Defined and repeatable process – reduced time to do same volume of work • Reduces risk and increases safety and reliability • Better software architecture designs • More maintainable systems • Enable NASA to handle more complex systems in the future • Put the responsibility where it belongs from a organizational perspective - focuses accountability

  7. Future Directions for Performance Analysis • Automation of modeling and data collection – increased efficiency and accuracy • A more useful, better reliability model • useful = user friendly (enable the masses not just the domain experts), increased usability of the data (learn more from what we have) • better = greater accuracy and predictability • Define and follow repeatable methods/processes for data collection and analysis including: • education and training • use of simulation • gold nugget = accurate and complete data

  8. Future Directions for Performance Analysis (Cont.) • Develop a method for establishing accurate performance predictions earlier in life cycle • Evolve to refine system level assessment • factor in the human element • Establish and define an approach to performing trade-off of attributes – reliability, etc. • Need for early guidance on criticality of components • Optimize a defect removal model • Methods and metrics for calculating/defending return on investment of conducting performance analysis

  9. Why not • Standard traps - Obstacles • Uncertainty about scalability • User friendliness • Lack of generality • “Not invented here” syndrome • Costs and benefits • Difficult to assess and quantify • Long term project benefit tracking recommended

  10. Technology Readiness Level • Integrating Software into PRA – Taxonomy (7) • Test-Based Approach for Integrating SW in PRA (3) • Empirical Assurance of Embedded Software Using Realistic Simulated Failure Modes (5) • Maintaining system and SW test consistency (8) • System Reliability (3) • Software Reliability (9) • Compositional Approach to Formal Models (2) • Software Safety Assurance of Programmable Logic (2) • Injecting Faults for Software Error Evaluation of Flight Software (9) • Risk Assessment of Software Architectures (5)

  11. Research Project Summaries

  12. Integrating Software Into PRADr. Carol Smidts, Bin Li Objective: • PRA is a methodology to assess the risk of large technological systems • The objective of this research is to extend current classical PRA methodology to account for the impact of software onto mission risk

  13. Integrating Software Into PRA (Cont) Achievements • Developed a software related failure mode taxonomy • Validated the taxonomy on multiple projects (ISS, Space Shuttle, X38) • Proposed a step-by-step approach to integration in the classical PRA framework with quantification of input and functional failures.

  14. TRIAKIS Corporation Analyze/Test/V&V Analyze/Test/Verify Requirements Build Model,Simulate,Prototype,ES, etc. SYSTEM SW Interpretation Integration Testing Design/Debug Design/Debug Problem Most embedded SW faults found at integ. test traceable to Rqmts. & interface misunderstanding Disconnect exists between System and software development loops

  15. TRIAKIS Corporation Approach • Develop & simulate entire system design using executable specifications (ES) • Verify total system design with suite of tests • Simulate controller hardware • Replace controller ES with simulated HW running object (flight) software • Test SW using system verification tests When SW passes all system verification tests, it has correctly implemented all of the tested requirements

  16. Problem: FMEA Limitations Expensive & time-consuming List of possible failure modes extensive Focuses on prioritized subset of failure modes Approach: Test SW w/sim’d Failures Create pure virtual simulation of Mini-AERCam HW & flight environment running on PC Induce realistic component/subsystem failures Observe flight SW response to induced failures TRIAKIS Corporation IV&V Facility Mini-AERCam Empirical Assurance of Embedded SWUsing Realistic Simulated Failure Modes • Can we improve coverage by testing SW resp. to sim’d failures? • Compare results with project-sponsored FMEA, FTA, etc.: #Failure modes evaluated? #Issues uncovered? Effort involved?

  17. Software and System ReliabilityDolores Wallace, Bill Farr, Swapna Gokhale • Addresses the need to evaluate and assess the reliability and availability of large complex software intensive systems by predicting (with associated confidence intervals): • The number of software/system faults, • Mean time to failure and restore/repair, • Availability, • Estimated release time from testing.

  18. 2003 & 2004 Research 2003 (Software Based) • Literature search completed • New models were selected: 1) Enhanced Schneidewind (includes risk assessment and trade-off analysis) and 2) Hypergeometric Model • Incorporated the new software models into the established public domain tool SMERFS^3 • Applied the new models on a Goddard software project • Made the latest version of SMERFS^3 available to the general public 2004 (System Based) • Conducted similar research effort for System Reliability and Availability • Will enhance SMERFS^3 and validate the system models on a Goddard data set

  19. A Compositional approach to Validation of Formal Models Dejan Desovski, Bojan Cukic • Problem • Significant number of faults in real systems can be traced back to specifications. • Current methodologies of specification assurance have problems: • Theorem Proving: Complex • Model Checking: State explosion problems • Testing: Incomplete. • Approach • Combine them! • Use test coverage to build abstractions. • Abstractions reduce the size of the state space for model checking. • Develop visual interfaces to improve the usability of the method.

  20. Identify Interfaces and Critical Sections Obtain Source Code and Documentation Error/Fault Research Start Sufficient time and funds? Estimate Effort Required Select Subset Importance Analysis Yes Feedback to FCF Project Fault Injection Testing Test Case Generation End Document Results, Metrics, Lessons Learned Software Fault Injection ProcessKalynnda Berens, Dr. John Crigler, Richard Plastow • Standardized approach to test systems with COTS and hardware interfaces • Provides a roadmap of where to look to determine what to test

  21. Programmable Logic at NASAKalynnda Berens, Jacqueline Somos • Issues • Lack of good assurance of PLCs and PLDs • Increasing complexity = increasing problems • Usage and Assurance Survey - SA involved in less than 1/3 of the projects; limited knowledge • Recommendations • Trained SA for PLCs • PLDs – determine what is complex; use process assurance (SA or QA) • Training Created • Basic PLC and PLD training aimed at SA • Process assurance for hardware QA

  22. Year 2 of Research • What is industry and other government agencies doing for assurance and verification? • An intensive literature search of white papers, manuals, standards, and other documents that illustrated what various organizations were doing. • Focused interviews with industry practitioners. Interviews were conducted with assurance personnel (both hardware and software) and engineering practitioners in various industries, including biomedical, aerospace, and control systems. • Meeting with FAA representatives. Discussions with FAA representatives lead to a more thorough understanding of their approach and the pitfalls they have encountered along the way. • Position paper, with recommendations for NASA Code Q

  23. Current Effort • Implement some of the recommendations • Develop coursework to educate software and hardware assurance engineers • Three courses • PLCs for Software Assurance personnel • PLDs for Software Assurance personnel • Process Assurance for Hardware QA • Guidebook • Other recommendations • For Code Q to implement if desired • Follow-up CSIP to try software-style assurance on complex electronics

  24. Severity Analysis MethodologyHanny Ammar, Katerina Goseva-Popstojanova, Ajith Guedem, Kalaivani Appukutty, Walib AbdelMoez, and Ahmad Hassan • We have developed a methodology to assess severity of failures of components, connectors, and scenarios based on UML models • This methodology is applied on NASA’s Earth Observing System (EOS)

  25. Requirement Risk Analysis Methodology • We have developed a methodology for assessing requirements based risk using normalized dynamic complexity and severity of failures. This can be used in the DDP process developed at JPL. Risk factor of scenario S1 in Failure mode FM2 • According to Dr. Martin Feather’s DDP Process, “The Requirements matrix maps the impacts of each failure mode on each requirement.” • Requirements are mapped to UML use case and scenarios • A failure mode refers to the way in which a scenario fails to achieve its requirement

  26. What to Read • Key works in the field • Tutorials • Web sites • Will be completed at a later time

More Related