Presentation is loading. Please wait.

Presentation is loading. Please wait.

IV&V Facility 1 Software Reliability Corroboration Bojan Cukic, Erdogan Gunel, Harshinder Singh, Lan Guo West Virginia University Carol Smidts University.

Similar presentations


Presentation on theme: "IV&V Facility 1 Software Reliability Corroboration Bojan Cukic, Erdogan Gunel, Harshinder Singh, Lan Guo West Virginia University Carol Smidts University."— Presentation transcript:

1 IV&V Facility 1 Software Reliability Corroboration Bojan Cukic, Erdogan Gunel, Harshinder Singh, Lan Guo West Virginia University Carol Smidts University of Maryland (WVU UI: Integrating Formal Methods and Testing in a Quantitative Software Reliability Assessment Framework 2002)

2 IV&V Facility 2 Overview Introduction and motivation. Software reliability assessment and NASA IV&V. Bayesian hypothesis testing approach. A methodology for formulating priors. Case study Accounting for severities and risks. Summary

3 IV&V Facility 3 Introduction Improvement of software V&V practices, especially for high assurance systems. Quantification of the effects of V&V activities is always desirable. Is software reliability quantification practical for safety/mission critical systems? –Time and cost considerations may limit the appeal. Reliability growth applicable only to integration testing, the tail end of V&V. Estimation of operational usage profiles is rare.

4 IV&V Facility 4 Is SRE Impractical for NASA IV&V? Most IV&V techniques are qualitative in nature. Mature software reliability estimation methods based exclusively on operational (system) testing. Neglects the investment made in other IV&V techniques –Requirements readings, inspections, problem reports and tracking, unit level tests… Req Design Code Test (Verification & Validation) Unit IntegrationAcceptance Life cycle long IV&V Implementation Tradional SW Rel. Assessment

5 IV&V Facility 5 Regulatory Viewpoint Regulatory view: DO178-B (software considerations in airborne systems and equipment certification) “… methods for estimating the post-verification probabilities of software errors were examined. The goal was to develop numerical requirements for such probabilities for software in computer-based airborne systems of equipment. The conclusion reached, however, was that currently available methods do not provide results in which the confidence can be placed to the level required for this purpose... If the applicant proposes to use software reliability models for certification credit, rationale for the model should be included in the plan for software aspects of certification, and agreed with by the certification authority.”

6 IV&V Facility 6 Contribution Develop software reliability assessment methods that build on: –Stable and mature development environments. –Lifecycle long IV&V activities. –Utilize all relevant available information. –Qualitative (formal and informal) IV&V methods? Strengthening the case for IV&V all across NASA enterprise.

7 IV&V Facility 7 Assessment vs. Corroboration Current thinking –Software reliability “tested into” the product through the integration and acceptance testing. Our thinking –Why “waste” the results of all the qualitative IV&V activities. –Testing should corroborate that the life-cycle long IV&V techniques are giving the “usual” results, that the project follows usual quality patterns.

8 IV&V Facility 8 Reliability Assessment (No Prior Assumptions) P(    Required testing effort (N), from random sampling: Number of failure free test cases as a function of the required failure rate, with C=0.99 Value of  Number of Tests 10 -2 458 10 -3 4,602 10 -4 46,048 10 -5 460,514 10 -6 4,605,167 Required testing effort not realistic.

9 IV&V Facility 9 Bayesian Inference Allows for the inclusion of a subjective probability of failure. Subjective estimate based on observed behavior, reflects beliefs. Hypothesis on the event occurrence probability is combined with new evidence, which may change the degree of belief..

10 IV&V Facility 10 Bayesian Estimation (Non Ignorance Priors) Needs the following assumption: –The system has achieved desired reliability prior to acceptance testing. –This “guess” should be “reasonably accurate.” Use random tests (operational profile) to corroborate assumed system failure probability. How many failure free random tests U should be performed?

11 IV&V Facility 11 Benefits What if corroboration testing is not failure free? –Keep adjusting the target number of tests [Littlewood 97, 98]

12 IV&V Facility 12 Bayesian Hypothesis Testing (BHT) Problem of Bayesian estimation: –Categorical assumption that the program meets required reliability. BHT makes this a probability statement, P(H 0 ). Corroboration testing now looks for the evidence in favor of the hypothesized reliability. H o :   o alternative hypothesis.

13 IV&V Facility 13 The number of corroboration tests according to BHT theory.  P(H o ) n o n 1 n2 0.01 4574764970.0010.012378267129750.00010.01683110648145010.000010.01934933176636490.0000010.019752101273282007 0.010.12282582890.0010.1636101714020.00010.1853315761500.000010.18869646272810.0000010.189030067123725 0.010.490128167 0.0010.41384117390.00010.4146125132600.000010.41473889147240.0000010.41471222267468 0.010.65087126 0.0010.663269552 0.00010.6658272458 0.000010.6652584111730.0000010.665813951351

14 IV&V Facility 14 Formulating Priors Formulation of prior beliefs is the most important research issue. Historical data on failure occurrences under the same V&V regime. –Historical data on failure occurrence reduction related to the application of specific verification techniques (very few studies). –Process effectiveness measures [Smidts 98]. –Transforming fault density into failure intensity [Smidts 01]. –Represent the application of a specific verification method by an appropriate number of random tests [Miller et. al. 94].

15 IV&V Facility 15 Can This Be Done? Is it realistic to expect software developers to hypothesize on the operational reliability? Experiment (Smidts et. al.). –A panel of experts ranked 32 measures related to software reliability. Ranks normalized to a [0, 1] range. –Highly ranked measures: Failure rate (0.98), test coverage (0.90), fault density (0.73). –Low ranked measures: Mutation testing(0.48), function point analysis (0.00), bugs per line of code (Gaffney estimate, 0.00).

16 IV&V Facility 16 Controlled Experiment A company contracted to develop a program (smart card based access control system, PACS). –Controlled requirements document (NSA specs). Five software engineering measures monitored: –Defect density, test coverage, requirements traceability, function points, Gaffney. Each measure can used within a reliability prediction system (RPS). Accurate RPS. –Defect density, test coverage and requirements traceability. Inaccurate RPS: function points and Gaffney.

17 IV&V Facility 17 Software Reliability Corroboration Accurate predictors are adequate for the corroboration approach. –A weighted linear combination of the three measures (RPS) gives a very accurate reliability prediction. –Low levels of trust in the prediction accuracy. No experience in repeatability. Low value of P(H 0 ) still requires substantial but realistic reliability corroboration effort.

18 IV&V Facility 18 Accounting for Failure Severities Not all the failures encountered in corroboration testing are equally important. Instead of counting generic failures, test failures stratified according to their severity. If a high severity failure encountered in corroboration testing, strong evidence in favor of the alternative hypothesis. Allows for the tolerance towards low severity failures.

19 IV&V Facility 19 Approach Recap Software quality Measures (SQM) Reliability Prediction Systems (RPS) RPS Combination Techniques SW Reliability Corroboration Testing SQM1 SQM3 SQM2 SQM4 SQM6 SQM5 SQMi SQMj RPS1RPS2RPSkRPSm... RPS Combination (Experience, Learning, Dempster-Schafer…) BHT software reliability corroboration Null Hypothesis, H 0 Alternative Hypothesis, H a Software Development Lifecycle Software Reliability Measure

20 IV&V Facility 20 Status and Perspectives Software reliability corroboration allows: –Inclusion of IV&V quality measures and activities into the reliability assessment. –A significant reduction in the number of (corroboration) tests. –Software reliability of safety/mission critical systems can be assessed with a reasonable effort. Research directions. –Sound formulation of prior beliefs from IV&V. –Further experimentation (other measures, repetition). –Can prior beliefs be based on the “formality” of the IV&V methods (formal methods)?


Download ppt "IV&V Facility 1 Software Reliability Corroboration Bojan Cukic, Erdogan Gunel, Harshinder Singh, Lan Guo West Virginia University Carol Smidts University."

Similar presentations


Ads by Google