A Brief Essay on Software Testing
1 / 23

A Brief Essay on Software Testing Antonia Bertolino, Eda Marchetti Presented by Gargi Chipalkatti Software Engineering - PowerPoint PPT Presentation

  • Uploaded on

A Brief Essay on Software Testing Antonia Bertolino, Eda Marchetti Presented by Gargi Chipalkatti (Software Engineering II - EEL 6883). Testing. Testing refers to many different activities used to check a piece of software

I am the owner, or an agent authorized to act on behalf of the owner, of the copyrighted work described.
Download Presentation

PowerPoint Slideshow about 'A Brief Essay on Software Testing Antonia Bertolino, Eda Marchetti Presented by Gargi Chipalkatti Software Engineering ' - hagen

An Image/Link below is provided (as is) to download presentation

Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author.While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server.

- - - - - - - - - - - - - - - - - - - - - - - - - - E N D - - - - - - - - - - - - - - - - - - - - - - - - - -
Presentation Transcript
Slide1 l.jpg

A Brief Essay on Software Testing

Antonia Bertolino, Eda Marchetti

Presented by Gargi Chipalkatti

(Software Engineering II - EEL 6883)

Testing l.jpg


Testing refers to many different activities used to check a piece of software

Even after successful completion of an extensive testing, the software can still contain faults

Testing can never prove the absence of defects, it can only possibly reveal the presence of faults by provoking malfunctions

Definition testing l.jpg

Definition - Testing


Software Testing consists of the dynamic verification of the behavior of a program on a finite set of test cases, suitably selected from the usually infinite executions domain, against the specified expected behavior.


Dynamic – means that testing implies executing the program on (valued) inputs

Finite – means that only a limited number of test cases can be executed during the testing phase, chosen from the whole test set, that can generally be considered infinite

Selected – refers to the test techniques adopted for selecting the test cases (and testers must be aware that different selection criteria may yield vastly different effectiveness)

Expected – points out to the decision process adopted for establishing whether the observed outcomes of program execution are acceptable or not

Goals for testing l.jpg

Goals for Testing

Detection of “bugs” in the software

Increase confidence in the proper functioning of the software

Assist with the evaluation of functional and nonfunctional properties (as well as performance characteristics)

Exposing potential design flaws

Exposing deviations from user’s requirements

Measuring the operational reliability

Producing estimates of software reliability

Tasks associated with testing l.jpg

Tasks associated with Testing

Deriving an adequate suite of test cases, according to a feasible and cost-effective test selection technique

The ability to launch the selected tests (in a controlled host environment, or worse in the tight target environment of an embedded system)

Deciding whether the test outcome is acceptable or not (which is referred to as the test oracle problem)

Evaluating the impact of the failure and finding its direct cause (the Fault), and the indirect one (via Root Cause Analysis)

Judging whether testing is sufficient and can be stopped

Measuring the effectiveness of the tests

Terms used in testing l.jpg

Terms used in Testing


Fault – The cause of a failure, e.g., a missing or incorrect piece of code

Error – An intermediate unstable state

Failure – The manifested inability of the program to perform the function required, i.e., a system malfunction evidenced by incorrect output, abnormal termination or unmet time and space constraints.


A fault may remain undetected long time, until some event activates it.

If an when the error propagates to the output, it eventually causes the failure.

Fault  Error  Failure

This chain can recursively iterate: a fault in turn can be caused by the failure of some other interacting system.

Software reliability l.jpg

Software Reliability

Software reliability is a probabilistic estimate, and measures the probability that the software will execute without failure in a given environment for a given period of time.

Thus, the value of software reliability depends on how frequently those inputs that cause a failure will be exercised by the final users.

The notion of reliability is specific to “a given environment”, the tests must be drawn from an input distribution that approximates as closely as possible the future usage in operation, which is called the operational distribution.

Types of test techniques l.jpg

Types of Test Techniques

Static Analysis Techniques –

Based solely on the (manual or automated) examination of project documentation of software models and cod, and of other related information about requirements and design

Generally yields valid results, but may be weak in precision

Dynamic Test Techniques –

Exercise the software in order to expose possible failures

Behavioral and performance properties of the program are also observed

Yields more precise results, but only holding for the examined executions

Static techniques l.jpg

Static Techniques

Can be applied at the requirements phase; at the design phase; and during the implementation phase

Traditional Static Techniques – heavily manual, error-prone, time consuming

Software inspection – The step-by-step analysis of the documents (deliverables) produced, against a compiled checklist of common and historical defects

Software reviews – The process by which different aspects of the work product are presented to project personnel (managers, users, customer etc) and other interested stakeholders for comment or approval

Code reading – The desktop analysis of the produced code for discovering typing errors that do not violate style or syntax

Algorithm analysis and tracing – The process in which the complexity of algorithms employed and the worst case, average-case and probabilistic analysis evaluations can be derived

Static Analysis Techniques relying on use of Formal Methods

Dynamic techniques l.jpg

Dynamic Techniques

Dynamic Techniques –

Testing – Based on the execution of the code on valued inputs (definition of the parameters and environmental conditions that characterize a system state must be included when necessary )

Profiling – A program profile records the number of times some entities of interest occur during a set of controlled executions

Test levels l.jpg

Test Levels

Unit Test –

A unit is the smallest testable piece of software, which may consist of hundreds or even just a few lines of source code, and generally represents the result of the work of one programmer

The purpose is to ensure that the unit satisfies its functional specification and/or that its implemented structure matches the intended design structure

Integration Test –

Integration is the process by which software pieces or components are aggregated to create a larger component.

Specifically aimed at exposing the problems that can arise at this stage.

Strategies – Big-Bang, Top-Down, Bottom-Up

Test levels continued l.jpg

Test Levels Continued…

System Test –

System test involves the whole system embedded in its actual hardware environment

The primary goals of system Testing –

Discovering the failures that manifest themselves only at system level and hence were not detected during unit or integration testing

Increasing the confidence that the developed product correctly implements the required capabilities

Collecting information useful for deciding the release of the product.

System Testing includes testing for performance, security, reliability, stress testing and recovery

Acceptance Test –

An extension of system test. It is a test session conducted over the whole system, which mainly focuses on the usability requirements more than on the compliance of the implementation against some specification.

Aim is to verify that the effort required from end-users to learn to use and fully exploit the system functionalities is acceptable

Regression test l.jpg

Regression Test

Refers to the retesting of a unit, a combination of components or a whole system after modification, in order to ascertain that the change has not introduced new faults

Corrective and evolutive modifications may be performed quite often. To re-run after each change all previously executed test cases would be prohibitively expensive.

Therefore various types of techniques have been developed to reduce regression testing costs and to make it more effective.

Selective regression test techniques help in selecting a (minimized) subset of the existing test cases by examining the modifications (for instance at code level, using control flow and data flow analysis).

Objectives of testing l.jpg

Objectives of Testing

Acceptance/Qualification Testing

Installation Testing

Alpha Testing

Beta Testing

Reliability Achievement

Conformance Testing/Functional Testing

Regression Testing

Performance Testing

Usability Testing

Test-Driven Development

Strategies for test case selection l.jpg

Strategies for Test Case Selection

Test Criterion provides a decision procedure for selecting the test cases

Random Testing –

The test inputs are picked purely randomly from the whole input domain according to a specified distribution, i.e., after assigning to the inputs different “weights” (more properly probabilities)

Partition Testing –

The program input domain is divided into sub-domains within which it is assumed that the program behaves the same, i.e., for every point within a sub-domain the program either succeeds or fails

Structural Testing –

Known as Code-based testing or White-box testing

By monitoring code coverage one tries to exercise thoroughly all “program elements”

Code-based criteria should be more properly used as adequacy criteria

Specification-Based Testing –

Depending on how the program specifications are expressed, different techniques are possible - equivalence classes, boundary conditions, cause-effect graphs

Other test criteria l.jpg

Other Test Criteria

Criteria based on Tester’s Intuition & Experience –

Ad-hoc testing techniques in which tests are derived relying on the tester’s skill, intuition, and experience with similar programs

Exploratory testing – simultaneous learning, test design, and test execution; that is, the tests are not defined in advance in an established test plan, but are dynamically designed, executed, and modified

Fault-Based Testing –

Devise test cases specifically aimed at revealing categories of likely or pre-defined faults, with different degrees of formalization

Error guessing – Test cases are designed by testers trying to figure out the most plausible faults in a given program. A good source of information is the history

of faults discovered in earlier projects, as well as the tester’s expertise

Mutation testing – The underlying assumption of mutation testing is, by looking for simple syntactic faults, more complex, but real, faults will be found

Criteria based on Operational Usage –

The idea is to infer, from the observed test results, the future reliability of the software when in actual use

Test first programming l.jpg

Test-First Programming

Traditional Test Process – Test Planning, Test Design, Test Execution and Test Results Evaluation

Test-First programming focuses on the derivation of (unit and acceptance) tests before coding

Principle of the approach is to make development more lightweight by keeping design simple and reducing as much as possible the rules and the activities of traditional processes felt by developers as overwhelming and unproductive

Test execution l.jpg

Test Execution

Launching the Tests –

Forcing the execution of the test cases (manually or automatically) derived according to one of the test criteria

Test Oracles –

A test is meaningful only if it is possible to decide about its outcome

An oracle is any (human or mechanical) agent that decides whether the program behaved correctly on a given test.

The oracle is specified to output a reject verdict if it observes a failure (or even an error, for smarter oracles), and approve otherwise.

In the event that the oracle cannot reach a decision while executing a test case, the test output is classified as inconclusive

Test Tools –

The usage of appropriate tools can alleviate the burden of clerical, tedious operations, and make them less error-prone, while increasing testing efficiency and effectiveness

Test harness (drivers, stubs), Test generators, Capture/Replay, Oracle/file comparators/assertion checking, Coverage analyzer/Instrumented, Tracers, Reliability evaluation tools

Test documentation l.jpg

Test Documentation

Documentation is an integral part of the formalization of the test process, which contributes to the coordination and control of the testing phase.

Documentation –

Test Plan

Test Design Specification

Test Case Specification

Test Procedure Specification

Test Log

Test Incident or Problem Report

Test management l.jpg

Test Management

Different management activities

Scheduling the timely completion of tasks

Estimation of the effort and the resources needed to execute the tasks

Quantification of the risk associated with the tasks

Effort/Cost estimation

Quality control measures to be employed: several

Test measures l.jpg

Test Measures

Evaluation of the Program under Test

Linguistic measures – based on proprieties of the program or of the specification text (Lines of Code (LOC), statements, number of unique operands or operators)

Structural measures – based on structural relations between objects in the program and comprise control flow or data flow complexity (frequency with which modules call each other)

Hybrid measures – may result from the combination

Fault density – measured by the ratio between the number of faults found and the size of the program

Life testing, reliability evaluation

Evaluation of the Test performed

Coverage/thoroughness measure


Conclusions l.jpg


Approaches overviewed include both traditional and modern techniques

Software Testing is a complex activity

Resources need to be devoted to this activity for its success

A test approach that guarantees a perfect product cannot be found