Chapter 8 Software Testing. Chapter 8 So-ware tes0ng

Similar documents
Chapter 8 Software Testing. Chapter 8 Software testing

Software Testing. Software Testing. in the textbook. Chapter 8. Verification and Validation. Verification and Validation: Goals

Lecture 15 Software Testing

Chapter 8 Software Testing

The testing process. Component testing. System testing

Software testing. Objectives. Contents

Verification and Validation. Ian Sommerville 2004 Software Engineering, 7th edition. Chapter 22 Slide 1

Software Testing. Software Testing. in the textbook. Chapter 8. Verification and Validation. Verification Techniques

Software testing. Ian Sommerville 2006 Software Engineering, 8th edition. Chapter 23 Slide 1

Verification and Validation. Assuring that a software system meets a user s needs. Verification vs Validation. The V & V Process

Software Engineering

Part 5. Verification and Validation

Topic: Software Verification, Validation and Testing Software Engineering. Faculty of Computing Universiti Teknologi Malaysia

Software Engineering I. Chapters 8 & 9. SW Testing & Maintenance. Slide 1. 12/17/2017 SW Testing and Maintenance

W.C.Uduwela. Dept. of Mathematics & Computer Science

Recap on SDLC Phases & Artefacts. Topic: Software Verification, Validation and Testing Software Engineering

Software Testing. Massimo Felici IF

Overview. State-of-the-Art. Relative cost of error correction. CS 619 Introduction to OO Design and Development. Testing.

Verification and Validation. Ian Sommerville 2004 Software Engineering, 7th edition. Chapter 22 Slide 1

Software Testing. Software Testing

Chapter 9. Software Testing

Verification and Validation

Ian Sommerville 2006 Software Engineering, 8th edition. Chapter 22 Slide 1

Computer Science and Software Engineering University of Wisconsin - Platteville 9-Software Testing, Verification and Validation

Objectives. Chapter 19. Verification vs. validation. Topics covered. Static and dynamic verification. The V&V process

Aerospace Software Engineering

Software Engineering

Introduction to Software Engineering

Verification and Validation. Verification and validation

Topics in Software Testing

Testing. Prof. Clarkson Fall Today s music: Wrecking Ball by Miley Cyrus

The requirements engineering process

Lecture 9. Verification and Validation. Assuring that a software system meets a user's needs. Tutorial Questions. Verification and Validation

Verification and Validation

Software Testing. Lecturer: Sebastian Coope Ashton Building, Room G.18

Chapter 9 Quality and Change Management

Pearson Education 2007 Chapter 9 (RASD 3/e)

Verification and Validation

Testing & Debugging TB-1

All the subjective part of 2011 papers solved complete reference numbers

Software LEIC/LETI. Lecture 18

Software Engineering Testing and Debugging Testing

Types of Software Testing: Different Testing Types with Details

ΗΜΥ 317 Τεχνολογία Υπολογισμού

Testing. Topics. Types of Testing. Types of Testing

Write perfect C code to solve the three problems below.

Software Quality. Richard Harris

Software Engineering 2 A practical course in software engineering. Ekkart Kindler

Software Engineering Theory. Lena Buffoni (slides by Kristian Sandahl/Mariam Kamkar) Department of Computer and Information Science

Higher-order Testing. Stuart Anderson. Stuart Anderson Higher-order Testing c 2011

Software testing A.A. 2018/2019

2. SYSTEM MODELS, DESIGN AND IMPLEMENTATION

CSE 403: Software Engineering, Fall courses.cs.washington.edu/courses/cse403/16au/ Unit Testing. Emina Torlak

Chapter 11, Testing. Using UML, Patterns, and Java. Object-Oriented Software Engineering

INTRODUCTION TO SOFTWARE ENGINEERING

In this Lecture you will Learn: Testing in Software Development Process. What is Software Testing. Static Testing vs.

Lecture 26: Testing. Software Engineering ITCS 3155 Fall Dr. Jamie Payton

Black-box Testing Techniques

Sample Exam Syllabus

Examination Questions Time allowed: 1 hour 15 minutes

Chapter 10. Testing and Quality Assurance

Computational Systems COMP1209

People tell me that testing is

Chapter 5 System modeling

Three General Principles of QA. COMP 4004 Fall Notes Adapted from Dr. A. Williams

Testing! Prof. Leon Osterweil! CS 520/620! Spring 2013!

Software Testing Interview Question and Answer

Darshan Institute of Engineering & Technology for Diploma Studies

Patterns and Testing

Design and Implementa3on

Testing. ECE/CS 5780/6780: Embedded System Design. Why is testing so hard? Why do testing?

An Introduction to Systematic Software Testing. Robert France CSU

Facts About Testing. Cost/benefit. Reveal faults. Bottom-up. Testing takes more than 50% of the total cost of software development

CS 4387/5387 SOFTWARE V&V LECTURE 4 BLACK-BOX TESTING

Bridge Course On Software Testing

No Source Code. EEC 521: Software Engineering. Specification-Based Testing. Advantages

Software Design Models, Tools & Processes. Lecture 6: Transition Phase Cecilia Mascolo

MONIKA HEINER.

Software Quality Assurance & Testing

Learning outcomes. Systems Engineering. Debugging Process. Debugging Process. Review

Ingegneria del Software II academic year: Course Web-site: [

Agile Manifesto & XP. Topics. Rapid software development. Agile methods. Chapter ) What is Agile trying to do?

Black Box Testing. EEC 521: Software Engineering. Specification-Based Testing. No Source Code. Software Testing

Ch 4: Requirements Engineering. What are requirements?

Sample Question Paper. Software Testing (ETIT 414)

Software specification and modelling. Requirements engineering

Lecture 20: SW Testing Presented by: Mohammad El-Ramly, PhD

Testing Objectives. Successful testing: discovers previously unknown errors

Software Engineering (CSC 4350/6350) Rao Casturi

Software Engineering Fall 2015 (CSC 4350/6350) TR. 5:30 pm 7:15 pm. Rao Casturi 11/10/2015

Testing is executing a system in order to identify any gaps, errors, or missing requirements in contrary to the actual requirements.

Efficient Regression Test Model for Object Oriented Software

Manuel Oriol, CHCRC-C, Software Testing ABB

Testing. Unit, integration, regression, validation, system. OO Testing techniques Application of traditional techniques to OO software

XVIII. Software Testing. Laurea Triennale in Informatica Corso di Ingegneria del Software I A.A. 2006/2007 Andrea Polini

Software Engineering Fall 2014

Verification and Validation

Verification and Validation

Quality Assurance = Testing? SOFTWARE QUALITY ASSURANCE. Meaning of Quality. How would you define software quality? Common Measures.

IT323 - Software Engineering 2 1

Transcription:

Chapter 8 Software Testing 1

Topics covered ² Introduction to testing ² Stages for testing software system are: Development testing Release testing User testing ² Test-driven development as interleave approach. 2

What is program testing? ² Testing is intended to show that a program does what it is intended to do and to discover program defects before it is put into use. ² When you test software, you execute a program using artificial data. ² You check the results of the test run for errors, anomalies or information about the program s non-functional attributes. ² Can reveal the presence of errors NOT their absence. ² Testing is part of a more general verification and validation process, which also includes static validation techniques. 3

Program testing goals ² To demonstrate to the developer and the customer that the software meets its requirements. For custom software, this means that there should be at least one test for every requirement in the requirements document. For generic software products, it means that there should be tests for all of the system features, plus combinations of these features, that will be incorporated in the product release. ² To discover situations in which the behavior of the software is incorrect, undesirable or does not conform to its specification. Defect testing is concerned with rooting out undesirable system behavior such as system crashes, unwanted interactions with other systems, incorrect computations and data corruption. 4

Validation and defect testing ² The first goal leads to validation testing You expect the system to perform correctly using a given set of test cases that reflect the system s expected use. ² The second goal leads to defect testing The test cases are designed to expose defects. The test cases in defect testing can be deliberately obscure and need not reflect how the system is normally used. 5

Testing process goals ² Validation testing To demonstrate to the developer and the system customer that the software meets its requirements A successful test shows that the system operates as intended. ² Defect testing To discover faults or defects in the software where its behaviour is incorrect or not in conformance with its specification A successful test is a test that makes the system perform incorrectly and so exposes a defect in the system. 6

An input-output model of program testing Input test data I e Inputs causing anomalous behaviour System Output test results Oe Outputs which reveal the presence of defects 7

Verification vs validation ² Verification: "Are we building the product right. The software should conform to its specification. ² Validation: "Are we building the right product. The software should do what the user really requires. 8

Comparison between V & V Verifica(on Main purpose is to detect defects in the ar0facts of the system under development. System meets designer specifica0ons ² Includes reviews, inspec0ons, walkthroughs, unit tes0ng and integra0on tes0ng. Valida(on Main purpose is to show that the system under development meets user needs, requirements, and expecta0ons. System meets customer needs. Includes program reviews, system tes0ng, customer acceptance tes0ng. 9

V & V confidence ² Aim of V & V is to establish confidence that the system is fit for purpose. ² Level of confidence depends on the following factors: Software purpose The level of confidence depends on how critical the software is to an organisation. User expectations Users may have low expectations of certain kinds of software. Marketing environment Getting a product to market early may be more important than finding defects in the program. 10

Inspections and testing ² Software inspectionsconcerned with analysis of the static system representation to discover problems (static verification) May be supplement by tool-based document and code analysis. Discussed in Chapter 15. ² Software testingconcerned with exercising and observing product behaviour (dynamic verification) The system is executed with test data and its operational behaviour is observed. 11

V & V techniques Sta0c Methods Techniques applied to ar0facts without execu0on. Inspec'ons and Reviews Techniques Dynamic Methods Techniques applied to ar0facts through execu0on. Tes'ng 12

Inspections and testing Inspections Requirements specification Software architecture UML design models Database schemas Program System prototype Testing 13

Software inspections ² These involve people examining the source representation with the aim of discovering anomalies and defects. ² Inspections not require execution of a system so may be used before implementation. ² They may be applied to any representation of the system (requirements, design,configuration data, test data, etc.). ² They have been shown to be an effective technique for discovering program errors. 14

Advantages of inspections ² During testing, errors can mask (hide) other errors. Because inspection is a static process, you don t have to be concerned with interactions between errors. ² Incomplete versions of a system can be inspected without additional costs. If a program is incomplete, then you need to develop specialized test harnesses to test the parts that are available. ² As well as searching for program defects, an inspection can also consider broader quality attributes of a program, such as compliance with standards, portability and maintainability. 15

Inspections and testing ² Inspections and testing are complementary and not opposing verification techniques. ² Both should be used during the V & V process. ² Inspections can check conformance with a specification but not conformance with the customer s real requirements. ² Inspections cannot check non-functional characteristics such as performance, usability, etc. 16

A model of the software testing process Test cases Test data Test results Test reports Design test cases Prepare test data Run program with test data Compare results to test cases 17

Stages of testing ² Development testing, where the system is tested during development to discover bugs and defects. ² Release testing, where a separate testing team test a complete version of the system before it is released to users. ² User testing, where users or potential users of a system test the system in their own environment. 18

Question ² What do you think about testing process? ² should we use a tool to test system?? Or tester should be involved? 19

Defect Testing and Debugging?? What do you think is the difference? 20

I. Development testing ² Development testing includes all testing activities that are carried out by the team developing the system. Unit testing, where individual program units or object classes are tested. Unit testing should focus on testing the functionality of objects or methods. Component testing, where several individual units are integrated to create composite components. Component testing should focus on testing component interfaces. System testing, where some or all of the components in a system are integrated and the system is tested as a whole. System testing should focus on testing component interactions. 21

I.1 Unit testing ² Unit testing is the process of testing individual components in isolation. ² It is a defect testing process. ² Units may be: Individual functions or methods within an object Object classes with several attributes and methods Composite components with defined interfaces used to access their functionality. 22

Object class testing ² Complete test coverage of a class involves Testing all operations associated with an object Setting and interrogating all object attributes Exercising the object in all possible states. ² Inheritance makes it more difficult to design object class tests as the information to be tested is not localised. 23

EX. The weather station object interface identifier WeatherStation reportweather ( ) reportstatus ( ) powersave (instruments) remotecontrol (commands) reconfigure (commands) restart (instruments) shutdown (instruments) 24

EX. Weather station testing ² Need to define test cases for reportweather, calibrate, test, startup and shutdown. ² Using a state model, identify sequences of state transitions to be tested and the event sequences to cause these transitions ² For example: Shutdown -> Running-> Shutdown Configuring-> Running-> Testing -> Transmitting -> Running Running-> Collecting-> Running-> Summarizing -> Transmitting -> Running 25

Automated unit testing ² Whenever possible, unit testing should be automated so that tests are run and checked without manual intervention. ² In automated unit testing, you make use of a test automation framework (such as JUnit) to write and run your program tests. ² Unit testing frameworks provide generic test classes that you extend to create specific test cases. They can then run all of the tests that you have implemented and report, often through some GUI, on the success of otherwise of the tests. 26

An automated test has three parts: ² A setup part, where you initialize the system with the test case, namely the inputs and expected outputs. ² A call part, where you call the object or method to be tested. ² An assertion part where you compare the result of the call with the expected result. If the assertion evaluates to true, the test has been successful if false, then it has failed. 27

Unit test effectiveness ² The unit test cases should show that, when used as expected, the component that you are testing does what it is supposed to do. ² If there are defects in the component, these should be revealed by test cases. ² This leads to 2 types of unit test case: The first of these should reflect normal operation of a program and should show that the component works as expected. The other kind of test case should be based on testing experience of where common problems arise. It should use abnormal inputs to check that these are properly processed and do not crash the component. 28

Testing strategies ² Partition testing, where you identify groups of inputs that have common characteristics and should be processed in the same way. You should choose tests from within each of these groups. ² Guideline-based testing, where you use testing guidelines to choose test cases. These guidelines reflect previous experience of the kinds of errors that programmers often make when developing components. 29

I.1.a Partition testing ² Input data and output results often fall into different classes where all members of a class are related. ² Each of these classes is an equivalence partition or domain where the program behaves in an equivalent way for each class member. ² Test cases should be chosen from each partition. 30

Equivalence partitioning Input equivalence partitions Output partitions System Possible inputs Correct outputs Possible outputs 31

Equivalence partitions Example 1. Program accepts 4-10 integer numbers 2. Numbers contain exactly 5 digits greater than 10,000 3 4 7 11 10 Less than 4 Between 4 and 10 More than 10 Number of input values 9999 10000 50000 100000 99999 Less than 10000 Between 10000 and 99999 More than 99999 Input values 32

Activities: Equivalence partitions ² Draw equivalence partitioning to check whether the number of a day of the week is valid or not? ² You have been asked to test a method called catwhitespace in a Paragraph object that, within the paragraph, replaces sequences of blank characters with a single blank character. Identify testing partitions for this example and derive a set of tests for the catwhitespace method.

34

I.1.b Guidelines-based testing (sequences) ² Test software with sequences which have only a single value. ² Use sequences of different sizes in different tests. ² Derive tests so that the first, middle and last elements of the sequence are accessed. ² Test with sequences of zero length. 35

Ex. of Partition testing: Search Routine Specification è Search Rou0ne Input Par00ons Inputs which conform to the pre- condi0ons. Inputs where a pre- condi0on does not hold. Inputs where the key element is a member of the array. Inputs where the key element is not a member of the array. 36

Ex. of Partition testing using guidelines (sequences): Search Routine Specification Sequence Single value Single value More than 1 value More than 1 value More than 1 value More than 1 value Element In sequence Not in sequence First element in sequence Last element in sequence Middle element in sequence Not in sequence Input sequence (T) Key (Key) Output (Found, L) 17 17 true, 1 17 0 false,?? 17, 29, 21, 23 17 true, 1 41, 18, 9, 31, 30, 16, 45 45 true, 7 17, 18, 21, 23, 29, 41, 38 23 true, 4 21, 23, 29, 33, 38 25 false,?? 37

General testing guidelines ² Choose inputs that force the system to generate all error messages ² Design inputs that cause input buffers to overflow ² Repeat the same input or series of inputs numerous times ² Force invalid outputs to be generated ² Force computation results to be too large or too small. 38

I.2 Component testing ² Software components are often composite components that are made up of several interacting objects. For example, in the weather station system, the reconfiguration component includes objects that deal with each aspect of the reconfiguration. ² You access the functionality of these objects through the defined component interface. ² Testing composite components should therefore focus on showing that the component interface behaves according to its specification. You can assume that unit tests on the individual objects within the component have been completed. 39

Interface testing Test cases A B C 40

Interface testing ² Objectives are to detect faults due to interface errors or invalid assumptions about interfaces. ² Interface types Parameter interfaces Data passed from one method or procedure to another. Shared memory interfaces Block of memory is shared between procedures or functions. Procedural interfaces Sub-system encapsulates a set of procedures to be called by other sub-systems. Message passing interfaces Sub-systems request services from other sub-systems 41

Interface errors ² Interface misuse A calling component calls another component and makes an error in its use of its interface e.g. parameters in the wrong order. ² Interface misunderstanding A calling component embeds assumptions about the behaviour of the called component which are incorrect. ² Timing errors The called and the calling component operate at different speeds and out-of-date information is accessed. 42

Interface testing guidelines ² Design tests so that parameters to a called procedure are at the extreme ends of their ranges. ² Always test pointer parameters with null pointers. ² Design tests which cause the component to fail. ² Use stress testing in message passing systems. ² In shared memory systems, vary the order in which components are activated. 43

I.3 System testing ² System testing during development involves integrating components to create a version of the system and then testing the integrated system. ² The focus in system testing is testing the interactions between components. ² System testing checks that components are compatible, interact correctly and transfer the right data at the right time across their interfaces. ² System testing tests the emergent behaviour of a system. 44

System and component testing ² During system testing, reusable components that have been separately developed and off-the-shelf systems may be integrated with newly developed components. The complete system is then tested. ² Components developed by different team members or sub-teams may be integrated at this stage. System testing is a collective rather than an individual process. In some companies, system testing may involve a separate testing team with no involvement from designers and programmers. 45

System and component testing ² Component testing Testing of individual program components; Usually the responsibility of the component developer; Tests are derived from the developer s experience. ² System testing Testing of groups of components integrated to create a system or sub-system; The responsibility of an independent testing team; Tests are based on a system specification. 46

Use-case testing ² The use-cases developed to identify system interactions can be used as a basis for system testing. ² Each use case usually involves several system components so testing the use case forces these interactions to occur. ² The sequence diagrams associated with the use case documents the components and interactions that are being tested. 47

Collectweather data sequence chart Weather information system SatComms WeatherStation Commslink WeatherData request (report) acknowledge reportweather () acknowledge get (summary) summarise () send (report) reply (report) acknowledge acknowledge 48

Testing policies ² Exhaustive system testing is impossible so testing policies which define the required system test coverage may be developed. ² Examples of testing policies: All system functions that are accessed through menus should be tested. Combinations of functions (e.g. text formatting) that are accessed through the same menu must be tested. Where user input is provided, all functions must be tested with both correct and incorrect input. 49

Ex. Testing policies ² All functions accessed through menus should be tested; 50

Ex. Testing policies ² Combinations of functions accessed through the same menu should be tested; Column formatting + Inserting Footnotes 51

Ex. Testing policies ² Where user input is required, all functions must be tested with correct and incorrect input. 52

II. Release testing ² Release testing is the process of testing a particular release of a system that is intended for use outside of the development team. ² The primary goal of the release testing process is to convince the supplier of the system that it is good enough for use. Release testing, therefore, has to show that the system delivers its specified functionality, performance and dependability, and that it does not fail during normal use. ² Release testing is usually a black-box testing process where tests are only derived from the system specification. 53

Release testing and system testing ² Release testing is a form of system testing. ² Important differences: A separate team that has not been involved in the system development, should be responsible for release testing. System testing by the development team should focus on discovering bugs in the system (defect testing). The objective of release testing is to check that the system meets its requirements and is good enough for external use (validation testing). 54

Requirements based testing ² Requirements-based testing involves examining each requirement and developing a test or tests for it. ² MHC-PMS requirements: If a patient is known to be allergic to any particular medication, then prescription of that medication shall result in a warning message being issued to the system user. If a prescriber chooses to ignore an allergy warning, they shall provide a reason why this has been ignored. 55

Requirements tests ² Set up a patient record with no known allergies. Prescribe medication for allergies that are known to exist. Check that a warning message is not issued by the system. ² Set up a patient record with a known allergy. Prescribe the medication to that the patient is allergic to, and check that the warning is issued by the system. ² Set up a patient record in which allergies to two or more drugs are recorded. Prescribe both of these drugs separately and check that the correct warning for each drug is issued. ² Prescribe two drugs that the patient is allergic to. Check that two warnings are correctly issued. ² Prescribe a drug that issues a warning and overrule that warning. Check that the system requires the user to provide information explaining why the warning was overruled. 56

Features tested by scenario ² Authentication by logging on to the system. ² Downloading and uploading of specified patient records to a laptop. ² Home visit scheduling. ² Encryption and decryption of patient records on a mobile device. ² Record retrieval and modification. ² Links with the drugs database that maintains side-effect information. ² The system for call prompting. 57

A usage scenario for the MHC-PMS Kate is a nurse who specializes in mental health care. One of her responsibilities is to visit patients at home to check that their treatment is effective and that they are not suffering from medication side -effects. On a day for home visits, Kate logs into the MHC-PMS and uses it to print her schedule of home visits for that day, along with summary information about the patients to be visited. She requests that the records for these patients be downloaded to her laptop. She is prompted for her key phrase to encrypt the records on the laptop. One of the patients that she visits is Jim, who is being treated with medication for depression. Jim feels that the medication is helping him but believes that it has the side -effect of keeping him awake at night. Kate looks up Jim s record and is prompted for her key phrase to decrypt the record. She checks the drug prescribed and queries its side effects. Sleeplessness is a known side effect so she notes the problem in Jim s record and suggests that he visits the clinic to have his medication changed. He agrees so Kate enters a prompt to call him when she gets back to the clinic to make an appointment with a physician. She ends the consultation and the system re-encrypts Jim s record. After, finishing her consultations, Kate returns to the clinic and uploads the records of patients visited to the database. The system generates a call list for Kate of those patients who she has to contact for follow-up information and make clinic appointments. 58

Performance testing ² Part of release testing may involve testing the emergent properties of a system, such as performance and reliability. ² Tests should reflect the profile of use of the system. ² Performance tests usually involve planning a series of tests where the load is steadily increased until the system performance becomes unacceptable. ² Stress testing is a form of performance testing where the system is deliberately overloaded to test its failure behaviour. 59

III. User testing ² User or customer testing is a stage in the testing process in which users or customers provide input and advice on system testing. ² User testing is essential, even when comprehensive system and release testing have been carried out. The reason for this is that influences from the user s working environment have a major effect on the reliability, performance, usability and robustness of a system. These cannot be replicated in a testing environment. 60

Types of user testing ² Alpha testing Users of the software work with the development team to test the software at the developer s site. ² Beta testing A release of the software is made available to users to allow them to experiment and to raise problems that they discover with the system developers. ² Acceptance testing Customers test a system to decide whether or not it is ready to be accepted from the system developers and deployed in the customer environment. Primarily for custom systems. 61

The acceptance testing process Test criteria Test plan Tests Test results Testing report Define acceptance criteria Plan acceptance testing Derive acceptance tests Run acceptance tests Negotiate test results Accept or reject system 62

Stages in the acceptance testing process ² Define acceptance criteria ² Plan acceptance testing ² Derive acceptance tests ² Run acceptance tests ² Negotiate test results ² Reject/accept system 63

Agile methods and acceptance testing ² In agile methods, the user/customer is part of the development team and is responsible for making decisions on the acceptability of the system. ² Tests are defined by the user/customer and are integrated with other tests in that they are run automatically when changes are made. ² There is no separate acceptance testing process. ² Main problem here is whether or not the embedded user is typical and can represent the interests of all system stakeholders. 64

Test-driven development ² Test-driven development (TDD) is an approach to program development in which you inter-leave testing and code development. ² Tests are written before code and passing the tests is the critical driver of development. ² You develop code incrementally, along with a test for that increment. You don t move on to the next increment until the code that you have developed passes its test. ² TDD was introduced as part of agile methods such as Extreme Programming. However, it can also be used in plan-driven development processes. 65

Test-driven development Identify new functionality pass Write test Run test fail Implement functionality and refactor 66

TDD process activities ² Start by identifying the increment of functionality that is required. This should normally be small and implementable in a few lines of code. ² Write a test for this functionality and implement this as an automated test. ² Run the test, along with all other tests that have been implemented. Initially, you have not implemented the functionality so the new test will fail. ² Implement the functionality and re-run the test. ² Once all tests run successfully, you move on to implementing the next chunk of functionality. 67

Benefits of test-driven development ² Code coverage Every code segment that you write has at least one associated test so all code written has at least one test. ² Regression testing A regression test suite is developed incrementally as a program is developed. ² Simplified debugging When a test fails, it should be obvious where the problem lies. The newly written code needs to be checked and modified. ² System documentation The tests themselves are a form of documentation that describe what the code should be doing. 68

Regression testing ² Regression testing is testing the system to check that changes have not broken previously working code. ² In a manual testing process, regression testing is expensive but, with automated testing, it is simple and straightforward. All tests are rerun every time a change is made to the program. ² Tests must run successfully before the change is committed. 69

Key points ² Testing can only show the presence of errors in a program. It cannot demonstrate that there are no remaining faults. ² Development testing is the responsibility of the software development team. A separate team should be responsible for testing a system before it is released to customers. ² Development testing includes unit testing, in which you test individual objects and methods component testing in which you test related groups of objects and system testing, in which you test partial or complete systems. 70

Key points ² When testing software, you should try to break the software by using experience and guidelines to choose types of test case that have been effective in discovering defects in other systems. ² Wherever possible, you should write automated tests. The tests are embedded in a program that can be run every time a change is made to a system. ² Test-first development is an approach to development where tests are written before the code to be tested. ² Scenario testing involves inventing a typical usage scenario and using this to derive test cases. ² Acceptance testing is a user testing process where the aim is to decide if the software is good enough to be deployed and used in its operational environment. 71