1 / 53

Advanced Software Engineering: Software Testing COMP 3702

Advanced Software Engineering: Software Testing COMP 3702. Instructor: Anneliese Andrews. News & Project. News Updated course program Reading instructions The book, deadline 23/3 Project IMPORTANT to read the project description thoroughly Schedule, Deadlines , Activities

havyn
Download Presentation

Advanced Software Engineering: Software Testing COMP 3702

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Advanced Software Engineering: Software TestingCOMP 3702 Instructor:Anneliese Andrews

  2. News & Project • News • Updated course program • Reading instructions • The book, deadline 23/3 • Project • IMPORTANT to read the project description thoroughly • Schedule, Deadlines, Activities • Requirements (7-10 papers), project areas • Report, template, presentation A Andrews - Software Engineering: Software Testing'06

  3. Lecture • Chapter 4 (Lab 1) • Black-box testing techniques • Chapter 12 (Lab 2) • Statistical testing • Usage modelling • Reliability A Andrews - Software Engineering: Software Testing'06

  4. Why test techniques? • Exhaustive testing (use of all possible inputs and conditions) is impractical • must use a subset of all possible test cases • want must have high probability of detecting faults • Need processes that help us selecting test cases • Different people – equal probability to detect faults • Effective testing – detect more faults • Focus attention on specific types of fault • Know you’re testing the right thing • Efficient testing – detect faults with less effort • Avoid duplication • Systematic techniques are measurable A Andrews - Software Engineering: Software Testing'06

  5. Dimensions of testing • Testing combines techniques that focus on • Testers – who does the testing • Coverage – what gets tested • Potential problems – why you're testing (risks / quality) • Activities – how you test • Evaluation – how to tell whether the test passed or failed • All testing should involve all five dimensions • Testing standards (e.g. IEEE) A Andrews - Software Engineering: Software Testing'06

  6. Black-box testing A Andrews - Software Engineering: Software Testing'06

  7. Equivalence partitioning mouse picks on menu Partitioning is based on input conditions output format requests user queries responses to prompts numerical data command key input A Andrews - Software Engineering: Software Testing'06

  8. Equivalence partitioning Input condition: • is a range • one valid and two invalid classes are defined • requires a specific value • one valid and two invalid classes are defined • is a boolean • one valid and one invalid class are defined A Andrews - Software Engineering: Software Testing'06

  9. Test Cases • Which test cases have the best chance of successfully uncovering faults? • as near to the mid-point of the partition as possible • the boundaries of the partition and • Mid-point of a partition typically represents the “typical values” • Boundary values represent the atypical or unusual values • Usually identify equivalence partitions based on specs and experience A Andrews - Software Engineering: Software Testing'06

  10. Equivalence Partitioning Example • Consider a system specification which states that a program will accept between 4 and 10 input values (inclusive), where the input values must be 5 digit integers greater than or equal to 10000 • What are the equivalence partitions? A Andrews - Software Engineering: Software Testing'06

  11. Example Equivalence Partitions A Andrews - Software Engineering: Software Testing'06

  12. mouse picks on menu output format requests user queries responses to prompts numerical data command key input Boundary value analysis output domain A Andrews - Software Engineering: Software Testing'06

  13. Boundary value analysis • Range a..b  a, b, just above a, just below b • Number of values: max, min, just below min, just above max • Output bounds should be checked • Boundaries of externally visible data structures shall be checked (e.g. arrays) A Andrews - Software Engineering: Software Testing'06

  14. Some other black-box techniques • Risk-based testing, random testing • Stress testing, performance testing • Cause-and-effect graphing • State-transition testing A Andrews - Software Engineering: Software Testing'06

  15. Error guessing • Exploratory testing, happy testing, ... • Always worth including • Can detect some failures that systematic techniques miss • Consider • Past failures (fault models) • Intuition • Experience • Brain storming • ”What is the craziest thing we can do?” • Lists in literature A Andrews - Software Engineering: Software Testing'06

  16. Characteristics Accessibility Responsiveness Efficiency Comprehensibility Environments Free form tasks Procedure scripts Paper screens Mock-ups Field trial Usability testing A Andrews - Software Engineering: Software Testing'06

  17. Specification-based testing • Formal method • Test cases derived from a (formal) specification (requirements or design) Model (state chart) Specification Test case generation Test execution A Andrews - Software Engineering: Software Testing'06

  18. Model Usage Requirements Model-based Testing VALIDATION Specification Top-levelDesign Integration Detailed Design Unit Test Test phase Coding A Andrews - Software Engineering: Software Testing'06

  19. Statistical testing /Usage based testing A Andrews - Software Engineering: Software Testing'06

  20. Usage specification models A Andrews - Software Engineering: Software Testing'06

  21. Usage specification models A Andrews - Software Engineering: Software Testing'06

  22. Operational profiles A Andrews - Software Engineering: Software Testing'06

  23. Operational profiles A Andrews - Software Engineering: Software Testing'06

  24. Code Statistical testing / Usage-based testing Usage model Random sample A Andrews - Software Engineering: Software Testing'06

  25. Each transition corresponds to an external event Probabilities are set according to the future use of the system Reliability prediction Usage Modelling Invoke Click on OK with non-valid hour Right-click Move Dialog Box Main Window Resize CANCEL or OK with Valid Hour Close Window Terminate A Andrews - Software Engineering: Software Testing'06

  26. P41 P21 N1 P12 N2 P24 P31 P14 P13 P34 N4 N3 Markov model • System states, seen as nodes • Probabilities of transitions Conditions for a Markov model: • Probabilities are constants • No memory of past states A Andrews - Software Engineering: Software Testing'06

  27. F Model of a program • The program is seen as a graph • One entry node (invoke) and one exit node (terminate) • Every transition from node Ni to node Nj has a probability of Pij • If no connection between Ni and Nj, then Pij= 0 P21 N1 P12 N2 Input P24 P31 P14 P13 Output P34 N4 N3 A Andrews - Software Engineering: Software Testing'06

  28. Clock Software A Andrews - Software Engineering: Software Testing'06

  29. Input Domain – Subpopulations • Human users – keystrokes, mouse clicks • System clock – time/date input • Combination usage - time/date changes from the OS while the clock is executing • Create one Markov chain to model the input from the user A Andrews - Software Engineering: Software Testing'06

  30. Operation modes of the clock • Window = {main window, change window, info window} • Setting = {analog, digital} • Display = {all, clock only} • Cursor = {time, date, none} A Andrews - Software Engineering: Software Testing'06

  31. State of the system • A state of the system under test is an element of the set S, where S is the cross product of the operational modes. • States of the clock {main window, analog, all, none} {main window, analog, clock-only, none} {main window, digital, all, none} {main window, digital, clock-only, none} {change window, analog, all, time} {change window, analog, all, date} {change window, digital, all, time} {change window, digital, all, date} {info window, analog, all, none} {info window, digital, all, none} A Andrews - Software Engineering: Software Testing'06

  32. Top Level Markov Chain Window operational mode is chosen as the primary modeling mode Rules for Markov chainsEach arc is assigned a probability between 0 and 1 inclusive,The sum of the exit arc probabilities from each state is exactly 1. A Andrews - Software Engineering: Software Testing'06

  33. Top Level Model – Data Dictionary A Andrews - Software Engineering: Software Testing'06

  34. Level 2 Markov Chain Submodel for the Main Window A Andrews - Software Engineering: Software Testing'06

  35. Data Dictionary – Level 2 A Andrews - Software Engineering: Software Testing'06

  36. Data Dictionary – Level 2 A Andrews - Software Engineering: Software Testing'06

  37. Level 2 Markov Chain Submodel for the Change Window A Andrews - Software Engineering: Software Testing'06

  38. Data Dictionary A Andrews - Software Engineering: Software Testing'06

  39. Software Reliability • Techniques • Markov models • Reliability growth models A Andrews - Software Engineering: Software Testing'06

  40. Dimensions of dependability A Andrews - Software Engineering: Software Testing'06

  41. Costs of increasing dependability C o s t Dependability L o w M e d i u m H i g h V e r y U l t r a - h i g h h i g h A Andrews - Software Engineering: Software Testing'06

  42. Availability and reliability • Reliability • The probability of failure-free system operation over a specified time in a given environment for a given purpose • Availability • The probability that a system, at a point in time, will be operational and able to deliver the requested services Both of these attributes can be expressed quantitatively A Andrews - Software Engineering: Software Testing'06

  43. Reliability terminology A Andrews - Software Engineering: Software Testing'06

  44. Usage profiles / Reliability Removing X% of the faults in a system will not necessarily improve the reliability by X%! A Andrews - Software Engineering: Software Testing'06

  45. Reliability achievement • Fault avoidance • Minimise the possibility of mistakes • Trap mistakes • Fault detection and removal • Increase the probability of detecting and correcting faults • Fault tolerance • Run-time techniques A Andrews - Software Engineering: Software Testing'06

  46. Reliability quantities • Execution time • is the CPU time that is actually spent by the computer in executing the software • Calendar time • is the time people normally experience in terms of years, months, weeks, etc • Clock time • is the elapsed time from start to end of computer execution in running the software A Andrews - Software Engineering: Software Testing'06

  47. Reliability metrics A Andrews - Software Engineering: Software Testing'06

  48. Nonhomogeneous Poisson Process (NHPP) Models • N(t) follows a Poisson distribution. The probability that N(t) is a given integer n is: • m(t) = (t) is called mean value function, it describes the expected cumulative number of failures in [0,t) A Andrews - Software Engineering: Software Testing'06

  49. The Goel-Okumoto (GO) model Assumptions • The cumulative number of failures detected at time t follows a Poisson distribution • All failures are independent and have the same chance of being detected • All detected faults are removed immediately and no new faults are introduced • The failure process is modelled by an NHPP model with mean value function (t) given by: A Andrews - Software Engineering: Software Testing'06

  50. Goel-Okumoto The shape of the mean value function ((t)) and the intensity function ((t)) of the GO model (t) (t) t A Andrews - Software Engineering: Software Testing'06

More Related