Testing the system - PowerPoint PPT Presentation

1 / 34
About This Presentation
Title:

Testing the system

Description:

System testing process ... Techniques used in system testing ... Based on system functional requirement ... – PowerPoint PPT presentation

Number of Views:33
Avg rating:3.0/5.0
Slides: 35
Provided by: shari73
Category:

less

Transcript and Presenter's Notes

Title: Testing the system


1
Testing the system
  • System testing involves working with the entire
    development team and ensuring that system does
    what the customer wants.
  • Unlike unit and integrating testing which could
    be done by one self or with others. The testing
    ensures that the code implements the design
    properly

2
Causes of faults
Requirement analysis Incorrect or unclear
requirement or translation. System design
Incorrect or unclear translation or design
specification. Program design Misinterpretation
of system design, incorrect or unclear design
specification. Program Implementation
Misinterpretation of program design, incorrect
documentation, syntax or semantics. Unit/integrati
on testing Incomplete test procedures, new
faults introduced as new one is
corrected System testing Incomplete test
procedures Maintenance Incorrect user
documentation, Poor human factors, changes in
requirement, new faults introduced as old one is
corrected.
3
Steps in system testing process
Integration test
Functiontest
Performance test
Acceptance test
Installation test
user environment
System functional requirement
Other Software requirement
customer requirement specification
4
System testing process
  • Function testing does the integrated system
    perform as promised by the requirements
    specification? example can a bank account package
    credit a deposit, calculate interest, print
    balance
  • Performance testing are the non-functional
    requirements met? Like speed, reliability of
    results, security of say pin

5
System testing process
  • System is now said to be verified by designer
    and if checked with customer requirements it is
    said to be validated.
  • Acceptance testing is the system what the
    customer expects? Testing done by the customer
  • Installation testing does the system run at the
    customer site(s)?

6
Techniques used in system testing
  • Build or spin plan for gradual testing
  • Build plan for telecommunication systems
  • spin function Test Start
    Test End
  • 0 Exchange 1 Sept.
    15 Sept.
  • 1 Area code 30 Sept.
    15 Oct.
  • 2 State 25 Oct.
    5 Nov.
  • 3 Country 10 Nov.
    20 Nov.
  • 4 International 1 Dec.
    15 Dec.

7
Techniques used cont
  • A system configuration is a collection of system
    components delivered to a particular customer.
    Example separate packages for Unix, Windows and
    Mac machines
  • Configuration management is managing changes in
    the different configurations.

8
Techniques used cont
  • Configuration management
  • Versions (n) and releases (m) n.m
  • production system vs. development system
  • deltas (difference between versions), separate
    files for each version and conditional
    compilation for only new and old version
  • change control (used to show how change affects
    design, documentation, people fixing problems
    simultaneosly

9
Techniques used in system testing
  • Regression testing Test applied to new versions
    to see that they worked as the old version

10
Test team
  • Professional testers organize and run the tests
  • Analysts who created requirements
  • System designers understand the proposed
    solution
  • Configuration management specialists to help
    control fixes
  • Users to evaluate issues that arise with ease
    of use and other human factors
  • NB no one from the implementation team can be
    on the test team

11
Function testing
  • Testing the various functions
  • It is closed box approach
  • Based on system functional requirement
  • The set of actions associated with a function is
    called a thread
  • We will use Cause-and-effect graphs to generate
    test cases from requirement.

12

13
Cause-and-effect graphs (2)
Causes
14
Cause-and-effect graphs (3)
Effects
Intermediate nodes
15
Graph (see page 398-400)
10
1
E3
11
2
3
E1
4
5
E2
16
The columns are test cases. NB that 5 causes each
with 2 states will yield 25 32 test cases.
Table 9.2. Decision table for cause-and-effect
graph.
Test 1
Test 2
Test 3
Test 4
Test 5
Cause 1
I
I
I
S
I
Cause 2
I
I
I
X
S
Cause 3
I
S
S
X
X
Cause 4
S
I
S
X
X
Cause 5
S
S
I
X
X
Effect 1
P
P
A
A
A
Effect 2
A
A
P
A
A
Effect 3
A
A
A
P
P
17
Performance tests
  • Stress tests
  • Volume tests
  • Configuration tests
  • Compatibility tests
  • Regression tests
  • Security tests
  • Timing tests
  • Environmental tests
  • Quality tests
  • Recovery tests
  • Maintenance tests
  • Documentation tests
  • Human factors (usability) tests

18
Performance Testing cont
  • Software reliability The probability that a
    system will operate without failure under given
    condition for a given interval
  • Software availability The probability that the
    system is operating successfully according to
    specification at a given time.
  • Software maintainability The probability that
    for a given condition of use, a maintainance
    activity can be carried out within a stated time
    interval and using stated procedures and resources

19
Table 9.3. Inter-failure times (read left to
right, in rows).
3
30
113
81
115
9
2
91
112
15
138
50
77
24
108
88
670
120
26
114
325
55
242
68
422
180
10
1146
600
15
36
4
0
8
227
65
176
58
457
300
97
263
452
255
197
193
6
79
816
1351
148
21
233
134
357
193
236
31
369
748
0
232
330
365
1222
543
10
16
529
379
44
129
810
290
300
529
281
160
828
1011
445
296
1755
1064
1783
860
983
707
33
868
724
2323
2930
1461
843
12
261
1800
865
1435
30
143
108
0
3110
1247
943
700
875
245
729
1897
447
386
446
122
990
948
1082
22
75
482
5509
100
10
1071
371
790
6150
3321
1045
648
5485
1160
1864
4116
20
Table 9.4. Successive failure times for
Jelinski-Moranda.
i
Mean time to ith failure
Simulated time to ith failure
1
22
11
2
24
41
3
26
13
4
28
4
5
30
30
6
33
77
7
37
11
8
42
64
9
48
54
10
56
34
11
67
183
12
83
83
13
111
17
14
167
190
15
333
436
21
Reliability, Availability and Maintainability
  • Mean Time to Failure (MTTF)
  • Mean time to Repair (MTTR)
  • Mean Time Between Failures (MTBF)
  • MTBF MTTF MTTR
  • Reliability R MTTF/(1MTTF)
  • Availability A MTBF / (1 MTBF)
  • Maintainability M 1 / (a MTTR)

22
Acceptance tests
  • Benchmark test customer prepares test cases to
    be tried out.
  • Pilot test install on experimental basis for
    all types of test cases be tried out
  • Alpha test in-house test
  • Beta test customer pilot
  • Parallel testing new system operates in
    parallel with old system

23
Installation Testing
  • Final round of testing involving installing the
    system at the user sites.
  • Automated testing could also be used. That is
    using simulated software for testing.

24
Test documentation
  • Test plan describes system and plan for
    exercising all functions and characteristics
  • Test specification and evaluation details each
    test and defines criteria for evaluating each
    feature
  • Test description test data and procedures for
    each test
  • Test analysis report results of each test

25
(No Transcript)
26
(No Transcript)
27
Test Analysis Report
  • During testing, data and faults are captured in a
    problem report form.
  • Discrepancy report form describes occurrences of
    problems where actual system behaviors dont
    match what is expected
  • Fault report form explains how a fault was found
    and fixed

28
Problem report forms cont
  • Problem report forms should include
  • Location Where did the problem occur
  • Timing When did it occur
  • Symptom What was observed
  • End result What were the consequenses
  • Mechanism How did it occur
  • Cause Why did it occur
  • Severity how much user and business affected
  • Cost How much did it cost

29
Testing safety-critical systems
  • Design diversity use different kinds of
    designs, designers and compare results
  • Software safety cases make explicit the ways
    the software addresses possible problems
  • failure modes and effects analysis
  • hazard and operability studies
  • Cleanroom certifying software with respect to
    the specification

30
Software safety cases
Table 9.6. Perspectives for safety analysis.
Known cause
Unknown cause
Known effect
Description of system behavior
Deductive analysis, including fault tree
analysis
Unknown effect
Inductive analysis, including failure
Exploratory analysis, including hazard
modes and effects analysis
and operability studies
31
Hazard and operability studies
Table 9.7. HAZOP guide words.
Guide word
Meaning
no
no data or control signal sent or received
more
data volume is too high or fast
less
data volume is too low or slow
part of
data or control signal is incomplete
other than
data or control signal has additional component
early
signal arrives too early for system clock
late
signal arrives too late for system clock
before
signal arrives earlier in sequence than expected
after
signal arrives later in sequence than expected
32
Table 9.8. SHARD guide words.
Failure categorization
Flow
Provision
Timing
Value
Protocol
Type
Omission
Early
Late
Subtle
Coarse
Commission
Pool
Boolean
no update
unwanted
N/A
old data
stuck at ...
N/A
update
Value
no update
unwanted
N/A
old data
wrong in
out of
update
tolerance
tolerance
Complex
no update
unwanted
N/A
old data
incorrect
inconsiste
update
nt
Channel
Boolean
no data
extra data
early
late
stuck at ...
N/A
Value
no data
extra data
early
late
wrong
out of
tolerance
tolerance
Complex
no data
extra data
early
late
incorrect
inconsiste
nt
33
Cleanroom control structures and correctness
conditions
34
(No Transcript)
Write a Comment
User Comments (0)
About PowerShow.com