Title: Software Performance Testing Based on Workload Characterization
1Software Performance Testing Based on Workload
Characterization
- Elaine Weyuker
- Alberto Avritzer
- Joe Kondek
- Danielle Liu
- ATT Labs
2Workload Characterization
A probability distribution associated with the
input domain that describes how the system is
used when it is operational in the field. Also
called an operational profile or operational
distribution. It is derived by monitoring field
usage. We have used it to select (correctness)
test cases, predict risk, assess software
reliability, predict scalability,
select performance test cases, among other things.
3Steps in Characterizing the Workload
- Model the software system.
- Identify key parameters that characterize the
systems behavior. - Getting the granularity right.
- Collect data while the system is operational
- or from related operational system.
- Analyze the data and determine the probability
distribution.
4System Description
Automated customer care system, built by another
company, that can be accessed by both customer
care agents and customers. It contains a large
database with a web browser front-end and a cache
facility. For this system data was collected for
2 ½ months and analyzed page hits at 15 minute
intervals.
5Implementation Information
The system was implemented as an extension to the
http web server daemon, with a mutex semaphore
used to implement database locking. The system
is single threaded. Queued processes executed
spin lock operations until the semaphore was
free.
6Before Performance Testing
Prior to doing performance testing, users were
complaining about poor performance, and the
database was hanging several times a day. The
hypothesis was that these problems
were capacity-related, and the vendor was
contacted but unable to solve the problems.
7Performance Testing Goals
- Help them determine
- Which resources were overloaded.
- Effects of the database size on performance.
- Effects of single vs. multiple transactions.
- Effects of the cache hit rate.
- Effects of the number of http servers.
8System Information
- The web-server database was modeled as an M/D/1
queue. - The arrival process was assumed to be Poisson.
- The cache hit rate was determined to be central
to the systems performance. It ranged between
80-87, with the average being 85.
9Distribution of User Requests
10Computing the Cache Hit Probability
11System Observations
- Heavier workload for agents on weekdays, peak
hours in the afternoon, with little day-to-day
variation. - Customer peak hours occurred during the
evening. - Little change in workload as users become
familiar with the system. (Agents are already
expert users and execute a well-defined process,
while individual customers tend to use the system
rarely and therefore also maintain the same usage
pattern over time.)
12What We Achieved
- Characterized the agent and customer workload,
and used it as a basis for performance testing. - Identified performance limits and as a result
detected a software bottleneck. - Provided recommendations for performance
improvement. - Increased the understanding of the use of data
collection for performance issues.
13What We Learned
- The system was currently running at about 20
- utilization.
- The CISCO routers were not properly load
balanced. - The spin lock operations consumed CPU time which
- led to a steep increase in the response time. (We
used - the SUN SE toolkit to record the number of spin
- locks).
14No Caching
7
6
5
4
Customer
Agent
avg CPU cost(sec)/request
3
2
1
0
0
0.5
1
1.5
2
2.5
Load (hits/sec)
15No Caching
40000
35000
30000
25000
Customer
avg response time(ms)
20000
Agent
15000
10000
5000
0
0
0.5
1
1.5
2
2.5
Load (hits/sec)
16All Requests Retrieved From Cache
5000
4500
4000
3500
3000
avg response time(ms)
Customer
2500
Agent
2000
1500
1000
500
0
0
20
40
60
80
100
120
140
160
180
Load (hits/sec)
17Simulated 85 Cache Hit Rate
60000
50000
40000
30000
Customer
avg response time (ms)
Agent
20000
10000
0
0
0.5
1
1.5
2
2.5
3
3.5
4
-10000
Load (hits/sec)
18In Particular
- Delay strongly depends on caching
- Found in cache 100ms
- Retrieved from database 5 secs
- Current Available capacity
- Customer 2 hits/sec
- Agent 2.5 hits/sec
- Average demand
- Customer 10,000 hits/day 0.12 hits/sec
- Agent 25,000 hits/day 0.29 hits/sec
- Busy Hour demand
- Customer 784 hits/hour 0.22 hits/sec
- Agent 2228 hits/hour 0.62 hits/sec
19Effect of Database SizeCustomer Cache Off
60000
50000
40000
avg response time (ms)
Small DB (200 MB)
Medium DB (400 MB)
30000
Large DB (600 MB)
20000
10000
0
0
0.2
0.4
0.6
0.8
1
1.2
1.4
1.6
1.8
2
Load (hits/sec)
20Effect of Database SizeAgent Cache Off
35000
30000
25000
20000
Small DB (200 MB)
avg response time (ms)
Medium DB (400 MB)
Large DB (600 MB)
15000
10000
5000
0
0
0.2
0.4
0.6
0.8
1
1.2
1.4
1.6
1.8
2
Load (hits/sec)
21Adding Servers
For this system, n servers meant n service
queues, each operating independently, and hence
less lock contention. This led to a significant
increase in the workload that could be
handled. However, since each server maintains
its own caching mechanism, there was a distinct
decrease in the cache hit probability, and an
associated increase in the response time. The
response time is dominated by the cache hit
probability when the load is low as the load
increases the queuing for the database also
increases.
22Multiple Servers
60000
50000
40000
1 web server
avg response time (ms)
2 web servers
3 web servers
30000
4 web servers
5 web servers
6 web servers
20000
10000
0
0
1
2
3
4
5
6
7
8
Load (hits/sec)
23Recommendations
- Projects should collect traffic data on a daily
basis. - Performance measurements should be made while the
software is being testing in the lab both for
new systems and when changes are being made. - Workload-based testing is a very cost effective
way to do performance testing.
24THE END
25No Caching
70000
60000
50000
40000
Error
Result
avg resp time(ms)
Form
Static
30000
20000
10000
0
0
0.5
1
1.5
2
2.5
hits/sec
26Cache Off
30
25
20
Error
Result
15
Form
avg CPU cost(sec)/request
Static
10
5
0
0
0.5
1
1.5
2
2.5
hits/sec