Scheduling in computational grids with reservations - PowerPoint PPT Presentation

About This Presentation
Title:

Scheduling in computational grids with reservations

Description:

Scheduling in computational grids. with reservations. Denis Trystram. LIG-MOAIS ... Recently, there was a rapid and deep evolution of high-performance execution ... – PowerPoint PPT presentation

Number of Views:40
Avg rating:3.0/5.0
Slides: 129
Provided by: emm133
Category:

less

Transcript and Presenter's Notes

Title: Scheduling in computational grids with reservations


1
Scheduling in computational gridswith
reservations
  • Denis Trystram
  • LIG-MOAIS
  • Grenoble University, France
  • AEOLUS, march 9, 2007

2
General Context
Recently, there was a rapid and deep evolution of
high-performance execution platforms
supercomputers, clusters, computational grids,
global computing, Need of efficient tools for
resource management for dealing with these new
systems. This talk will investigate some
scheduling problems and focus on reservations.
3
Parallel computing today.
  • Different kinds of platforms
  • Clusters, collection of clusters, grid, global
    computing
  • Set of temporary unused resources
  • Autonomous nodes (P2P)
  • Our view of grid computing (reasonable
    trade-off)
  • Set of computing resources under control (no hard
    authentication problems, no random addition of
    computers, etc.)

4
Content
  • Some preliminaries (Parallel tasks model)
  • Scheduling and packing problems
  • On-line versus off-line batch scheduling
  • Multi-criteria
  • Reservations

5
A national french initiativeGRID5000
Several local computational grids (like
CiGri) National project with shared resources and
competences with almost 4000 processors today
with local administration but centralized control.
6
(No Transcript)
7
Target Applications
New execution supports created new applications
(data-mining, bio-computing, coupling of codes,
interactive, virtual reality, ). Interactive
computations (human in the loop), adaptive
algorithms, etc.. See MOAIS project for more
details.
8
Scheduling problem(informally)
Given a set of tasks, the problem is to
determine when and where to execute the tasks
(according to the precedence constraints - if any
- and to the target architecture).
9
Central Scheduling Problem
The basic problem P prec, pj Cmax is NP-hard
Ulmann75. Thus, we are looking for  good 
heuristics.
10
Central Scheduling Problem
The basic problem P prec, pj Cmax is NP-hard
Ulmann75. Thus, we are looking for  good 
heuristics.
based on theoretical analysis good approximation
factor
low cost
11
Available models
Extension of  old  existing models
(delay) Parallel Tasks Divisible load
12
Delay if two consecutive tasks are allocated on
different processors, we have to pay a
communication delay.
13
Delay if two consecutive tasks are allocated on
different processors, we have to pay a
communication delay.If L is large, the problem
is very hard (no approximation algorithm is
known)
L
14
Extensions of delay
Some tentatives have been proposed (like
LogP). Not adequate for grids (heterogeneity,
large delays, hierarchy, incertainties)
15
Parallel Tasks
Extension of classical sequential tasks each
task may require more than one processor for its
execution Feitelson and Rudolph.
16
Job
17
(No Transcript)
18
(No Transcript)
19
(No Transcript)
20
(No Transcript)
21
(No Transcript)
22
overhead
Computational area
23
Classification
rigid tasks
24
Classification
moldable tasks
25
Classification
moldable tasks
decreasing time
26
Classification
moldable tasks
increasing area
27
Classification
malleable tasks
28
Classification
extra overhead
malleable tasks
29
Divisible load
Also known as  bag of tasks  Big amount of
arbitrary small computational units.
30
Divisible load
Also known as  bag of tasks  Big amount of
arbitrary small computational units.
31
Divisible load
(asymptotically) optimal for some criteria
(throughput). Valid only for specific
applications with regular patterns. Popular for
best effort jobs.
32
Resource management in clusters
33
Users queue
time
job
34
Users queue
time
35
Users queue
time
36
Users queue
time
37
Users queue
time
38
Users queue
time
39
Users queue
time
40
Users queue
time
41
Users queue
time
42
Users queue
time
43
Users queue
time
44
Users queue
time
45
Users queue
time
46
Users queue
time
47
Integrated approach
48
m
49
m
50
m
51
m
52
m
53
m
54
m
55
m
56
m
57
m
58
m
59
m
60
(strip) Packing problems
  • The schedule is divided into two successive
    steps
  • Allocation problem
  • Scheduling with preallocation (NP-hard in general
    Rayward-Smith 95).

61
Scheduling on-line vs off-line
On-line no knowledge about the future
We take the scheduling decision while other jobs
arrive
62
Scheduling on-line vs off-line
Off-line we have a finite set of works
We try to find a good arrangement
63
Off-line scheduler
Problem Schedule a set of independent moldable
jobs (clairvoyant). Penalty functions have
somehow to be estimated (using complexity
analysis or any prediction-measurement method
like the one obtained by the log analysis).
64
Example
Let us consider 7 MT to be scheduled on m10
processors.
65
Canonical Allotment
1
W/m
66
Canonical Allotment
1
m
Maximal number of processors needed for executing
the tasks in time lower than 1.
67
2-shelves scheduling
Idea to analyze the structure of the optimum
where the tasks are either greater than 1/2 or
not. Thus, we will try to fill two shelves with
these tasks.
68
2 shelves partitioning
1/2
1
m
Knapsack problem minimizing the global surface
under the constraint of using less than m
processors in the first shelf.
69
Dynamic programming
  • For i 1..n // of tasks
  • for j 1..m // proc.
  • Wi,j min(
  • Wi,j-minalloc(i,1) work(i,minalloc(i,1))
  • Wi,j work(i,minalloc(i,1))
  • )
  • work Wn,m
  • lt work of an optimal solution
  • but the half-sized shelf may be overloaded

70
2 shelves partitioning
1/2
1
m
71
Drop down
1/2
1
m
72
Insertion of small tasks
1/2
1
m
73
Analysis
  • These transformations donot increase the work
  • If the 2nd shelf is used more than m, it is
    always possible to do one of the transformations
    (using a global surface argument)
  • It is always possible to insert the  small 
    sequential tasks (again by a surface argument)

74
Guaranty
  • The 2-shelves algorithm has a performance
    guaranty of 3/2e
  • (SIAM J. on Computing, to appear)
  • Rigid case 2-approximation algorithm (Graham
    resource constraints)

75
Batch scheduling
Principle several jobs are treated at once using
off-line scheduling.
76
Principle of batch
jobs arrival
time
77
Start batch 1
78
(No Transcript)
79
Batch chaining
Batch i
80
Batch chaining
Batch i
81
Batch chaining
Batch i
82
Batch chaining
Batch i
83
Batch chaining
Batch i
84
Batch chaining
Batch i
Batch i1
85
Constructing a batch scheduling
Analysis there exists a nice (simple) result
which gives a guaranty for an execution in batch
mode using the guaranty of the off-line
scheduling policy inside the batches.
86
Analysis Shmoys
previous last batch last batch
Cmax
r (last job)
n
87
D
D
k
K-1
previous last batch last batch
Cmax
r
n
k
T
88
Proposition
89
Analysis
  • Tk is the duration of the last batch
  • On another hand, and
  • Thus

90
Application
  • Applied to the best off-line algorithm for
    moldable jobs (3/2-approximation), we obtain a
    3-approximation on-line batch algorithm for Cmax.
  • This result holds also for rigid jobs (using the
    2-approximation Graham resource constraints),
    leading to a 4-approximation algorithm.

91
Multi criteria
  • Cmax is not always the adequate criterion.
  • User point of view
  • Average completion time (weighted or not)
  • Other criteria
  • Stretch, Asymptotic throughput, fairness,

92
How to deal with this problem?
  • Hierachal approach one criterion after the other
  • (Convex) combination of criteria
  • Transforming one criterion in a constraint
  • Better - but harder - ad hoc algorithms

93
A first solution
Construct a feasible schedule from two
schedules of guaranty r for minsum and r for
makespan with a guaranty (2r,2r) Stein et
al.. Instance 7 jobs (moldable tasks) to be
scheduled on 5 processors.
94
Schedules s and s
3
5
1
2
7
Schedule s (minsum)
6
4
7
1
Schedule s (makespan)
2
4
6
5
3
95
New schedule
3
5
1
2
7
6
4
rCmax
7
1
2
4
6
5
3
96
New schedule
3
5
1
2
7
6
4
7
6
5
97
New schedule
3
1
2
4
7
6
5
98
New schedule
3
1
2
7
4
6
5
2rCmax
99
New schedule
3
1
2
7
4
6
5
2rCmax
Similar bound for the first criterion
100
Analysis
The best known schedules are 8 Schwiegelsohn
for minsum and 3/2 Mounie et al. for makespan
leading to (163). Similarly for the weighted
minsum (ratio 8.53 for minsum).
101
Improvement
We can improve this result by determining the
Pareto curves (of the best compromises) (1l)/
l r and (1 l)r Idea take the first part of
schedule s up to l rCmax
102
Pareto curve
103
Pareto curve
104
Another way for designing better schedules
We proposed SPAA2005 a new solution for a
better bound which has not to consider explicitly
the schedule for minsum (based on a dynamic
framework). Principle recursive doubling with
smart selection (using a knapsack) inside each
interval. Starting from the previous algorithm
for Cmax, we obtain a (66) approximation.
105
Bi criteria Cmax and SwiCi
  • Generic On-line Framework Shmoys et al.
  • Exponantially increasing time intervals
  • Uses a max-weight r approximation algorithm
  • If the optimal schedule of length d has weight
    w, provides a schedule of length rd and weight
    ? w
  • Yields a (4r, 4r) approximation algorithm
  • For moldable tasks, yields a (12, 12)
    approximation
  • With the 2-shelf algorithm, yields a (6, 6)
    approximation Dutot et al.

106
Example for r 2
Shortest job
Schedule for makespan
107
Example for r 2
"Contains more weight"
t
16
0
2
4
8
108
A last trick
  • The intervals are shaked (like in 2-opt local
    optimization techniques).
  • This algorithm has been adapted for rigid tasks.
  • It is quite good in practice, but there is no
    theoretical guaranty

109
Reservations
  • Motivation
  • Execute large jobs that require more than m
    processors.

time
110
Reservations
111
Reservations
  • The problem is to schedule n independent parallel
    rigid tasks such that the last finishing time is
    minimum.

q
m
At each time t, r(t)m processors are not
available
112
State of the art
  • Most existing results deal with sequential tasks
    (qj1).

Without preemption Decreasing reservations Only
one reservation per machine
With preemption Optimal algorithms for
independent tasks Optimal algorithms for some
simple task graphs
113
Without reservation
2
3
1
4
5
FCFS with backfilling
114
Without reservation
2
3
1
4
5
FCFS with backfilling
115
Without reservation
3
1
4
2
5
FCFS with backfilling
116
Without reservation
3
4
1
2
5
FCFS with backfilling
117
Without reservation
4
3
1
2
5
FCFS with backfilling
118
Without reservation
4
5
3
1
2
FCFS with backfilling
119
Without reservation
  • List algorithms use available processors for
    executing the first possible task in the list.

4
4
5
3
3
1
1
5
2
2
FCFS with backfilling
list algorithm
120
Without reservation
  • Proposition list algorithm is a 2-1/m
    approximation.
  • This is a special case of Graham 1975 (resource
    constraints), revisited by Eyraud et al. IPDPS
    2007.
  • The bound is tight (same example as in the
    well-known case in 1969 for sequential tasks).

121
With reservation
  • The guaranty is not valid.
  • This is a special case of Graham 1975 (resource
    constraints), revisited by Eyraud et al. IPDPS
    2007.
  • The bound is tight (same example as in the
    well-known case in 1969 for sequential tasks).

122
Complexity
  • The problem is already NP-hard with no
    reservation.
  • Even worse, an optimal solution with arbitrary
    reservation may be delayed as long as we want

Cmax
123
Complexity
  • The problem is already NP-hard with no
    reservation.
  • Even worse, an optimal solution with arbitrary
    reservation may be delayed as long as we want

Cmax
124
Complexity
  • The problem is already NP-hard with no
    reservation.
  • Even worse, an optimal solution with arbitrary
    reservation may be delayed as long as we want

Conclusion can not be approximated unless PNP,
even for m1
125
Two preliminary results
  • Decreasing number of available processors
  • Restricted reservation problem always a given
    part a of the processors is available r(t) (1-
    a)m and for all task i, qi am.

(1-a) m
am
126
Analysis
  • Case 1. The same approximation bound 2-1/m is
    still valid
  • Case 2. The list algorithm has a guaranty of 2/a
  • Insight of the proof while the optimal uses m
    processors, list uses only a processors with a
    approximation of 2

127
Analysis
  • Case 1. The same approximation bound 2-1/m is
    still valid
  • Case 2. The list algorithm has a guaranty of 2/a
  • Insight of the proof while the optimal uses m
    processors, list uses only a processors with a
    approximation of 2
  • There exists a lower bound which is arbitrary
    close to this bound
  • 2/a -1 a/2 if 2/a is an integer

128
Conclusion
It remains a lot of interesting open problems
with reservations. Using preemption Not rigid
reservations Better approximation (more costly)
Write a Comment
User Comments (0)
About PowerShow.com