Title: Tony Doyle University of Glasgow
1ScotGrid, GridPPand EU DataGrid
- Joint Edinburgh/Glasgow SHEFC JREI-funded project
to develop a prototype Tier-2 centre for LHC
computing. - UK-wide project to develop a prototype Grid for
Particle Physics Applications. - EU-wide project to develop middleware for
Particle Physics, Bioinformatics and Earth
Observation applications - Emphasis on local developments
2Outline
- Introduction
- ScotGrid
- Motivation and Overview
- How Does the Grid Work?
- Middleware Development
- Grid Data Management
- Testbed Status
- Is the Middleware Robust?
- Tier-1 and -2 Centre Resources
- GridPP Achievements
- Timeline
3ScotGRID
- ScotGRID Processing nodes at Glasgow
- 59 IBM X Series 330 dual 1 GHz Pentium III with
2GB memory - 2 IBM X Series 340 dual 1 GHz Pentium III with
2GB memory and dual ethernet - 3 IBM X Series 340 dual 1 GHz Pentium III with
2GB memory and 100 1000 Mbit/s ethernet - 1TB disk
- LTO/Ultrium Tape Library
- Cisco ethernet switches
- ScotGRID Storage at Edinburgh
- IBM X Series 370 PIII Xeon with 512 MB memory 32
x 512 MB RAM - 70 x 73.4 GB IBM FC Hot-Swap HDD
Applications
Middleware
Hardware
4Starting Point (Dec 2000)
5Rare Phenomena Huge Background
All interactions
9 orders of magnitude
The HIGGS
6LHC Computing Challenge
PBytes/sec
Online System
100 MBytes/sec
Offline Farm20 TIPS
- One bunch crossing per 25 ns
- 100 triggers per second
- Each event is 1 Mbyte
100 MBytes/sec
Tier 0
CERN Computer Centre gt20 TIPS
Gbits/sec
or Air Freight
HPSS
Tier 1
RAL Regional Centre
US Regional Centre
French Regional Centre
Italian Regional Centre
HPSS
HPSS
HPSS
HPSS
Tier 2
Tier2 Centre 1 TIPS
Tier2 Centre 1 TIPS
Tier2 Centre 1 TIPS
Gbits/sec
Tier 3
Physicists work on analysis channels Each
institute has 10 physicists working on one or
more channels Data for these channels should be
cached by the institute server
Institute 0.25TIPS
Institute
Institute
Institute
Physics data cache
100 - 1000 Mbits/sec
Tier 4
Workstations
7How Does theGrid Work?
Tutorial _at_ NeSC Next Mon-Tuesday
1. Authentication grid-proxy-init 2. Job
submission dg-job-submit 3. Monitoring and
control dg-job-status dg-job-cancel dg-job-get-
output 4. Data publication and replication globus
-url-copy, GDMP 5. Resource scheduling use of
Mass Storage Systems JDL, sandboxes, storage
elements
0. Web User Interface
8Middleware Development
9Grid Data Management
- Secure access to metadata
- metadata where are the files on the grid?
- database client interface
- grid service using standard web services
- develop with UK e-science programme
- Input to OGSA-DAI
- Optimised file replication
- simulations required
- economic models using CPU, disk, network inputs
- OptorSim
10Testbed Status
11Is the Middleware Robust?
- Code Base
- Software Evaluation Process
- Testbed Infrastructure WP-specific
?Development?Certification?Application. - Code Development Platforms
12Application Interfacesunder development
13Tier-2 Web-BasedMonitoring
Prototype
Accumulated CPU Use
Total Disk Use
ScotGrid reached its 100,000th processing hour on
Wednesday 13th November 2002.
Documentation
Instantaneous CPU Use
14(No Transcript)
15Lattice QCD
- QCD is the theory of the strong interactions
(QCD) between quarks and gluons. In numerical
simulations it is very expensive, but also very
important to include quark/antiquark pairs
popping in and out of the vacuum. - On ScotGrid we have been studying new
formulations for quarks which will allow us to do
realistic calculations on UKQCDs JIF-funded
machine in 2003/4.
16Tier-1 and -2 Centre Resources
- Estimated resources at end of 2003 (from
Institute returns)
Tier-1
- Tier-1 600 CPUs 150 TB
- Tier-2 e(4000 CPUs 200 TB)
17Demonstrations
- Super Computing 2002 Baltimore, US
- Major event last week
- GridPP participated in three successful Worldwide
demos - WorldGrid
- Replica Location Service
- SAMGrid
- These and other Web-based demos available online
from http//www.gridpp.ac.uk/demos/
18Year 0 Year 1
The GridPP Project has now completed one year .
19Achievements I
- Dedicated people actively developing a Grid
- All with personal certificates
- Using the largest UK grid testbed
- (16 sites and more than 100 servers)
- Deployed within EU-wide programme
- Linked to Worldwide Grid testbeds
20Achievements II
- Grid Deployment Programme Defined The Basis
for LHC Computing - Active Tier-1/A Production Centre meeting
International Requirements - Latent Tier-2 resources being monitored
- Significant middleware development programme
- First simple applications using the Grid testbed
(open approach)
21Outlook
General deployment of e-Science methods
2002
2005
2004
2003
2001
GridPP
Certs.
From Prototype (hundreds) to Production
(thousands) Grid..
22 Timeline
DataGrid
GridPP-Procure, Install, Compute, Data
Develop, Test, Refine
LHC Computing Grid
Prototypes
Production
Initial Grid Tests
Worldwide Grid Demonstrations
Q1 Q2 Q3 Q4
Q1 Q2 Q3 Q4
Q1 Q2 Q3 Q4
Q1 Q2 Q3 Q4
Q1 Q2 Q3 Q4
2001
2002
2005
2004
2003
GridPP II Grid Service
EGEE (DataGrid II?)
Middleware and Hardware upgrades
Transition and Planning Phase
23Summary
Hardware
Middleware
Applications
- Grid Data Management - fundamental to Grid
Development - University strategic investment -
refurbishment, running costs, networking - Software prototyping (GDM) and stress-testing
(Applications) - Long-term commitment (LHC era)
- development/deployment
- Partnership of Computing Science, Particle
Physics, Edinburgh, Glasgow, IBM - Working locally as part of National, European and
International Grid development
ScotGRID