9th GridPP Collaboration Meeting - PowerPoint PPT Presentation

1 / 16
About This Presentation
Title:

9th GridPP Collaboration Meeting

Description:

Currently has 4 sites with 7 RAID disk nodes ... document for each data file. XML document storage ... Grid recovery tools including switching of control thread ... – PowerPoint PPT presentation

Number of Views:20
Avg rating:3.0/5.0
Slides: 17
Provided by: dranna8
Category:

less

Transcript and Presenter's Notes

Title: 9th GridPP Collaboration Meeting


1
9th GridPP Collaboration Meeting
QCDgrid Status and Future Alan Irving University
of Liverpool
2
UKQCD and the Grid QCDgrid architecture
  • PPARC support
  • GridPP1 Phase 1 data grid ?
  • GridPP1 Phase 2 pilot scheme for
  • distributed processing ?
  • GridPP2 Full distributed processing
  • GridPP2 International Lattice Data
  • Grid activities (ILDG)

3
QCDOC Columbia IBM UKQCD BNL
10,000 processors 10 Tflops, 6.6M, July 2004
128 procs Nov 03
4
Stop press ....
  • Following exhaustive tests of the ASIC, orders
    have now been placed for some 14,720 ASICS for
  • 2048 node development machine ( gt 1 Tflop
    sustained) for assembly in March
  • 12,000 node main machine, for assembly in May

5
UKQCD computing strategy with QCDOC
  • Distributed computing ? Grid
  • International standards (ILDG)
  • SCIDAC US strategy
  • Local resources compute/data
  • Tier 1 Edinburgh
  • Tier 2 Edinburgh, Liverpool, Swansea,
    Southampton ( RAL)
  • UKQCD approved simulations
  • International cooperation with MILC,
    Columbia,..
  • Data grid for configuration acquisition and
    storage
  • International nodes available
  • Job submission software (JSS) for homogeneous
    physics analysis within UKQCD
  • Need for significant clusters at computational
    nodes, (Liverpool, RAL, ...)

QCDOC
FE
Grid
Node
6
Basics of the QCDgrid datagrid
  • Currently has 4 sites with 7 RAID disk nodes
  • Main design and implementation by EPCC (James
    Perry)
  • Admin by C Maynard (Physics/Edinburgh) local
    sys admins
  • User requirement/testing driven by Liverpool (C
    McNeile)
  • File replication managed by custom written
    software built on Globus 2
  • Central control thread ensures at least 2 copies
    of each file at different sites
  • Replica catalogue maps logical names to physical
    locations
  • Metadata catalogue associates physical parameters
    with files
  • XML document for each data file
  • XML document storage in eXist XML database,
    queried by Xpath

7
Operation of the QCDgrid datagrid
  • Initial queries via browser GUI
  • Production running via command line tools
  • Current developments
  • Simple interface for data/metadata submission
    under development
  • Grid administration tools
  • Grid recovery tools including switching of
    control thread
  • EDG software for virtual organisation management
    and security.
  • Data binding in QCDOC codes

8
QCDgrid metadata browser
9
Pilot version of job submission software
  • Globus toolkit
  • EDG software for VO management
  • and security
  • Integrated with datagrid SW
  • Pilot running on test grid at EPCC
  • Command line job submission
  • Job IO can go to user console
  • Output files returned automatically
  • Soon ...
  • Deploy on main grid
  • Integrate with batch systems (PBS ..)
  • Better user interface (GUI..)
  • GridPP2..
  • Full system with real analysis code

10
job submission test
alan_at_qcdtest gridwork qcdgrid-job-submit
qcdother.epcc.ed.ac.uk \
/home/alan/gridwork/testrn \ -input
/home/alan/gridwork/in_seed.dat Storing results
in local directory qcdgridjob000002 Storing
results in remote directory /tmp/qcdgridjob000024
RSL(executable/opt/qcdgrid/qcdgrid-job-controll
er) (arguments/tmp/qcdgridjob000024/jobdesc) (e
nvironment(LD_LIBRARY_PATH /opt/globus/lib/opt/q
cdgrid)) Connecting to port 16395... OUTPUT
iter r.n. 0 0.586089 1
0.651327 r.n. seeds written to out_seed.dat
testrn finished Ok! Job has
completed Retrieving jobdesc Retrieving
controller.log Retrieving wrapper.log Retrieving
stdout Retrieving stderr Retrieving
out_seed.dat alan_at_qcdtest gridwork
11
International Lattice Data Grid
  • UKQCD launched this in 2002 in Boston
  • Participants from USA(Scidac), Japan,
    Germany,..
  • Enable data sharing
  • Agree standards
  • Steering group of national reps ..
  • 2 working groups
  • Metadata WG
  • XML schema
  • gauge formats etc
  • Middleware WG
  • Web service standards
  • Storage Resource Manager

Feb 3 CP-PACS (Japan) launch ILDG node at
http//www.lqa.rccp.tsukuba.ac.jp/
12
3-continent file browsing
UKQCD
JLAB
LATT03
13
ILDG file browser
14
QCDgrid and GridPP2
  • Extend Job Submission Software, resource
    brokering,..
  • XML mark-up within main QCDOC production codes
  • Web services implementation of replica and
    metadata catalogues
  • Web services ILDG replica and metadata catalogues
  • Web services based compute grid using UK and
    non-UK nodes

15
QCDgrid websites
  • QCDgrid home page (at GridPP?)
  • http//www.gridpp.ac.uk/qcdgrid
  • QCDgrid project page at NeSCForge development
    site
  • http//qcdgrid.forge.nesc.ac.uk/
  • ILDG project page at JLAB, USA
  • http//qcdgrid.lqcd.org/ildg

16
CONCLUSIONS
  • UKQCD has operational data grid (QCDgrid)
  • QCDOC preparations are well advanced
  • Tier 2 nodes have been (are being) installed
  • Work continues on XML tools
  • Prototype job submission SW exists and is being
    developed
  • International activity is increasing (
    )
  • Open software development via NeSC Forge
Write a Comment
User Comments (0)
About PowerShow.com